Team: Anirudh Sharma, Sriganesh Madhvanath, Mark Billinghurst

According to a survey 80 percent of people want to model/create to visualize their imagination using their computers, however the difficult UI of such tools prevents them from doing so. Even to model something the user has to go through the obtrusive set of icons/toolbars/features which are rarely used.

We propose a computer modeling interface to bring it to the common layman who wish to rapidly visualize their imagination in 3D. This work is motivated to employ natural expression with the fewest restrictions to free CAD users from the tedious command buttons and menu items. We have explored both the hardware and software aspects of the interface, specifically, the use of intuitive speech commands and multitouch gestures on an inclined interactive surface. 

The very initial TUIO and OSC touch-based integration was done at Google SoC. Touch+Speech Multimodal fusion with SriG at HP Labs. 

Thanks: Srig, NUIgroup forums. CCV Forum.

Dec 2010: Our research paper on Mozart selected for publication at ACM IndiaHCI 2011[/stextbox] 

Sept 2011: MozArt published and demo-ed at 13th ACM International Conference on Multimodal Interfaces 2011(ICMI). Mozart ACM ICMI Research Paper