MAGI: Enabling multi-device gestural applications

We describe our vision of a multiple mobile or wearable device environment and share our initial exploration of our vision in multi-wrist gesture recognition. We explore how multi-device input and output might look, giving four scenarios of everyday multi-device use that show the technical challenge...

Full description

Saved in:
Bibliographic Details
Main Authors: TRAN HUY VU, CHOO TSU WEI, KENNY, LEE, Youngki, DAVIS, Richard Christopher, MISRA, Archan
Format: text
Language:English
Published: Institutional Knowledge at Singapore Management University 2016
Subjects:
Online Access:https://ink.library.smu.edu.sg/sis_research/3287
https://ink.library.smu.edu.sg/context/sis_research/article/4289/viewcontent/MAGI.pdf
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Singapore Management University
Language: English
id sg-smu-ink.sis_research-4289
record_format dspace
spelling sg-smu-ink.sis_research-42892018-03-08T01:04:34Z MAGI: Enabling multi-device gestural applications TRAN HUY VU, CHOO TSU WEI, KENNY, LEE, Youngki DAVIS, Richard Christopher MISRA, Archan We describe our vision of a multiple mobile or wearable device environment and share our initial exploration of our vision in multi-wrist gesture recognition. We explore how multi-device input and output might look, giving four scenarios of everyday multi-device use that show the technical challenges that need to be addressed. We describe our system which allows for recognition to be distributed between multiple devices, fusing recognition streams on a resource-rich device (e.g., mobile phone). An Interactor layer recognises common gestures from the fusion engine, and provides abstract input streams (e.g., scrolling and zooming) to user interface components called Midgets. These take advantage of multi-device input and output, and are designed to simplify the process of implementing multi-device gestural applications. Our initial exploration of multi-device gestures led us to design a modified pipelined HMM with early elimination of candidate gestures that can recognize gestures in almost 0.2 milliseconds and scales well to large numbers of gestures. Finally, we discuss the open problems in multi-device interaction and our research directions. 2016-03-01T08:00:00Z text application/pdf https://ink.library.smu.edu.sg/sis_research/3287 info:doi/10.1109/PERCOMW.2016.7457168 https://ink.library.smu.edu.sg/context/sis_research/article/4289/viewcontent/MAGI.pdf http://creativecommons.org/licenses/by-nc-nd/4.0/ Research Collection School Of Computing and Information Systems eng Institutional Knowledge at Singapore Management University Cellular telephone systems Ubiquitous computing User interfaces Multiple devices Resource-rich devices Wearable devices Gesture recognition Computer Sciences Software Engineering
institution Singapore Management University
building SMU Libraries
continent Asia
country Singapore
Singapore
content_provider SMU Libraries
collection InK@SMU
language English
topic Cellular telephone systems
Ubiquitous computing
User interfaces
Multiple devices
Resource-rich devices
Wearable devices
Gesture recognition
Computer Sciences
Software Engineering
spellingShingle Cellular telephone systems
Ubiquitous computing
User interfaces
Multiple devices
Resource-rich devices
Wearable devices
Gesture recognition
Computer Sciences
Software Engineering
TRAN HUY VU,
CHOO TSU WEI, KENNY,
LEE, Youngki
DAVIS, Richard Christopher
MISRA, Archan
MAGI: Enabling multi-device gestural applications
description We describe our vision of a multiple mobile or wearable device environment and share our initial exploration of our vision in multi-wrist gesture recognition. We explore how multi-device input and output might look, giving four scenarios of everyday multi-device use that show the technical challenges that need to be addressed. We describe our system which allows for recognition to be distributed between multiple devices, fusing recognition streams on a resource-rich device (e.g., mobile phone). An Interactor layer recognises common gestures from the fusion engine, and provides abstract input streams (e.g., scrolling and zooming) to user interface components called Midgets. These take advantage of multi-device input and output, and are designed to simplify the process of implementing multi-device gestural applications. Our initial exploration of multi-device gestures led us to design a modified pipelined HMM with early elimination of candidate gestures that can recognize gestures in almost 0.2 milliseconds and scales well to large numbers of gestures. Finally, we discuss the open problems in multi-device interaction and our research directions.
format text
author TRAN HUY VU,
CHOO TSU WEI, KENNY,
LEE, Youngki
DAVIS, Richard Christopher
MISRA, Archan
author_facet TRAN HUY VU,
CHOO TSU WEI, KENNY,
LEE, Youngki
DAVIS, Richard Christopher
MISRA, Archan
author_sort TRAN HUY VU,
title MAGI: Enabling multi-device gestural applications
title_short MAGI: Enabling multi-device gestural applications
title_full MAGI: Enabling multi-device gestural applications
title_fullStr MAGI: Enabling multi-device gestural applications
title_full_unstemmed MAGI: Enabling multi-device gestural applications
title_sort magi: enabling multi-device gestural applications
publisher Institutional Knowledge at Singapore Management University
publishDate 2016
url https://ink.library.smu.edu.sg/sis_research/3287
https://ink.library.smu.edu.sg/context/sis_research/article/4289/viewcontent/MAGI.pdf
_version_ 1770573075116457984