03-05-2011, 04:17 PM
[attachment=13300]
Abstract
This paper presents a new approach for controlling
mouse movement using a real-time camera. Most ex-
isting approaches involve changing mouse parts such
as adding more buttons or changing the position of the
tracking ball. Instead, we propose to change the hard-
ware design. Our method is to use a camera, image
comparison technology and motion detection technol-
ogy to control mouse movement.
Introduction
As computer technology continues to develop, people
have smaller and smaller electronic devices and want
to use them ubiquitously. There is a need for new
interfaces designed specically for use with devices.
Increasingly we are recognizing the importance of hu-
man computing interaction (HCI), and in particular
vision-based gesture and object recognition. Simple
interfaces already exist, such as embedded keyboard,
folder-keyboard and mini-keyboard. However, these
interfaces need some amount of space to use and can-
not be used while moving. Touch screens are also a
good control interface and nowadays it is used globally
in many applications. However, touch screens can-
not be applied to desktop systems because of cost and
other hardware limitations. By applying vision tech-
nology and controlling the mouse by natural hand ges-
tures, we can reduce the work space required. In this
paper, we propose a novel approach that uses a video
device to control the mouse system.
2 Related Work
2.1 MouseFree
Vision-Based Human-Computer Interaction through
Real-Time Hand Tracking and Gesture Recognition
Vision-based interaction is an appealing option for re-
placing primitive human-computer interaction (HCI)
using a mouse or touchpad. We propose a system for
using a webcam to track a users hand and recognize
gestures to initiate specic interactions. The contri-
butions of our work will be to implement a system for
hand tracking and simple gesture recognition in real
time.
This is reference to biblio [1]
2.2 A Method for Controlling Mouse
Movement using a Real- Time
Camera
This is a new approach for controlling mouse move-
ment using a real-time camera. Most existing ap-
proaches involve changing mouse parts such as adding
more buttons or changing the position of the tracking
ball. Instead, we propose to change the hardware de-
sign. Our method is to use a camera and computer vi-
sion technology, such as image segmentation and ges-
ture recognition.Our method is to use a camera and
computer vision technology, such as image segmenta-
tion and gesture recognition, to control mouse tasks
(left and right clicking, double-clicking, and scrolling)
and we show how it can perform everything current
mouse devices can. This paper shows how to build
this mouse control system.
This is reference to biblio [2]
2.3 SixthSense
SixthSense' is a wearable gestural interface that aug-
ments the physical world around us with digital in-
formation and lets us use natural hand gestures to
interact with that information. The SixthSense pro-
totype is comprised of a pocket projector, a mirror
and a camera. The hardware components are coupled
in a pendant like mobile wearable device. Both the
projector and the camera are connected to the mo-
bile computing device in the users pocket. The pro-
jector projects visual information enabling surfaces,
walls and physical objects around us to be used as in-
terfaces; while the camera recognizes and tracks user's
hand gestures and physical objects using computer-
vision based techniques.