Survey Paper Based On Hand Gesture Hex Color Matrix Vector: Sukrit Mehra, Prashant Verma, Harshit Bung, Deepak Bairagee
Survey Paper Based On Hand Gesture Hex Color Matrix Vector: Sukrit Mehra, Prashant Verma, Harshit Bung, Deepak Bairagee
Abstract
In order to enable a more natural communication with reality systems, automatic hand gesture recognition appears as a suitable means. Hand gesture recognition making use of digital images has been a research topic for many years. The aim of this paper is the proposal of real time vision system for its application within visual interaction environments through hand gesture recognition, using general purpose low cost software, so any user could make use of it in his office or home. The basis of our approach is a fast segmentation process to obtain the moving hand from the whole image, which is able to deal with a large number of hand shapes against different background and lighting conditions. The most important part of the recognition process is robust shape comparison carried out through Hidden Markov Model approach, which operates on edge maps. The visual memory use allows the system to handle variation within the gestures.
I. INTRODUCTION
Hand gesture recognition is an important research issue in the field of Human Interaction with Computer, because of its extensive applications in virtual reality, sign language recognition, and computer games. Despite lots of previous work, building a robust hand gesture recognition system that is applicable for real-life applications.Body language is an important way of communication among humans. Gesture recognition is the mathematical interpretation of a human motion by a computing device. Gesture recognition, along with facial recognition, voice recognition, eye tracking and lip movement recognition are components of what developers refer to as a perceptual user interface (PUI)[1]. The goal of PUI is to enhance the efficiency and ease of use for the underlying logical design of a stored program, a design discipline known as usability. In personal computing, gestures are most often used for input commands. Recognizing gestures as input allows computers to be more accessible for the physically-impaired and makes interaction more natural in a gaming or 3-D virtual environment.Body language is an important way of communication among humans, adding emphasis to voice messages or even being a complete message by itself. Thus, automatic posture recognition systems could be used for improving human-machine interaction. This kind of human-machine interfaces would allow a human user to control remotely through hand postures a wide variety of devices. Different applications have been suggested. Interactions between human and computer are currently[2] performed using keyboards, mice or different hectic devices. In addition to being different from our natural way of interacting. In gesture recognition technology, a camera reads the movements of the human body and communicates the data to a computer that uses the gestures as input to control devices or applications.
www.ijceronline.com
||April||2013||
Page 39
Multiple reflections In contrast to laser scanning systems, where only a single point is illuminated at once, the time-of-flight cameras illuminate a whole scene. Due to multiple reflections, the light may reach the objects along several paths and therefore, the measured distance may be greater than the true distance. c) Nave Bayes Classifier Limitation Assumption of class conditional independence usually does not hold Dependencies among these cannot be modeled by Nave Bayesian Classifier [6]. 2.2 Data Glove Based Limitation [1] Cost: Even though glove-based technology has come down in price (under $500 for the 5DT Glove), the cost of robust and complex posture and gesture recognition is going to be high if a glove-based solution is used. The cost of a tracking device and a robust glove is in the thousands of dollars. On the other hand, a vision-based solution is relatively inexpensive, especially since modern-day workstations are equipped with cameras. [2] User Comfort: With a glove-based solution, the user must wear a tracking device and glove that are connected to a computer. Putting these devices on takes time, can be quite cumbersome, and can limit ones range of motion. With a vision-based solution, the user may have to wear a glove, but the glove will be extremely lightweight, easy to put on, and not connected to the computer. Applications, in which no gloves are used, give the user complete freedom of motion and provides a cleaner way to interact and perform posture and gesture recognition. [3] Hand Size: Human hands vary in shape and size. This is a significant problem with glove-based solutions: some users cannot wear these input devices because their hands are too big or too small. [4] Hand Anatomy: Glove-based input devices may not always fit well enough to prevent their position sensors from moving relative to the joints the sensors are trying to measure. This problem reduces recognition accuracy after extended periods of use and forces users to recalibrate the devices which can be a nuisance. [5] Accuracy: In both vision- and glove-based solutions for hand posture and gesture recognition, accuracy is one of the most critical components to providing robust recognition. Both these solutions provide the potential for high levels of accuracy depending on the technology and recognition algorithms used. Accuracy also depends on the complexity and quantity of the postures and gestures to be recognized. Obviously, the quantity of possible postures and gestures and their complexity greatly affect accuracy no matter what raw data collection system is used. [6] Calibration: Calibration is important in both vision- and glove-based solutions but, due to the anatomy of the hand, it is more critical with glove-based solutions. In general, a calibration procedure or step is required for every user and, in some cases, every time a user wants to run the system.
www.ijceronline.com ||April||2013|| Page 40
Figure 1: Proposed System Framework The purposed technique is depending on the following approach: A pre image of hand is stored in the template. Capture the hand at initial point through frame. Match the captured hand with the pre hand stored in the template and marked or set the captured image as a reference image. Take the video frame of hand through webcam and convert the YUV color format to RGB color format. The RGB formatted image is saved in the form of matrix. Taking continues frame and match with the RGB matrix, the number of changes in the matrix is know and the movement of hand can be recognized. 3.1 Method Used: The Hidden Markov model is a stochastic process built on the top of another stochastic process, the Markov process. A time domain process exhibits first-order Markov property if the conditional probability density of the current event, given all past and present events, depends only on the most recent events. In a HMM [7], the 3.2Using YUV Colorspace:YUV colorspace is a bit unusual. The Y component determines the brightness of the color (referred to as luminance or luma), while the U and V components determine the color itself (the chroma) [9]. Y ranges from 0 to 1 (or 0 to 255 in digital formats), while U and V range from -0.5 to 0.5 (or -128 to 127 in signed digital form, or 0 to 255 in unsigned form). Some standards further limit the ranges so the outof-bounds values indicate special information like synchronization.
www.ijceronline.com ||April||2013|| Page 41
Figure2: YUV Image on Y-axis These images show us rods at various points on the UV plane, extending through Y. This allows us to see how each UV point's color is changed as the Y value is increased or decreased. 3.3 YUV RGB Conversion: There are many slightly different formulas to convert between YUV and RGB. The only major difference is a few decimal places. The CCIR 601 Standard [10] specifies the correct coefficients. Since I'm lazy and haven't looked up this spec, I don't know if the following coefficients are correct or not. In any event, I've used them for many conversions with no obvious discoloration. These formulas assume U and V are unsigned bytes. R = Y + 1.4075 * (V - 128) G = Y - 0.3455 * (U - 128) - (0.7169 * (V - 128)) B = Y + 1.7790 * (U - 128) Y = R * .299000 + G * .587000 + B * .114000 U = R * -.168736 + G * -.331264 + B * .500000 + 128 V = R * .500000 + G * -.418688 + B * -.081312 + 128 JPEG/JFIF - RGB Conversion: JPEG/JFIF files store compressed images in a YUV-like colorspace [11][12] that uses slightly different coefficients to convert to RGB. These formulas are: R = Y + 1.40200 * (U - 128) G = Y - 0.34414 * (V - 128) - 0.71414 * (U - 128) B = Y + 1.77200 * (V - 128)
IV. CONCLUSION
The mouse works on 120dpi and the frame captured through webcam is 30-25 fps. Using HMM with color hex method we would in increase the accuracy mouse movement interface and the process of the image in the form of vector i.e. in the form of matrix would give an 60% optimum result.
REFERENCES
[1] Zhou Ren, Junsong Yuan, Zhengyou Zhang, Robust Hand Gesture With Kinetic Sensor , Proceeding Of The 19th ACM International Conference On Multimedia, November 28-December 01, 2011 Scottsdale, Arizona USA, Pp. 759760. [2] Fukunaga, Keinosuke; Larry D. Hostetler (January 1975). "The Estimation Of The Gradient Of A Density Function, With Applications In Pattern Recognition". IEEE Transactions On Information Theory(IEEE) 21 (1): 3240. Doi:10.1109/TIT.1975.1055330. Retrieved 2008-02-29. VADAKKEPAT, P ET.AL, MULTIMODAL APPROACH TO HUMAN-FACE DETECTION AND TRACKING, I NDUSTRIAL ELECTRONICS, IEEE TRANSACTIONS ON ISSUE DATE: MARCH 2008, VOLUME: 55 ISSUE:3, PP.1385 1393 E. KOLLORZ, J. H ORNEGGER AND A. BARKE, GESTURE RECOGNITION WITH A T IME-OF-FLIGHT CAMERA, D YNAMIC 3D IMAGING, INTERNATIONAL JOURNAL OF I NTELLIGENT SYSTEMS TECHNOLOGIES AND APPLICATIONS ISSUE: VOLUME 5, N UMBER 3-4 2008, PP.334 343. BO HME, M., H AKER, M., MARTINETZ, T., AND BARTH, E. (2007) A FACIAL FEATURE TRACKER FOR HUMANCOMPUTER INTERACTION B ASED ON 3D TOF CAMERAS, D YNAMIC 3D IMAGING (WORKSHOP IN CONJUNCTION WITH DAGM 2007). PUJAN ZIAIE ET.AL, USING A NAVE BAYES CLASSIFIER BASED ON K-N EAREST NEIGHBORS WITH D ISTANCE WEIGHTING FOR STATIC HAND-GESTURE RECOGNITION I N A H UMAN-ROBOT D IALOG SYSTEM ADVANCES I N COMPUTER SCIENCE AND ENGINEERING C OMMUNICATIONS IN COMPUTER AND I NFORMATION SCIENCE, 2009, VOLUME 6, PART 1, PART 8, PP.308-315. A. MARKOV. AN EXAMPLE OF STATISTICAL INVESTIGATION I N THE TEXT OF EUGENE ONYEGIN, ILLUSTRATING COUPLING OF TESTS IN CHAINS. PROCEEDINGS OF THE ACADEMY OF SCIENCES OF ST. PETERSBURG, 1913 L. BAUM ET. AL. A MAXIMIZATION TECHNIQUE OCCURING IN THE STATISTICAL ANALYSIS OF PROBABLISTIC FUNCTIONS OF MARKOV CHAINS. ANNALS OF MATHEMATICAL STATISTICS, 41:164171, 1970. N. SOONTRANON, S. ARAMVITH, AND T.H. CHALIDABHONGSE. FACE AND H ANDS LOCALIZATION AND TRACKING FOR SIGN LANGUAGE RECOGNITION. INTERNATIONAL SYMPOSIUM O N COMMU-N ICATIONS AND I NFORMATION TECHNOLOGIES 2004 (ISCIT 2004), SAPPORO, JAPAN, OCTOBER26-29, 2004 U. AHLVERS, U. ZLZER, R. RAJAGOPALAN. MODEL-FREE FACE DETECTION AND HEAD TRACKING WITH MORPHOLOGICAL HOLE MAPPING, EUSIPCO 05, ANTALYA, TURKEY. S. ASKAR, Y. KONDRATYUK, K. ELAZOUZI, P. KAUFF, O. SCHEER. VISION-BASED SKIN-COLOUR SEGMENTATION OF MOVING HANDS FOR REAL-TIME APPLICATIONS. PROC. OF. 1ST EUROPEANCONFERENCE ON VISUAL MEDIA PRODUCTION, CVMP, LONDON, UNITED KINGDOM , 2004. A. H ADID, M. PIETIKAINEN, B. MARTINKAUPPI. COLOR -BASED FACE DETECTION USING SKIN LO- CUS MODEL AND H IERARCHICAL FILTERING. PROC. OF 16TH I NTERNATIONAL CONFERENCE ON PATTERNRECOGNITION 2002, VOL4, PP. 196-200, 2002 ||April||2013|| Page 42
[3] [4]
[5] [6]
[10] [11]
[12]
www.ijceronline.com