Virtual Gesture Screen System Based on 3D Visual Information and Multi-Layer Perceptron

Active research is underway on virtual touch screens that complement the physical limitations of conventional touch screens. This paper discusses a virtual touch screen that uses a multi-layer perceptron to recognize and control three-dimensional (3D) depth information from a time of flight (TOF) camera. This system extracts an object-s area from the image input and compares it with the trajectory of the object, which is learned in advance, to recognize gestures. The system enables the maneuvering of content in virtual space by utilizing human actions.




References:
[1] Kim Hyung-joon, " Virtual Touch Screen System for Game
Applications" , Journal of Korea Game Society, vol. 6, no. 3, pp77-86,
2006.
[2] Martin Tosas and Bai Li, Lecture Notes in Computer Science, Heidelberg,
Springer Berlin, pp48-59, 2004.
[3] Eunjin Koh, Jongho Won, and Changseok Bae,"Vision-based Virtual
Touch Screen Interface", Proceeding of ICCE 2008, LasVegas, USA,
2008.
[4] http://en.wikipedia.org/wiki/Time-of-flight
[5] Gokturk. S. B, Yalcin. H, and Bamji. C., "A Time-Of-Flight Depth Sensor
- System Description, Issues and Solutions", CVPRW '04, p35, 2004.
[6] P. Hajela, B. Fu and L. Berke, Neural networks in structural analysis and
design: an overview. Comput. Syst. Engng 3 1-4 (1992), pp. 525-538.
[7] R.P. Lippmann, An introduction to computing with neural nets. IEEE
Acoust. Speech Signal Process 4 2 (1987), pp. 4-22.