04-07-2012, 04:04 PM
human computer interaction (HCI)
human computer interaction.docx (Size: 636.52 KB / Downloads: 27)
INTRODUCTION
The world today is largely dependent on computers; therefore not being aware of the tricks of this trade is bound to make a person feel left out [1]. Computers changed the world a lot. It helped man step forward into the future. Thanks to computers, space exploration came true, new designs of vehicles and other transportation were made; entertainment became more entertaining, medical science made more cures for diseases, etc. The computers impacted our lives in many ways. They did make life a lot easier. Without computers, the world would be a harder place to live in. Thanks to the computers, everyday life is easier for us [2]. There are many people who find the standard computer input devices—the keyboard and mouse—difficult to use due to a motor disability. The mobility impairment prevents them from moving the mouse or typing on a keyboard. A number of keyboard and mouse configuration options designed to overcome physical difficulties exist so-called adaptive or assistive technology – hardware or software that eliminates barriers to using a computer [3]. The “mobility-impaired” people have trouble using the hardware of their computers rather than understanding or interpreting information [4].
Imagine yourself in a world where human Interact with computers. You are sitting in front of you personal computer that can listen, talk, or even scream aloud. It has the ability to gather information about you and interact with you through special techniques like facial recognition, speech recognition, etc. It can even understand your emotions at the touch of the mouse. It verifies your identity, feels your presents, and starts interacting with you .You ask the computer to dial to your friend at his office. It realizes the urgency of the situation through the mouse, dials your friend at his office, and establishes a connection. The BLUE EYES technology aims at creating computational machines that have perceptual and sensory ability like those of human beings. Employing most modern video cameras and microphones to identifies the users actions through the use of imparted sensory abilities. The machine can understand what a user wants, where he is looking at, and even realize his physical or emotional states.
THEORY
Based on Paul Ekman’s facial expression work, we see a correlation between a person’s emotional state and a person’s physiological measurements. Selected works from Ekman and others on measuring facial behaviors describe Elman’s Facial Action Coding System (Ekman and Rosenberg, 1997). One of his experiments involved participants attached to devices to record certain measurements including pulse, galvanic skin response (GSR), and temperature, somatic
movement and blood pressure. He then recorded the measurements as the participants were instructed to mimic facial expressions which corresponded to the six basic emotions. He defined the six basic emotions as anger, fear, sadness, disgust, joy and surprise. From this work, Dryer (1993) determined how physiological measures could be used to distinguish various emotional states. The measures taken were GSR, heart rate, skin temperature and general somatic activity (GSA). These data were then subject to two analyses. For the first analysis, a multidimensional scaling (MDS) procedure was used to determine the dimensionality of the data.
HUMAN – COMPUTER INTERACTION
EMOTION MOUSE
One goal of human computer interaction (HCI) is to make an adaptive, smart computer system. This type of project could possibly include gesture recognition, facial recognition, eye tracking, speech recognition, etc. Another noninvasive way to obtain information about a person is through touch. People use their computers to obtain, store and manipulate data using their computer. In order to start creating smart computers, the computer must start gaining information about the user. Our proposed method for gaining user information through touch is via a computer input device, the mouse. From the physiological data obtained from the user, an emotional state may be determined which would then be related to the task the user is currently doing on the computer. Over a period of time, a user model will be built in order to gain a sense of the user's personality. The scope of the project is to have the computer adapt to the user in order to create a better working environment where the user is more productive. The first steps towards realizing this goal are described here.
HUMAN – COMPUTER INTERACTION
Computer programming is an important activity occupying a large number of students in our society. Disabled people have reduced opportunities in a lot of areas, and the field of education is no exception. A virtual environment offers the possibility to control or operate in the real world partly alleviating physical limitations. Most people use their arms, hands, and fingers when using a computer, namely a keyboard and/or mouse. Nevertheless, for people with motor disabilities, the need for controlled movements can be an obstacle for them to be able to interact with the computer. There are many types of physical disabilities that affect a pleasant human-computer interaction. Several diseases cause muscle deterioration and/or neurological disorders that can result in physical weakness, loss of muscle control, paralysis, and amputation. The main problem for people with physical disabilities is the capacity to access computer controls (i.e. power switch, disk drives, monitor adjustments) and the ability to type on the standard keyboard or move pointing devices like a mouse [5]. With today's Graphic User Interface (GUI) - based PC software, most of the human-to-computer interaction is based on selection operations, such as;