09-02-2010, 06:14 PM
Please can anyone provide full report on sixth sense technology!?
09-02-2010, 06:14 PM
Please can anyone provide full report on sixth sense technology!?
09-02-2010, 07:28 PM
TED (Technology, Entertainment, Design) is a working prototype of a multifunctional device that can become part of our lives in five years to ten. Set named sixth sense consists of a camera (it captures the movement of hands), the projector (it produces the image on any surface), the mobile phone (it is in your pocket and need only to communicate with the abstract database server) and four fingers on which to detect movements, wearing colorful caps, perceived by the camera
Sixth Sense Technology is a mini-projector coupled with a camera and a cellphoneâ€which acts as the computer and connected to the Cloud, all the information stored on the web. Sixth Sense can also obey hand gestures. The camera recognizes objects around a person instantly, with the micro-projector overlaying the information on any surface, including the object itself or hand. Also can access or manipulate the information using fingers. make a call by Extend hand on front of the projector and numbers will appear for to click. know the time by Draw a circle on wrist and a watch will appear. take a photo by Just make a square with fingers, highlighting what want to frame, and the system will make the photoâ€which can later organize with the others using own hands over the air.and The device has a huge number of applications , it is portable and easily to carry as can wear it in neck. The drawing application lets user draw on any surface by observing the movement of index finger. Mapping can also be done anywhere with the features of zooming in or zooming out. The camera also helps user to take pictures of the scene is viewing and later can arrange them on any surface. Some of the more practical uses are reading a newspaper. reading a newspaper and viewing videos instead of the photos in the paper. Or live sports updates while reading the newspaper. The device can also tell arrival, departure or delay time of air plane on tickets. For book lovers it is nothing less than a blessing. Open any book and find the Amazon ratings of the book. To add to it, pick any page and the device gives additional information on the text, comments and lot more add on feature read more http://gizmodo5167790/sixth-sense-techno...ld-forever http://theviewspapersixth-sense-technology-will-revolutionize-the-world/
14-02-2010, 10:11 PM
SixthSense: RFID-based Enterprise Intelligence
ABSTRACT: RFID is widely used to track the movement of goods through a supply chain.RFID platform is extended by presenting SixthSense here. SixthSense combines mobil- ity information obtained from RFID-based sensing with in- formation from enterprise systems such as calendar and pres- ence, to automatically draw inferences about the association and interaction amongst people, objects, and workspaces. INTRODUCTION: Radio Frequency Identification (RFID) is an electronic tagging technology that allows the detection and tracking of tags, and consequently the objects they are af-fixed to.SixthSense focuses on applying RFID to an enterprise setting, such as a corporate office or university department.In short SixthSense represents a form of mobile computing applied to non-computing entities. SIXTHSENSE ARCHITECTURE: Assumptions: SixthSense assumes an enterprise setting with widespread coverage of RFID readers and that people and objects are tagged with RFID. Users are free to pick up new tags and affix them to objects, as and when needed. Architectural Components: The key com- ponents of the system, including the databases, inference engine, and applications, are run centrally by the enter- prise.the various components of Sixth- Sense are: 1)RFID Monitor: It issues a read command every 500 ms to the RFID reader which then reports the EPC and the signal strength of the tags read via each of its antennas. Other Enterprise Monitors: 2)Calendar Monitor: monitors the time and location of the userâ„¢s appointments. 3)Presence Monitor: monitors the status of each userâ„¢s interaction with their desktop 4)Cameras:The camera feed is stored in a video database for future analysis, 5)Login Monitor: general login is a much stronger indica- tion of a user being present than simply a change in their machineâ„¢s idle. Raw database: The RFID monitor and the other enterprise monitors push data into the raw database. Processed database: It is populated by the inference en- gine with its inferences Inference engine: operates on the raw database us- ing the algorithms to draw inferences. Full report: http://www.stanford.edu/~piyushag/docs/mobisys08.pdf
22-03-2010, 10:39 PM
read http://istc711.wikispacesfile/view/Sixth..._twiss.ppt
for presentation of Sixth Sense Technology
12-04-2010, 11:54 AM
The basics¦
Sixth Sense allows people to use the internet without a screen or a keyboard It can turn any surface into a touch screen It is made with store-bought components The software program processes the video stream data captured by the camera and tracks the locations of the colored markers [that you wear on your fingers] using simple computer-vision techniques (http://www.pranavmistryprojects/sixthsen....htm#ABOUT) According to a number of sources, graduate students at MIT have been working on this for the past four months day and night Just a few things it can do¦ The device recognizes items in stores, and then gathers and projects information about products to the user. It also is able to provide quick signals to let users know what the best products are based on their needs.  Another amazing feature is the device™s ability to look at an airplane ticket and let the user know whether the flight is delayed or on time. It is also able to recognize books (based on an image or RFID) in a book store and then project reader reviews or author information from the Internet onto blank pages. The device can also recognize articles in newspapers, retrieve the latest related stories or video from the Internet and then display them on pages for the user. Say goodbye to your dictionary and encyclopedia. http://gadgets.tmcnettopics/gadgets/arti...ns-any.htm Cost of Sixth Sense $350 dollars The interesting thing here is that Sixth Sense can never really be more expensive than its individual components Purpose of Sixth Sense I believe Sixth Sense™s purpose would be to allow K-12 students and teachers to interact with information on a level not experienced before In my opinion, at least initially, its purpose would be to motivate¦motivate¦motivate Educational Use Engagement Research Collaboration Cost-effectiveness Engagement Teachers may find that Sixth Sense, because of its unique ability to work with many of Gardner™s multiple intelligences, will be a catalyst for students who might otherwise remain unengaged in learning. Sixth Sense uses the visual, kinesthetic, intrapersonal and interpersonal among others. If this technology does nothing else, the idea that students can simply use their hands to draw images on walls and take pictures with their hands will be (at least initially) quite motivating Research Sixth Sense will give new meaning to looking something up. Since we are essentially talking about the internet, students will be given the opportunity to portably research the world. They can all go on a virtual field trip and project the images on their own desks, for example. Collaboration Students will be given an opportunity through this technology to feel more connected to others. Although there does seem to be the sense that individuals could walk through life wired and lonely, I like to think the classroom would serve as the environment in which students and teachers using Sixth Sense would want to share. Not only could the sharing take place in the classroom, I suppose since the users are connected online, sharing could take place around the world. Cost-effectiveness As mentioned earlier, this product currently costs only $350. From an educator™s stand point, this is minimum given the amount of money spent on textbooks, computer labs and LCD projectors. For educator™s, Sixth Sense would be a way to put technology (literally) into each student™s hands, because of its size and its cost. Anticipated Time Frame According to a number of sources, the MIT students already have a patent on Sixth Sense. Still it is unclear as to when this device might actually become available for general purchasing. Microsoft Surface does currently exist (which is similar in theory). But Microsoft™s product is not portable like Sixth Sense Even if it came out next week, as educators we may need to wait some time in order to learn of the dangers that are sure to exist with the product. Advantages of Sixth Sense Portable Inexpensive Cool Multi-sensory Connectedness between the world and information Critical thinking skills would be somewhat vital here¦allowing teachers a fantastic opportunity for interdisciplinary study with critical thinking components Future Implications Students could become too wired¦and the element of collaboration I envision might be completely erased by how the individual interacts Availability of inappropriate information (how does this differ from current computer labs..this information is literally at students™ fingertips.) The technology will cost more because it is packaged and marketed¦and they can charge more What type of signal is needed..and might there be issues with getting signals in schools Getting Started My new thing is getting people to play with the technology No pressure Give them time Allow students and teachers to play with this together Offer support in the initial weeks/months of use Let Opinion Leaders emerge from their experience Start small! Relevant Issues from Rogers (2003) Opinion Leaders are key Consider the length of time of the innovation-decision period Think about homophily and heterophily when planning Sixth Sense expansion in a second and third school/classroom Realize that in most cases, individuals depend mainly on the communicated experience of others much like themselves who have already adopted a new idea (p. 331). My guinea pigs better be the right ones
01-05-2010, 11:05 PM
sixth sense FINAL ppt.pptx (Size: 3.06 MB / Downloads: 899) SIXTH SENSE TECHNOLOGY CONTENTS Introduction Components Working Applications INTRODUCTION Sixth Sense is a wearable gestural interface that augments the physical world around us with digital information and lets us use natural hand gestures to interact with that information. It was developed by Pranav Mistry, a PhD student in the Fluid Interfaces Group at the MIT Media Lab. SixthSense bridges the gap by bringing intangible, digital information out into the tangible world, and allowing us to interact with this information via natural hand SixthSense comprises a pocket projector, a mirror and a camera. The hardware components are coupled in a pendant like mobile wearable device COMPONENTS The hardware components are coupled in a pendant like mobile wearable device. Camera Projector Mirror Mobile Component Color Markers Camera Captures an object in view and tracks the user™s hand gestures It sends the data to smart phone It acts as a digital eye, connecting you to the world of digital information Projector The projector projects visual information enabling surfaces and physical objects to be used as interfaces The project itself contains a battery inside, with 3 hours of battery life. A tiny LED projector displays data sent from the smart phone on any surface in view“object, wall, or person. Mirror The usage of the mirror is significant as the projector dangles pointing downwards from the neck. Smart Phone A Web-enabled smart phone in the user™s pocket processes the video data Other software searches the Web and interprets the hand gestures Color Markers It is at the tip of the user™s fingers . Marking the user™s fingers with red, yellow, green, and blue tape helps the webcam recognize gestures The movements and arrangements of these makers are interpreted into gestures that act as interaction instructions for the projected application interfaces. HOW IT WORKS The hardware that makes Sixth Sense work is a pendant like mobile wearable interface It has a camera, a mirror and a projector and is connected wirelessly to a bluetooth smart phone that can slip comfortably into one™s pocket The camera recognizes individuals, images, pictures, gestures one makes with their hands Information is sent to the Smartphone for processing The downward-facing projector projects the output image on to the mirror Mirror reflects image on to the desired surface Thus, digital information is freed from its confines and placed in the physical world Make a call You can use the Sixth Sense to project a keypad onto your hand, then use that virtual keypad to make a call. Call up a map With the map application we can call up the map of our choice and then use thumbs and index fingers to navigate the map Check the time Draw a circle on your wrist to get a virtual watch that gives you the correct time Create multimedia reading experiences Sixth Sense can be programmed to project related videos onto newspaper articles you are reading Drawing application The drawing application lets the user draw on any surface by tracking the fingertip movements of the user™s index finger Zooming features The user can zoom in or zoom out using intuitive hand movements Get product information Sixth Sense uses image recognition or marker technology to recognize products we pick up, then feeds us information on those products Get book information The system can project Amazon ratings on that book, as well as reviews and other relevant information Get flight updates The system will recognize your boarding pass and let you know whether your flight is on time and if the gate has changed. Feed information on people The system will project relevant information about a person such as what they do, where they work, and so on. Take pictures If you fashion your index fingers and thumbs into a square ("framing" gesture), the system will snap a photo. After taking the desired number of photos, we can project them onto a surface, and use gestures to sort through the photos, and organize and resize them. CONCLUSION Sixth Sense recognizes the objects around us, displaying information automatically and letting us to access it in any way we need The Sixth Sense prototype implements several applications that demonstrate the usefulness, viability and flexibility of the system Allowing us to interact with this information via natural hand gestures The potential of becoming the ultimate "transparent" user interface for accessing information about everything around us hey please read https://seminarproject.net/Thread-sixth-...logy--7186 and https://seminarproject.net/Thread-sixth-sense-computer for getting more information about sixth sense technology
01-05-2010, 11:21 PM
SixthSense (Report).doc (Size: 2.16 MB / Downloads: 1,064) Chapter 1 INTRODUCTION We™ve evolved over millions of years to sense the world around us. When we encounter something, someone or some place, we use our five natural senses which includes eye, ear, nose, tongue mind and body to perceive information about it; that information helps us make decisions and chose the right actions to take. But arguably the most useful information that can help us make the right decision is not naturally perceivable with our five senses, namely the data, information and knowledge that mankind has accumulated about everything and which is increasingly all available online. Although the miniaturization of computing devices allows us to carry computers in our pockets, keeping us continually connected to the digital world, there is no link between our digital devices and our interactions with the physical world. Information is confined traditionally on paper or digitally on a screen. SixthSense bridges this gap, bringing intangible, digital information out into the tangible world, and allowing us to interact with this information via natural hand gestures. ˜SixthSense™ frees information from its confines by seamlessly integrating it with reality, and thus making the entire world your computer. Sixth Sense Technology, it is the newest jargon that has proclaimed its presence in the technical arena. This technology has emerged, which has its relation to the power of these six senses. Our ordinary computers will soon be able to sense the different feelings accumulated in the surroundings and it is all a gift of the Sixth Sense Technology newly introduced. SixthSense is a wearable gesture based device that augments the physical world with digital information and lets people use natural hand gestures to interact with that information. It was developed by Pranav Mistry, a PhD student in the Fluid Interfaces Group at the MIT Media Lab. A grad student with the Fluid Interfaces Group at MIT, he caused a storm with his creation of SixthSense. He says that the movies Robocop and Minority Report gave him the inspiration to create his view of a world not dominated by computers, digital information and human robots, but one where computers and other digital devices enhance people™s enjoyment of the physical world. Right now, we use our devices (computers, mobile phones, tablets, etc.) to go into the internet and get information that we want. With SixthSense we will use a device no bigger than current cell phones and probably eventually as small as a button on our shirts to bring the internet to us in order to interact with our world! SixthSense will allow us to interact with our world like never before. We can get information on anything we want from anywhere within a few moments! We will not only be able to interact with things on a whole new level but also with people! One great part of the device is its ability to scan objects or even people and project out information regarding what you are looking at. Chapter 2 SIXTH SENSE TECHNOLOGY 2.1 What is SixthSense Figure 2.1: Six Senses Sixth Sense in scientific (or non-scientific) terms is defined as Extra Sensory Perception or in short ESP. It involves the reception of information not gained through any of the five senses. Nor is it taken from any experiences from the past or known. Sixth Sense aims to more seamlessly integrate online information and tech into everyday life. By making available information needed for decision-making beyond what we have access to with our five senses, it effectively gives users a sixth sense. 2.2 Earlier SixthSense Prototype Figure 2.2: Earlier Device Maes™ MIT group, which includes seven graduate students, were thinking about how a person could be more integrated into the world around them and access information without having to do something like take out a phone. They initially produced a wristband that would read an Radio Frequency Identification tag to know, for example, which book a user is holding in a store. They also had a ring that used infrared to communicate by beacon to supermarket smart shelves to give you information about products. As we grab a package of macaroni, the ring would glow red or green to tell us if the product was organic or free of peanut traces †whatever criteria we program into the system. They wanted to make information more useful to people in real time with minimal effort in a way that doesn™t require any behaviour changes. The wristband was getting close, but we still had to take out our cell phone to look at the information. That™s when they struck on the idea of accessing information from the internet and projecting it. So someone wearing the wristband could pick up a paperback in the bookstore and immediately call up reviews about the book, projecting them onto a surface in the store or doing a keyword search through the book by accessing digitized pages on Amazon or Google books. They started with a larger projector that was mounted on a helmet. But that proved cumbersome if someone was projecting data onto a wall then turned to speak to friend †the data would project on the friend™s face. 2.3 Recent Prototype Figure 2.3: Present Device Now they have switched to a smaller projector and created the pendant prototype to be worn around the neck. The SixthSense prototype is composed of a pocket projector, a mirror and a camera. The hardware components are coupled in a pendant-like mobile wearable device. Both the projector and the camera are connected to the mobile computing device in the user™s pocket. We can very well consider the Sixth Sense Technology as a blend of the computer and the cell phone. It works as the device associated to it is hanged around the neck of a person and thus the projection starts by means of the micro projector attached to the device. Therefore, in course, you turn out to be a moving computer in yourself and the fingers act like a mouse and a keyboard. The prototype was built from an ordinary webcam and a battery-powered 3M projector, with an attached mirror †all connected to an internet-enabled mobile phone. The setup, which costs less than $350, allows the user to project information from the phone onto any surface †walls, the body of another person or even your hand. Mistry wore the device on a lanyard around his neck, and colored Magic Marker caps on four fingers (red, blue, green and yellow) helped the camera distinguish the four fingers and recognize his hand gestures with software that Mistry created. Chapter 3 WORKING OF SIXTH SENSE TECHNOLOGY 3.1 Components The hardware components are coupled in a pendant like mobile wearable device. Camera Projector Mirror Mobile Component Color Markers 3.1.1 Camera Figure 3.1: Camera A webcam captures and recognises an object in view and tracks the user™s hand gestures using computer-vision based techniques. It sends the data to the smart phone. The camera, in a sense, acts as a digital eye, seeing what the user sees. It also tracks the movements of the thumbs and index fingers of both of the user's hands. The camera recognizes objects around you instantly, with the micro-projector overlaying the information on any surface, including the object itself or your hand. 3.1.2 Projector Figure 3.2: Projector Also, a projector opens up interaction and sharing. The project itself contains a battery inside, with 3 hours of battery life. The projector projects visual information enabling surfaces, walls and physical objects around us to be used as interfaces. We want this thing to merge with the physical world in a real physical sense. You are touching that object and projecting info onto that object. The information will look like it is part of the object. A tiny LED projector displays data sent from the smart phone on any surface in view“object, wall, or person. 3.1.3 Mirror Figure 3.3: Mirror The usage of the mirror is significant as the projector dangles pointing downwards from the neck. 3.1.4 Mobile Component Figure 3.4: Smartphone The mobile devices like Smartphone in our pockets transmit and receive voice and data anywhere and to anyone via the mobile internet. An accompanying Smartphone runs the SixthSense software, and handles the connection to the internet. A Web-enabled smart phone in the user™s pocket processes the video data. Other software searches the Web and interprets the hand gestures. 3.1.5 Color Markers Figure 3.5: Color Markers It is at the tip of the user™s fingers. Marking the user™s fingers with red, yellow, green, and blue tape helps the webcam recognize gestures. The movements and arrangements of these makers are interpreted into gestures that act as interaction instructions for the projected application interfaces. 3.2 Working Figure 3.6: Working The hardware that makes Sixth Sense work is a pendant like mobile wearable interface It has a camera, a mirror and a projector and is connected wirelessly to a Bluetooth or 3G or wifi smart phone that can slip comfortably into one™s pocket The camera recognizes individuals, images, pictures, gestures one makes with their hands Information is sent to the Smartphone for processing The downward-facing projector projects the output image on to the mirror Mirror reflects image on to the desired surface Thus, digital information is freed from its confines and placed in the physical world The entire hardware apparatus is encompassed in a pendant-shaped mobile wearable device. Basically the camera recognises individuals, images, pictures, gestures one makes with their hands and the projector assists in projecting any information on whatever type of surface is present in front of the person. The usage of the mirror is significant as the projector dangles pointing downwards from the neck. To bring out variations on a much higher plane, in the demo video which was broadcasted to showcase the prototype to the world, Mistry uses coloured caps on his fingers so that it becomes simpler for the software to differentiate between the fingers, demanding various applications. The software program analyses the video data caught by the camera and also tracks down the locations of the coloured markers by utilising single computer vision techniques. One can have any number of hand gestures and movements as long as they are all reasonably identified and differentiated for the system to interpret it, preferably through unique and varied fiducials. This is possible only because the ˜Sixth Sense™ device supports multi-touch and multi-user interaction. MIT basically plans to augment reality with a pendant picoprojector: hold up an object at the store and the device blasts relevant information onto it (like environmental stats, for instance), which can be browsed and manipulated with hand gestures. The "sixth sense" in question is the internet, which naturally supplies the data, and that can be just about anything -- MIT has shown off the device projecting information about a person you meet at a party on that actual person (pictured), projecting flight status on a boarding pass, along with an entire non-contextual interface for reading email or making calls. It's pretty interesting technology that, like many MIT Media Lab projects, makes the wearer look like a complete dork -- if the projector doesn't give it away, the colored finger bands the device uses to detect finger motion certainly might. The idea is that SixthSense tries to determine not only what someone is interacting with, but also how he or she is interacting with it. The software searches the internet for information that is potentially relevant to that situation, and then the projector takes over. All the work is in the software," says Dr Maes. "The system is constantly trying to figure out what's around you, and what you're trying to do. It has to recognize the images you see, track your gestures, and then relate it all to relevant information at the same time." The software recognizes 3 kinds of gestures: Multitouch gestures, like the ones you see in Microsoft Surface or the iPhone -- where you touch the screen and make the map move by pinching and dragging. Freehand gestures, like when you take a picture [as in the photo above]. Or, you might have noticed in the demo, because of my culture, I do a namaste gesture to start the projection on the wall. Iconic gestures, drawing an icon in the air. Like, whenever I draw a star, show me the weather. When I draw a magnifying glass, show me the map. You might want to use other gestures that you use in everyday life. This system is very customizable. The technology is mainly based on hand gesture recognition, image capturing, processing, and manipulation, etc. The map application lets the user navigate a map displayed on a nearby surface using hand gestures, similar to gestures supported by multi-touch based systems, letting the user zoom in, zoom out or pan using intuitive hand movements. The drawing application lets the user draw on any surface by tracking the fingertip movements of the user™s index finger. Chapter 4 RELATED TECHNOLOGIES SixthSense™ technology takes a different approach to computing and tries to make the digital aspect of our lives more intuitive, interactive and, above all, more natural. We shouldn™t have to think about it separately. It™s a lot of complex technology squeezed into a simple portable device. When we bring in connectivity, we can get instant, relevant visual information projected on any object we pick up or interact with The technology is mainly based on hand augmented reality, gesture recognition, computer vision based algorithm etc. 4.1 Augmented reality Augmented reality (AR) is a term for a live direct or indirect view of a physical real-world environment whose elements are augmented by virtual computer-generated imagery. It is related to a more general concept called mediated reality in which a view of reality is modified (possibly even diminished rather than augmented) by a computer. The augmentation is conventionally in real-time and in semantic context with environmental elements. Sixth sense technology which uses Augmented Reality concept to super imposes digital information on the physical world. With the help of advanced AR technology (e.g. adding computer vision and object recognition) the information about the surrounding real world of the user becomes interactive and digitally usable. Artificial information about the environment and the objects in it can be stored and retrieved as an information layer on top of the real world view. The main hardware components for augmented reality are: display, tracking, input devices, and computer. Combination of powerful CPU, camera, accelerometers, GPS and solid state compass are often present in modern Smartphone, which make them prospective platforms. There are three major display techniques for Augmented Reality: Head Mounted Displays Handheld Displays Spatial Displays Head Mounted Displays A Head Mounted Display (HMD) places images of both the physical world and registered virtual graphical objects over the user's view of the world. The HMD's are either optical see-through or video see-through in nature. Handheld Displays Handheld Augment Reality employs a small computing device with a display that fits in a user's hand. All handheld AR solutions to date have employed video see-through techniques to overlay the graphical information to the physical world. Initially handheld AR employed sensors such as digital compasses and GPS units for its six degree of freedom tracking sensors. Spatial Displays Instead of the user wearing or carrying the display such as with head mounted displays or handheld devices; Spatial Augmented Reality (SAR) makes use of digital projectors to display graphical information onto physical objects. Modern mobile augmented reality systems use one or more of the following tracking technologies: digital cameras and/or other optical sensors, RFID, wireless sensors etc. Each of these technologies have different levels of accuracy and precision. Most important is the tracking of the pose and position of the user's head for the augmentation of the user's view. For users with disabilities of varying kinds, AR has real potential to help people with a variety of disabilities. Only some of the current and future AR applications make use of a Smartphone as a mobile computing platform. 4.2 Gesture Recognition Gesture recognition is a topic in computer science and language technology with the goal of interpreting human gestures via mathematical algorithms. Gestures can originate from any bodily motion or state but commonly originate from the face or hand. Current focuses in the field include emotion recognition from the face and hand gesture recognition. Many approaches have been made using cameras and computer vision algorithms to interpret sign language. Gestures can exist in isolation or involve external objects. Free of any object, we wave, beckon, fend off, and to a greater or lesser degree (depending on training) make use of more formal sign languages. With respect to objects, we have a broad range of gestures that are almost universal, including pointing at objects, touching or moving objects, changing object shape, activating objects such as controls, or handing objects to others. Gesture recognition can be seen as a way for computers to begin to understand human body language, thus building a richer bridge between machines and humans than primitive text user interfaces or even GUIs (graphical user interfaces), which still limit the majority of input to keyboard and mouse. Gesture recognition enables humans to interface with the machine (HMI) and interact naturally without any mechanical devices. Gestures can be used to communicate with a computer so we will be mostly concerned with empty handed semiotic gestures. These can further be categorized according to their functionality. Symbolic gestures These are gestures that, within each culture, have come to a single meaning. An Emblem such as the OK gesture is one such example, however American Sign Language gestures also fall into this category. Deictic gestures These are the types of gestures most generally seen in HCI and are the gestures of pointing, or otherwise directing the listeners attention to specific event or objects in the environment. Iconic gestures As the name suggests, these gestures are used to convey information about the size, shape or orientation of the object of discourse. They are the gestures made when someone says The plane flew like this, while moving their hand through the air like the flight path of the aircraft. Pantomimic gestures: These are the gestures typically used in showing the use of movement of some invisible tool or object in the speaker™s hand. When a speaker says I turned the steering wheel hard to the left, while mimicking the action of turning a wheel with both hands, they are making a pantomimic gesture. Using the concept of gesture recognition, it is possible to point a finger at the computer screen so that the cursor will move accordingly. This could potentially make conventional input devices such as mouse, keyboards and even touch-screens redundant. Gesture recognition can be conducted with techniques from computer vision and image processing. The literature includes ongoing work in the computer vision field on capturing gestures or more general human pose and movements by cameras connected to a computer. 4.3 Computer vision based algorithm Computer vision is the science and technology of machines that see. As a scientific discipline, computer vision is concerned with the theory behind artificial systems that extract information from images. The image data can take many forms, such as video sequences, views from multiple cameras, or multi-dimensional data from a medical scanner. Computer vision, on the other hand, studies and describes the processes implemented in software and hardware behind artificial vision systems. The software tracks the user™s gestures using computer-vision based algorithms. Computer vision is, in some ways, the inverse of computer graphics. While computer graphics produces image data from 3D models, computer vision often produces 3D models from image data. There is also a trend towards a combination of the two disciplines, e.g., as explored in augmented reality. The fields most closely related to computer vision are image processing, image analysis and machine vision. Image processing and image analysis tend to focus on 2D images, how to transform one image to another. his characterization implies that image processing/analysis neither require assumptions nor produce interpretations about the image content. Computer vision tends to focus on the 3D scene projected onto one or several images, e.g., how to reconstruct structure or other information about the 3D scene from one or several images. Machine vision tends to focus on applications, mainly in manufacturing, e.g., vision based autonomous robots and systems for vision based inspection or measurement. The Recognition Algorithms The computer vision system for tracking and recognizing the hand postures that control the menus is based on a combination of multi-scale color feature detection, view based hierarchical hand models and particle filtering. The hand postures or states are represented in terms of hierarchies of multi-scale color image features at different scales, with qualitative inter-relations in terms of scale, position and orientation. In each image, detection of multiscale color features is performed. The hand postures are then simultaneously detected and tracked using particle filtering, with an extension of layered sampling referred to as hierarchical layered sampling. To improve the performance of the system, a prior on skin color is included in the particle filtering. 4.4 Technologies that uses Sixth Sense as Platform SixthSense technology takes a different approach to computing and tries to make the digital aspect of our lives more intuitive, interactive and, above all, more natural. When you bring in connectivity, you can get instant, relevant visual information projected on any object you pick up or interact with. So, pick up a box of cereal and your device will project whether it suits your preferences. Some of the technologies that uses this are Radio Frequency Identification, gesture gaming, washing machine. 4.4.1 Radio Frequency Identification SixthSense is a platform for Radio Frequency Identification based enterprise intelligence that combines Radio Frequency Identification events with information from other enterprise systems and sensors to automatically make inferences about people, objects, workspaces, and their interactions. Radio Frequency Identification is basically an electronic tagging technology that allows the detection and tracking of tags and consequently the objects that they are affixed to. This ability to do remote detection and tracking coupled with the low cost of passive tags has led to the widespread adoption of RFID in supply chains worldwide. Pranav Mistry, a researcher at the media lab of the Massachusetts Institute Technology, has developed a 'sixth sense' device “ a gadget worn on the wrist that can function as a 'touch screen' device for many modern applications. The gadget is capable of selecting a product either by image recognition or radio frequency identification (RFID) tags and project information, like an Amazon rating. The idea of SixthSense is to use Radio Frequency Identification technology in conjunction with a bunch of other enterprise systems such as the calendar system or online presence that can track user activity. Here, we consider an enterprise setting of the future where people (or rather their employee badges) and their personal objects such as books, laptops, and mobile phones are tagged with cheap, passive RFID tags, and there is good coverage of RFID readers in the workplace. SixthSense incorporates algorithms that start with a mass of undifferentiated tags and automatically infer a range of information based on an accumulation of observations. The technology is able to automatically differentiate between people tags and object tags, learn the identities of people, infer the ownership of objects by people, learn the nature of different zones in a workspace (e.g., private office versus conference room), and perform other such inferences. By combining information from these diverse sources, SixthSense records all tag-level events in a raw database. The inference algorithms consume these raw events to infer events at the level of people, objects, and workspace zones, which are then recorded in a separate processed database. Applications can either poll these databases by running SQL queries or set up triggers to be notified of specific events of interest. SixthSense infers when a user has interacted with an object, for example, when you pick up your mobile phone. It is a platform in that its programming model makes the inferences made automatically available to applications via a rich set of APIs. To demonstrate the capabilities of the platform, the researchers have prototyped a few applications using these APIs, including a misplaced object alert service, an enhanced calendar service, and rich annotation of video with physical events. 4.4.2 Sixth Sense Washing Machine Whirlpool AWOE 8758 White Washing Machine is a remarkable front loader that incorporates the unparalleled Sixth Sense technology. Whirlpool™s 2009 range of washing machines comes integrated with enhanced 6th sense technology that gives more optimisation of resources and also increased saving in terms of energy, water and time. Ideal washing machine for thorough washing that requires sixth sense to detect stubborn stains and adjust wash impact. It is a feature packed washing ally with Sixth Sense Technology and several customized programs to enhance the washing performance and dexterously assist you in heavy washing loads. The New Generation 6th Sense appliances from Whirlpool are helping to protect the environment and to reduce your energy bills. Whirlpool 6th Sense appliances are designed to be intelligent and energy efficient appliances that adapt their performance to better suit your needs. All Whirlpool appliances with intelligent 6th Sense technology work on three key principles; Sense, Adaption and Control, to ensure that they achieve optimal performance each and every time that they are used. Whirlpool 6th Sense washing machines can save you up to 50% less water, energy and time during the cycle. These intelligent machines sense the size of the load and adjust and control the cycle dependent on the load inside in order to optimise the use of water, energy and time. Some models also contain a detergent overdosing monitor to make sure that you do not use too much washing detergent. Tumble dryers use 6th Sense technology to minimise energy and time wastage by monitoring the humidity inside your laundry and adjusting the drying time accordingly. Chapter 5 APPLICATIONS The SixthSense prototype implements several applications that demonstrate the usefulness, viability and flexibility of the system. The SixthSense device has a huge number of applications. The following are few of the applications of Sixth Sense Technology. Make a call Call up a map Check the time Create multimedia reading experience Drawing application Zooming features Get product information Get book information Get flight updates Feed information on people Take pictures Check the email 5.1 Make a call Figure 5.1: Make a call You can use the Sixth Sense to project a keypad onto your hand, then use that virtual keypad to make a call. Calling a number also will not be a great task with the introduction of Sixth Sense Technology. No mobile device will be required, just type in the number with your palm acting as the virtual keypad. The keys will come up on the fingers. The fingers of the other hand will then be used to key in the number and call. 5.2 Call up a map Figure 5.2: Map The sixth sense also implements map which lets the user display the map on any physical surface and find his destination and he can use his thumbs and index fingers to navigate the map, for example, to zoom in and out and do other controls. 5.3 Check the time Figure 5.3: Wrist Watch Sixth Sense all we have to do is draw a circle on our wrist with our index finger to get a virtual watch that gives us the correct time. The computer tracks the red marker cap or piece of tape, recognizes the gesture, and instructs the projector to flash the image of a watch onto his wrist. 5.4 Create multimedia reading experiences Figure 5.4: Video in Newspaper The SixthSense system also augments physical objects the user is interacting with by projecting more information about these objects projected on them. For example, a newspaper can show live video news or dynamic information can be provided on a regular piece of paper. Thus a piece of paper turns into a video display. 5.5 Drawing application Figure 5.5: Drawing The drawing application lets the user draw on any surface by tracking the fingertip movements of the user™s index finger. 5.6 Zooming features Figure 5.6: Zoom in and Zoom out The user can zoom in or zoom out using intuitive hand movements. 5.7 Get product information Figure 5.7: Product information Maes says Sixth Sense uses image recognition or marker technology to recognize products you pick up, then feeds you information on those products. For example, if you're trying to shop "green" and are looking for paper towels with the least amount of bleach in them, the system will scan the product you pick up off the shelf and give you guidance on whether this product is a good choice for you. 5.8 Get book information Figure 5.8: Book information Maes says Sixth Sense uses image recognition or marker technology to recognize products you pick up, then feeds you information on books. The system can project Amazon ratings on that book, as well as reviews and other relevant information 5.9 Take pictures Figure 5.9: Take Pictures If we fashion our index fingers and thumbs into a square (the typical "framing" gesture), the system will snap a photo. After taking the desired number of photos, we can project them onto a surface, and use gestures to sort through the photos, and organize and resize them. 5.10 Get flight updates Figure 5.10: Flight updates The system will recognize your boarding pass and let you know whether your flight is on time and if the gate has changed. 5.11 Feed information on people Figure 5.11: Information on people Sixth Sense also is capable of "a more controversial use. When you go out and meet someone, projecting relevant information such as what they do, where they work, and also m it could display tags about the person floating on their shirt. It could be handy if it displayed their facebook relationship status so that you knew not to waste your time. Chapter 6 ADVANTAGES AND ENHANCEMENTS . 6.1 Advantages SixthSense is an user friendly interface which integrates digital information into the physical world and its objects, making the entire world your computer. SixthSense does not change human habits but causes computer and other machines to adapt to human needs. It uses hand gestures to interact with digital information. Supports multi-touch and multi-user interaction Data access directly from machine in real time It is an open source and cost effective and we can mind map the idea anywhere It is gesture-controlled wearable computing device that feeds our relevant information and turns any surface into an interactive display. It is portable and easy to carry as we can wear it in our neck. The device could be used by anyone without even a basic knowledge of a keyboard or mouse. There is no need to carry a camera anymore. If we are going for a holiday, then from now on wards it will be easy to capture photos by using mere fingers 6.2 Future Enhancements To get rid of color markers To incorporate camera and projector inside mobile computing device. Whenever we place pendant- style wearable device on table, it should allow us to use the table as multi touch user interface. Applying this technology in various interest like gaming, education systems etc. To have 3D gesture tracking. To make sixth sense work as fifth sense for disabled person. CONCLUSION The key here is that Sixth Sense recognizes the objects around you, displaying information automatically and letting you access it in any way you want, in the simplest way possible. Clearly, this has the potential of becoming the ultimate "transparent" user interface for accessing information about everything around us. If they can get rid of the colored finger caps and it ever goes beyond the initial development phase, that is. But as it is now, it may change the way we interact with the real world and truly give everyone complete awareness of the environment around us. REFERENCES 1. www.blendernationsixth-sense-technology/ 2. http://boingboing2009/11/12/sixth-sense-technolo.html 3. http://gizmodo5167790/sixth-sense-techno...ld-forever 4. http://theviewspapersixth-sense-technology-will-revolutionize-the-world/ 5. http://lucasrichter.wordpress2009/03/13/...ping-this/
23-06-2010, 06:36 PM
Presentation On Sixth Sense.pptx (Size: 2.68 MB / Downloads: 521) SIXTH SENSE WEARABLE GEASTURES (WEAR UR WORLD) INVENTED BY PRANAB MISTRY (MIT STUDENT) Presented By: Md shahid PRANAB MISTRY ¢ Pranav Mistry is the genius behind Sixth Sense ¢ A PhD student at MIT's Media Lab. Before his studies at MIT, he worked with Microsoft as a UX researcher. Mistry is passionate about integrating the digital informational experience with our real-world interactions. WEAR UR WORLD 'SixthSense' is a wearable gestural interface that augments the physical world around us with digital information Actually, it is named so because you can operate this with a wearable gestural , it is user friendly interface which links the physical world around us and uses the hand gestures the same way you use other things like reading newspaper, turning pages, etc and in the same way you can interact with them. Well, this device will bring us closer to reality and assists us in making right decisions by providing the relevant information, thereby, making the entire world a computer. APPLICATION OF WUW(WEAR UR WORLD) The SixthSense prototype implements several applications that demonstrate the usefulness, viability and flexibility of the system. The map application lets the user navigate a map displayed on a nearby surface using hand gestures, similar to gestures supported by Multi-Touch based systems, letting the user zoom in, zoom out or pan using intuitive hand movements. The drawing application lets the user draw on any surface by tracking the fingertip movements of the user™s index finger. SixthSense also recognizes user™s freehand gestures (postures). For example, the SixthSense system implements a gestural camera that takes photos of the scene the user is looking at by detecting the ˜framing™ gesture. The user can stop by any surface or wall and flick through the photos he/she has taken. SixthSense also lets the user draw icons or symbols in the air using the movement of the index finger and recognizes those symbols as interaction instructions. For example, drawing a magnifying glass symbol takes the user to the map application or drawing an ˜@™ symbol lets the user check his mail. The SixthSense system also augments physical objects the user is interacting with by projecting more information about these objects projected on them.
29-06-2010, 01:40 PM
SIXTH SENSE TECHNOLOGY DEVICE.pptx (Size: 540.88 KB / Downloads: 409) SEMINAR ON SIXTH SENSE TECHNOLOGY DEVICE PRESENTED BY: DIVYAKANT SAINI WHAT IS IT A WEARABLE GESTURE . DEVICE THAT ENHANCE REAL LIFE WITH DIGITAL INFORMATION. DEVICE THAT TURNS ANY SURFACE INTO A WEB INTERFACE AND ALLOWS THE USER TO PROJECT INFORMATION FROM THE PHONE ONTO ANY SURFACE †WALLS,NEWSPAPER. COMPONENTS USED Camera: A webcam captures an object in view and tracks the user's hand gestures. It sends the data to the smart phone. Colored Markers: Marking the user's fingers with red, yellow, green, and blue tape helps the webcam recognize gestures. Projector: A tiny LED projector displays data sent from the smart phone on any surface in view--object, wall, or person. Smart Phone: A Web-enabled smart phone in the user's pocket processes the video data, using vision algorithms to identify the object. Other software searches the Web and interprets the hand gestures. WORKING THE SOFTWARE RECOGNIZES 3 KINDS OF GESTURES: MULTITOUCH GESTURES, LIKE THE ONES WE SEE IN IPHONE -- WHERE YOU TOUCH THE SCREEN AND MAKE THE MAP MOVE BY PINCHING AND DRAGGING. FREEHAND GESTURES, LIKE WHEN WE TAKE A PICTURE. ICONIC GESTURES, DRAWING AN ICON IN THE AIR. LIKE, WHENEVER I DRAW A STAR, SHOW ME THE WEATHER. WHEN I DRAW A MAGNIFYING GLASS, SHOW ME THE MAP. THE GESTURES CAN BE AS SIMPLE AS USING HIS FINGERS AND THUMBS TO CREATE A PICTURE FRAME THAT TELLS THE CAMERA TO SNAP A PHOTO, WHICH IS SAVED TO HIS MOBILE PHONE. PROJECTS A CLOUD OF WORDS ON THE PERSON™S BODY TO PROVIDE MORE INFORMATION ABOUT HIM †HIS BLOG URL, THE NAME OF HIS COMPANY, HIS LIKES AND INTERESTS. THE CURRENT STATUS OF HIS FLIGHT AND GATE NUMBER HE™S RETRIEVED FROM THE FLIGHT-STATUS PAGE OF THE AIRLINE ONTO THE CARD. FOR CHECKING E-MAIL DRAW @ SYMBOL IN THE AIR WITH HIS FINGER. PROJECT A PHONE PAD ONTO HIS PALM AND DIAL A NUMBER WITHOUT REMOVING THE PHONE FROM HIS POCKET FUTURE SCOPES WORK IS IN PROGRESS FOR GESTURE-RECOGNITION ALGORITHMS THAT COULD ELIMINATE THE NEED FOR THE MARKERS. USE OF LASER PROJECTORS TO INCREASE THE BRIGHTNESS. SOLAR PANEL TO ENSURE LONG BACKUP. THANK YOU
22-07-2010, 06:50 PM
Here are two more powerpoint presentations of the topic:
28704893-Sixth-Sense-FINAL-Ppt.pptx (Size: 3.12 MB / Downloads: 448) Sixth-Sense-Technology.ppt (Size: 4.26 MB / Downloads: 589)
25-07-2010, 10:44 PM
Hey eventhough i downloaded the sixthsense(report).exe documentation, it can't be open, please
Hey eventhough i downloaded the sixthsense(report).exe documentation, it can't be open, please send the complete documentation to my mail:mohanraja07541[at]gmail.com. its urgent please send meee.
02-10-2010, 12:05 PM
sisth sense technology ppt.pptx (Size: 1.22 MB / Downloads: 218) Sixth Sense Technology SUPERVISOR: Dr. M.N.MISHRA PRESENTED BY: DIVYAKANT SAINI abstract Sixth Sense Technology is a mini-projector coupled with a camera and a cellphoneâ€which acts as the computer and connected to the Cloud, all the information stored on the web. Sixth Sense can also obey hand gestures. The camera recognizes objects around a person instantly, with the micro-projector overlaying the information on any surface, including the object itself or hand. Also can access or manipulate the information using fingers. make a call by Extend hand on front of the projector and numbers will appear for to click. know the time by Draw a circle on wrist and a watch will appear. take a photo by Just make a square with fingers, highlighting what want to frame, and the system will make the photoâ€which can later organize with the others using own hands over the air.and The device has a huge number of applications , it is portable and easily to carry as can wear it in neck. The drawing application lets user draw on any surface by observing the movement of index finger. Mapping can also be done anywhere with the features of zooming in or zooming out. The camera also helps user to take pictures of the scene is viewing and later can arrange them on any surface. Some of the more practical uses are reading a newspaper. reading a newspaper and viewing videos instead of the photos in the paper. Or live sports updates while reading the newspaper. The device can also tell arrival, departure or delay time of air plane on tickets. For book lovers it is nothing less than a blessing. Open any book and find the Amazon ratings of the book. To add to it, pick any page and the device gives additional information on the text, comments and lot more add on feature for more:- https://seminarproject.net/Thread-sixth-...logy--7993 https://seminarproject.net/Thread-sixth-...technology https://seminarproject.net/Thread-sixth-...logy--7446 http://www.seminarprojectstag/sixth-sens...nology-ppt
19-10-2010, 04:48 PM
Sixth_Sense_twiss.ppt (Size: 52.5 KB / Downloads: 401) SixthSense is a wearable gestural interface device that augments the physical world with digital information and lets pe (Size: 58.96 KB / Downloads: 100) Sixth Sense Codename WUW~ Wear Ur World “integrating information with the real world” The basics… Sixth Sense allows people to use the internet without a screen or a keyboard It can turn any surface into a touch screen It is made with store-bought components “The software program processes the video stream data captured by the camera and tracks the locations of the colored markers [that you wear on your fingers] using simple computer-vision techniques” (http://www.pranavmistryprojects/sixthsen....htm#ABOUT) According to a number of sources, graduate students at MIT have been working on this for the past four months “day and night” Just a few things it can do… “The device recognizes items in stores, and then gathers and projects information about products to the user. It also is able to provide quick signals to let users know what the best products are based on their needs.” “Another amazing feature is the device’s ability to look at an airplane ticket and let the user know whether the flight is delayed or on time. It is also able to recognize books (based on an image or RFID) in a book store and then project reader reviews or author information from the Internet onto blank pages.” “The device can also recognize articles in newspapers, retrieve the latest related stories or video from the Internet and then display them on pages for the user. Say goodbye to your dictionary and encyclopedia.”
20-10-2010, 01:37 PM
please send me abt Sixth Sense Technology
19-11-2010, 12:31 PM
Can u pls send me the full report and ppt of sixth sense technology as soon a spossible?
|
|