The thrilling potential of SixthSense technology | Pranav Mistry

Estimated read time: 1:20

    Learn to use AI like a Pro

    Get the latest AI workflows to boost your productivity and business performance, delivered weekly by expert consultants. Enjoy step-by-step guides, weekly Q&A sessions, and full access to our AI workflow archive.

    Canva Logo
    Claude AI Logo
    Google Gemini Logo
    HeyGen Logo
    Hugging Face Logo
    Microsoft Logo
    OpenAI Logo
    Zapier Logo
    Canva Logo
    Claude AI Logo
    Google Gemini Logo
    HeyGen Logo
    Hugging Face Logo
    Microsoft Logo
    OpenAI Logo
    Zapier Logo

    Summary

    Pranav Mistry explores the fascinating integration of the physical and digital worlds through his SixthSense technology. By leveraging everyday gestures and objects, he seeks to create interfaces that are more intuitive and seamless. His journey began with experiments using mouse rollers and sticky notes, leading to innovations like a 3D drawing pen and a head-mounted projector. Ultimately, Mistry developed the SixthSense device, which allows interaction with digital information via gestures in real-world contexts. This technology promises to bridge the digital divide, enabling richer, more human connections.

      Highlights

      • Pranav Mistry's innovation journey started with curious experiments like dismantling a mouse. 🔍
      • SixthSense technology enables digital interactions using everyday gestures and objects. 🖐️
      • Head-mounted projectors and cameras allow digital augmentation in real-world settings. 📽️
      • The technology offers new ways to interact with data without traditional devices. 💡
      • Mistry advocates for making technology accessible to everyone with open-source plans. 🌎

      Key Takeaways

      • SixthSense technology seamlessly blends the physical and digital worlds. 🌍
      • Pranav Mistry's inventions are inspired by everyday gestures and objects. 🤲
      • The technology uses intuitive gestures to interact with digital data. ✨
      • SixthSense aims to bridge the digital divide and enhance human interaction. 🤝
      • Open-source plans will allow anyone to create their own SixthSense device. 🌐

      Overview

      Pranav Mistry's journey into the world of SixthSense technology is a thrilling tale of curiosity and innovation. Starting with simple experiments like deconstructing a mouse, Mistry was driven by the need to merge the digital with the physical world in intuitive ways. His work aims to make technology more engaging and accessible through the use of everyday gestures and objects.

        The SixthSense device, developed by Mistry, is an exciting leap forward in how we interact with digital information. Instead of being confined to screens, digital data can now be accessed and manipulated using simple gestures in real-world environments. Whether it's drawing in 3D or checking flight delays with a boarding pass, SixthSense opens up new possibilities for intuitive technology use.

          Mistry's vision extends beyond just creating innovative technology. He is committed to making these advancements accessible to all, with plans to release open-source instructions for building a SixthSense device. His work not only bridges the gap between the digital and physical but also paves the way for technology to enhance our human experience, keeping us connected with the world around us.

            Chapters

            • 00:00 - 00:30: Introduction The introduction chapter discusses how humans have grown up interacting with numerous physical objects daily, highlighting their abundance and the enjoyment they provide compared to typical computing devices.
            • 00:30 - 01:40: Gestures and Physical Interactions This chapter explores the concept of gestures and their significance in interacting with both objects and people in everyday life. It highlights how certain gestures are universally understood, such as 'Namaste' for respect or a cricket gesture for scoring, and how these become a part of our learning and communication.
            • 01:40 - 04:30: Early Explorations and Experiments The chapter titled 'Early Explorations and Experiments' discusses the exploration of leveraging everyday objects and interactions with the physical world to interact with the digital world. It questions why traditional input devices like keyboards and mice must be used when there are more intuitive ways based on physical interactions. The exploration, which began about eight years ago, started with examining a mouse, not for its usual computer-related functions, but by physically opening it to understand its workings.
            • 04:30 - 10:00: Developing the SixthSense Device This chapter discusses the early development of the SixthSense device by exploring how the mouse rollers were repurposed for the project. The author was intrigued by the mechanics of old computer mice, which used a ball and rollers to detect movement. By borrowing additional mice and extracting the rollers, the author explored using these components to innovate and create something new. This creativity marked a step in developing the SixthSense, showcasing an inventive and resourceful approach.
            • 10:00 - 18:20: Demonstrations and Applications of SixthSense Pranav Mistry discusses a gesture-interface device he developed, which acts as a motion-sensing device created for just two dollars. This device allows physical movements to be replicated in the digital world. Mistry's interest in integrating physical and digital spaces led to this innovation, conceptualized around the year 2000. The chapter hints at the connection between physical and digital spaces through the idea of 'sticky notes.'
            • 18:20 - 23:00: Future Plans and Open Source Movement The chapter 'Future Plans and Open Source Movement' explores the concept of integrating physical sticky notes with the digital world. The discussion includes the ability for handwritten notes, such as messages or reminders, to automatically sync with digital systems like SMS or digital calendars. The chapter also speculates on future technologies where queries written on paper could interact seamlessly with digital databases, potentially serving as a paper input-output system.
            • 23:00 - 25:15: Closing Remarks and Recognition The concluding chapter discusses innovative ideas and inventions. It describes the development of a three-dimensional drawing pen intended for designers and architects to visualize and interact with their designs in a more intuitive manner. Additionally, it touches upon the concept of a physical Google Map, where physical objects placed on it can offer real-world information, such as showing a flight gate location when a boarding pass is placed on the map.

            The thrilling potential of SixthSense technology | Pranav Mistry Transcription

            • 00:00 - 00:30 We grew up interacting with the physical objects around us. There are an enormous number of them that we use every day. Unlike most of our computing devices, these objects are much more fun to use.
            • 00:30 - 01:00 When you talk about objects, one other thing automatically comes attached to that thing, and that is gestures: how we manipulate these objects, how we use these objects in everyday life. We use gestures not only to interact with these objects, but we also use them to interact with each other. A gesture of "Namaste!", maybe, to respect someone, or maybe, in India I don't need to teach a kid that this means "four runs" in cricket. It comes as a part of our everyday learning. So, I am very interested, from the beginning,
            • 01:00 - 01:30 how our knowledge about everyday objects and gestures, and how we use these objects, can be leveraged to our interactions with the digital world. Rather than using a keyboard and mouse, why can I not use my computer in the same way that I interact in the physical world? So, I started this exploration around eight years back, and it literally started with a mouse on my desk. Rather than using it for my computer, I actually opened it.
            • 01:30 - 02:00 Most of you might be aware that, in those days, the mouse used to come with a ball inside, and there were two rollers that actually guide the computer where the ball is moving, and, accordingly, where the mouse is moving. So, I was interested in these two rollers, and I actually wanted more, so I borrowed another mouse from a friend -- never returned to him -- and I now had four rollers. Interestingly, what I did with these rollers is, basically, I took them off of these mouses and then put them in one line.
            • 02:00 - 02:30 It had some strings and pulleys and some springs. What I got is basically a gesture-interface device that actually acts as a motion-sensing device made for two dollars. So, here, whatever movement I do in my physical world is actually replicated inside the digital world just using this small device that I made, around eight years back, in 2000. Because I was interested in integrating these two worlds, I thought of sticky notes. I thought, "Why can I not connect
            • 02:30 - 03:00 the normal interface of a physical sticky note to the digital world?" A message written on a sticky note to my mom, on paper, can come to an SMS, or maybe a meeting reminder automatically syncs with my digital calendar -- a to-do list that automatically syncs with you. But you can also search in the digital world, or maybe you can write a query, saying, "What is Dr. Smith's address?" and this small system actually prints it out -- so it actually acts like a paper input-output system, just made out of paper.
            • 03:00 - 03:30 In another exploration, I thought of making a pen that can draw in three dimensions. So, I implemented this pen that can help designers and architects not only think in three dimensions, but they can actually draw, so that it's more intuitive to use that way. Then I thought, "Why not make a Google Map, but in the physical world?" Rather than typing a keyword to find something, I put my objects on top of it. If I put a boarding pass, it will show me where the flight gate is.
            • 03:30 - 04:00 A coffee cup will show where you can find more coffee, or where you can trash the cup. So, these were some of the earlier explorations I did because the goal was to connect these two worlds seamlessly. Among all these experiments, there was one thing in common: I was trying to bring a part of the physical world to the digital world. I was taking some part of the objects, or any of the intuitiveness of real life, and bringing them to the digital world,
            • 04:00 - 04:30 because the goal was to make our computing interfaces more intuitive. But then I realized that we humans are not actually interested in computing. What we are interested in is information. We want to know about things. We want to know about dynamic things going around. So I thought, around last year -- in the beginning of the last year -- I started thinking, "Why can I not take this approach in the reverse way?" Maybe, "How about I take my digital world and paint the physical world with that digital information?"
            • 04:30 - 05:00 Because pixels are actually, right now, confined in these rectangular devices that fit in our pockets. Why can I not remove this confine and take that to my everyday objects, everyday life so that I don't need to learn the new language for interacting with those pixels? So, in order to realize this dream, I actually thought of putting a big-size projector on my head. I think that's why this is called a head-mounted projector, isn't it? I took it very literally,
            • 05:00 - 05:30 and took my bike helmet, put a little cut over there so that the projector actually fits nicely. So now, what I can do -- I can augment the world around me with this digital information. But later, I realized that I actually wanted to interact with those digital pixels, also. So I put a small camera over there that acts as a digital eye. Later, we moved to a much better, consumer-oriented pendant version of that, that many of you now know as the SixthSense device. But the most interesting thing about this particular technology
            • 05:30 - 06:00 is that you can carry your digital world with you wherever you go. You can start using any surface, any wall around you, as an interface. The camera is actually tracking all your gestures. Whatever you're doing with your hands, it's understanding that gesture. And, actually, if you see, there are some color markers that in the beginning version we are using with it. You can start painting on any wall. You stop by a wall, and start painting on that wall. But we are not only tracking one finger, here.
            • 06:00 - 06:30 We are giving you the freedom of using all of both of your hands, so you can actually use both of your hands to zoom into or zoom out of a map just by pinching all present. The camera is actually doing -- just, getting all the images -- is doing the edge recognition and also the color recognition and so many other small algorithms are going on inside. So, technically, it's a little bit complex, but it gives you an output which is more intuitive to use, in some sense. But I'm more excited that you can actually take it outside. Rather than getting your camera out of your pocket,
            • 06:30 - 07:00 you can just do the gesture of taking a photo, and it takes a photo for you. (Applause) Thank you. And later I can find a wall, anywhere, and start browsing those photos or maybe, "OK, I want to modify this photo a little bit and send it as an email to a friend." So, we are looking for an era where computing will actually merge with the physical world. And, of course, if you don't have any surface, you can start using your palm for simple operations.
            • 07:00 - 07:30 Here, I'm dialing a phone number just using my hand. The camera is actually not only understanding your hand movements, but, interestingly, is also able to understand what objects you are holding in your hand. For example, in this case, the book cover is matched with so many thousands, or maybe millions of books online, and checking out which book it is. Once it has that information, it finds out more reviews about that, or maybe New York Times has a sound overview on that,
            • 07:30 - 08:00 so you can actually hear, on a physical book, a review as sound. (Video) Famous talk at Harvard University -- This was Obama's visit last week to MIT. (Video) And particularly I want to thank two outstanding MIT -- Pranav Mistry: So, I was seeing the live [video] of his talk, outside, on just a newspaper. Your newspaper will show you live weather information rather than having it updated. You have to check your computer in order to do that, right? (Applause)
            • 08:00 - 08:30 When I'm going back, I can just use my boarding pass to check how much my flight has been delayed, because at that particular time, I'm not feeling like opening my iPhone, and checking out a particular icon. And I think this technology will not only change the way -- (Laughter) Yes. It will change the way we interact with people, also, not only the physical world. The fun part is, I'm going to the Boston metro, and playing a pong game inside the train on the ground, right?
            • 08:30 - 09:00 (Laughter) And I think the imagination is the only limit of what you can think of when this kind of technology merges with real life. But many of you argue, actually, that all of our work is not only about physical objects. We actually do lots of accounting and paper editing and all those kinds of things; what about that? And many of you are excited about the next-generation tablet computers to come out in the market. So, rather than waiting for that, I actually made my own, just using a piece of paper.
            • 09:00 - 09:30 So, what I did here is remove the camera -- All the webcam cameras have a microphone inside the camera. I removed the microphone from that, and then just pinched that -- like I just made a clip out of the microphone -- and clipped that to a piece of paper, any paper that you found around. So now the sound of the touch is getting me when exactly I'm touching the paper. But the camera is actually tracking where my fingers are moving. You can of course watch movies.
            • 09:30 - 10:00 (Video) Good afternoon. My name is Russell, and I am a Wilderness Explorer in Tribe 54." PM: And you can of course play games. (Car engine) Here, the camera is actually understanding how you're holding the paper and playing a car-racing game. (Applause) Many of you already must have thought, OK, you can browse. Yeah. Of course you can browse to any websites or you can do all sorts of computing on a piece of paper
            • 10:00 - 10:30 wherever you need it. So, more interestingly, I'm interested in how we can take that in a more dynamic way. When I come back to my desk, I can just pinch that information back to my desktop so I can use my full-size computer. (Applause) And why only computers? We can just play with papers. Paper world is interesting to play with. Here, I'm taking a part of a document, and putting over here a second part from a second place, and I'm actually modifying the information that I have over there.
            • 10:30 - 11:00 Yeah. And I say, "OK, this looks nice, let me print it out, that thing." So I now have a print-out of that thing. So the workflow is more intuitive, the way we used to do it maybe 20 years back, rather than now switching between these two worlds. So, as a last thought, I think that integrating information to everyday objects will not only help us to get rid of the digital divide,
            • 11:00 - 11:30 the gap between these two worlds, but will also help us, in some way, to stay human, to be more connected to our physical world. And it will actually help us not end up being machines sitting in front of other machines. That's all. Thank you. (Applause)
            • 11:30 - 12:00 Thank you. (Applause) Chris Anderson: So, Pranav, first of all, you're a genius. This is incredible, really. What are you doing with this? Is there a company being planned? Or is this research forever, or what? Pranav Mistry: So, there are lots of companies, sponsor companies of Media Lab interested in taking this ahead in one or another way. Companies like mobile-phone operators want to take this in a different way than the NGOs in India,
            • 12:00 - 12:30 thinking, "Why can we only have 'Sixth Sense'? We should have a 'Fifth Sense' for missing-sense people who cannot speak. This technology can be used for them to speak out in a different way maybe a speaker system." CA: What are your own plans? Are you staying at MIT, or are you going to do something with this? PM: I'm trying to make this more available to people so that anyone can develop their own SixthSense device, because the hardware is actually not that hard to manufacture or hard to make your own. We will provide all the open source software for them,
            • 12:30 - 13:00 maybe starting next month. CA: Open source? Wow. (Applause) CA: Are you going to come back to India with some of this, at some point? PM: Yeah. Yes, yes, of course. CA: What are your plans? MIT? India? How are you going to split your time going forward? PM: There is a lot of energy here. Lots of learning. All of this work that you have seen is all about my learning in India. And now, if you see, it's more about the cost-effectiveness: this system costs you $300
            • 13:00 - 13:30 compared to the $20,000 surface tables, or anything like that. Or maybe even the $2 mouse gesture system at that time was costing around $5,000? I showed that, at a conference, to President Abdul Kalam, at that time, and then he said, "OK, we should use this in Bhabha Atomic Research Centre for some use of that." So I'm excited about how I can bring the technology to the masses rather than just keeping that technology in the lab environment. (Applause)
            • 13:30 - 14:00 CA: Based on the people we've seen at TED, I would say you're truly one of the two or three best inventors in the world right now. It's an honor to have you at TED. Thank you so much. That's fantastic. (Applause)