Project Soli uses radar to bring hand gestures to devices


The folks at Google ATAP are doing it again. ATAP is responsible for some of the crazier things you’ve seen from Google in the past. Things like Project Ara and Project Jacquard. ATAP shoots for the moon, and their latest shot includes interacting with devices through radar.

Project Soli uses radar to recognize gestures and fine hand movements. Radar is the technology you’ve probably seen in submarines on TV and for tracking weather patterns. ATAP built this tech into a chip so small that they can use it with smartwatches. In the video above, you can see the watch detecting hand movements from a few inches away. Gestures control the UI without even touching it.


Project Soli isn’t just for smartwatches. ATAP also partnered with JBL to make a speaker. You can snap your fingers to turn music on or off and more. It’s easy to see the potential for something like Project Soli. We’re surrounded by devices that use touchscreens and buttons. Both require physical touches to operate. Expanding a devices ability to 3D space opens it up to do a lot of cool things.

ATAP will release a DevKit sometime next year for Project Soli. Is this something you would use in your everyday life?

[via Google]

Joe Fedewa
Ever since I flipped open my first phone I've been obsessed with the devices. I've dabbled in other platforms, but Android is where I feel most at home.

Official: Project Ara launching next year, is a little less modular than we hoped [VIDEO]

Previous article

Top 10 Android Stories of the Week (May 20th)

Next article

You may also like


Leave a reply

Your email address will not be published. Required fields are marked *