+ Monday July 22nd, 2019

At their ATAP event, which promised to ‘knock your socks off’, Google has announced a new way to interact with wearables, called Project Soli.

Based on a research paper called ‘The information capacity of the human motor system’ which was written in 1954 by Paul Fitts, and re-enforced by recent research by Ivan Pouprev, Project Soli sees the small screens on wearables as the perfect platform to introduce a gesture control system.

The system is actually using a radar based system at 10,000fps at 60GHz detecting user movement giving users potentially more controls than is currently allowed by a touch screen. The system uses Range Doppler, Micro Doppler, Spectogram and IQ to capture the motions which are fed into a machine learning algorithm which interpret gestures to interpret a range of inputs like swipes, or the gesture for turning a knob or dial to be interpreted into software.

The nature of the Google ATAP group, which is lean, and in there words ‘optimised for speed’ has allowed the group to go from proof of concept the size of a small shoe box to a microSD card sized sensor in just 10 months.
Project Soli Development

The final product hasn’t been announced as yet, but more information, including a dev kit will be seeded to developers later this year.

Daniel Tyson  


Dan is a die-hard Android fan. Some might even call him a lunatic. He's been an Android user since Android was a thing, and if there's a phone that's run Android, chances are he owns it (his Nexus collection is second-to-none) or has used it.

Dan's dedication to Ausdroid is without question, and he has represented us at some of the biggest international events in our industry including Google I/O, Mobile World Congress, CES and IFA.

Join the Ausdroid Conversation

Notify of

Check Also

Tinder avoids Google’s 30% cut on in-app purchases, risks getting booted from Google Play entirely

Those who’ve not developed apps may well be unaware that both Apple and Google take …