Your phone travels everywhere with you, potentially listening to every sound you make. These data
collectors could learn to understand you as well as your closest friend, and probably better.
The vast majority of people in developed countries now carry a smartphoneeverywhere. And while many
of us are already well aware of privacy issues associated with smartphones, such as their ability to
track our movements or even take surreptitious photos, an increasing number of people are starting
to worry that their smartphone is actually listening to everything they say.
There might not be much evidence for this but, it turns out, it isn’t far from the truth. Researchers
worldwide have begun developing many types of powerful audio analysis AI algorithms that can extract a
lot of information about us from sound alone. While this technology is only just beginning to emerge in
the real world, these growing capabilities – coupled with its 24/7 presence – could have serious
implications for our personal privacy.
Instead of analysing every word people say, much of the listening AI that has developed can actually learn
a staggering amount of personal information just from the sound of our speech alone. It can determine everything
from who you are and where you come from, your current location, your gender and age, and what language you’re
speaking – all just from the way your voice sounds when you speak.
If that isn’t creepy enough, other audio AI systems can detect if you’re lying, analyse your health and fitness
level, your current emotional state, and whether or not you’re intoxicated. There are even systems capable of
detecting what you’re eating when you speak with your mouth full, plus a slew of research looking into diagnosing
medical conditions from sound.
AI systems can also accurately interpret events from sound by listening to details like car crashes or gunshots,
or environments from their background noise. Other systems can identify a speaker’s attitude in a conversation,
pick up unspoken messages or detect conflicts between speakers. Another AI system developed last year can predict,
just by listening to the tone a couple used when speaking to each other, whether or not they will stay together.
These are all examples of current AI technology developed in research labs worldwide.
All of these technologies – no matter what they’re trying to learn about you – use machine learning. This involves
training an algorithm with large amounts of data that has been labelled to indicate what information the data contains.
By processing thousands or millions of recordings, the algorithm gradually begins to infer which characteristics of the
data – often just tiny fluctuations in the sound – are associated with which labels.
For example, a system used to detect your gender would record speech from your smartphone, and process it to extract
“features” – a small set of distinct values that compactly represent a bigger speech recording. Typically, features represent
amplitude and frequency information in each successive 20 millisecond period of speech. The way that these fluctuate
over time will be slightly different for male or female speech.
Machine learning systems will not only look at those features, but also how much, how often and in which way the features change
over time. While the recording happens in the smartphone itself, clips are sent to internet servers which will extract features,
compute their statistics and handle the machine learning part.AI was first created to perform conceptual tasks normally requiring
human intelligence. At the moment, most AI systems perform analysis and understanding tasks, which means they provide information
for humans to act on, rather than acting automatically.
For example, audio AI systems for road monitoring can alert traffic controllers to the sound of a vehicle crash, and audio-based
medical diagnosis AI would alert a doctor about findings of concern. But a human would still have to make a decision based on the
information provided to them by the AI.
But new AI technologies are changing. Many AI systems are starting to exceed human capabilities, with some devices even able to
act without human intervention. Amazon Echo and Google Home are both examples of AI that has thinking abilities. This kind of AI
can respond to commands directly and can also act on these commands, like when we ask Alexa to turn on the lights or draw our smart
curtains.While most AI systems today are designed to assist people, in the wrong hands, these technologies could look more like the
Thought Police from George Orwell’s Nineteen Eighty-Four. Audio (and video) surveillance can already detect our actions, but the AI
systems we have mentioned are starting to detect what is behind those actions – what we’re thinking, even if we never speak it aloud.
Most tech firms say their devices don’t record us unless we command them to, but there have been examples of Alexa making recordings
by mistake. And researchers have shown that it doesn’t take much to turn your phone into a permanent microphone. It may only be a
matter of time before advertisers and scammers start to use this technology to understand exactly how we think, and target our private
weaknesses.
Organisations like the World Privacy Forum, Fight for the Future and the Electronic Frontier Foundation are working to ensure people
have the right to privacy from digital sensing systems, or have the right to opt out from commercial surveillance. In the meantime,
when you next install an app or a game on your smartphone and it asks to access all sensors on your device, just remember what you
are potentially signing up to.These data collectors could learn to understand you as well as your closest friend and probably better, because your phone travels
everywhere with you, potentially listening to every sound you make. But while we can trust a true friend with our life, can we say
the same for those who are collecting our data today?