AI and Mass Spying

Spying and surveillance are different but related things. If I hired a private detective to spy on you, that detective could hide a bug in your home or car, tap your phone, and listen to what you said. At the end, I would get a report of all the conversations you had and the contents of those conversations. If I hired that same private detective to put you under surveillance, I would get a different report: where you went, whom you talked to, what you purchased, what you did.

Before the internet, putting someone under surveillance was expensive and time-consuming. You had to manually follow someone around, noting where they went, whom they talked to, what they purchased, what they did, and what they read. That world is forever gone. Our phones track our locations. Credit cards track our purchases. Apps track whom we talk to, and e-readers know what we read. Computers collect data about what we’re doing on them, and as both storage and processing have become cheaper, that data is increasingly saved and used. What was manual and individual has become bulk and mass. Surveillance has become the business model of the internet, and there’s no reasonable way for us to opt out of it.

Spying is another matter. It has long been possible to tap someone’s phone or put a bug in their home and/or car, but those things still require someone to listen to and make sense of the conversations. Yes, spyware companies like NSO Group help the government hack into people’s phones, but someone still has to sort through all the conversations. And governments like China could censor social media posts based on particular words or phrases, but that was coarse and easy to bypass. Spying is limited by the need for human labor.

AI is about to change that. Summarization is something a modern generative AI system does well. Give it an hourlong meeting, and it will return a one-page summary of what was said. Ask it to search through millions of conversations and organize them by topic, and it’ll do that. Want to know who is talking about what? It’ll tell you.

The technologies aren’t perfect; some of them are pretty primitive. They miss things that are important. They get other things wrong. But so do humans. And, unlike humans, AI tools can be replicated by the millions and are improving at astonishing rates. They’ll get better next year, and even better the year after that. We are about to enter the era of mass spying.

Mass surveillance fundamentally changed the nature of surveillance. Because all the data is saved, mass surveillance allows people to conduct surveillance backward in time, and without even knowing whom specifically you want to target. Tell me where this person was last year. List all the red sedans that drove down this road in the past month. List all of the people who purchased all the ingredients for a pressure cooker bomb in the past year. Find me all the pairs of phones that were moving toward each other, turned themselves off, then turned themselves on again an hour later while moving away from each other (a sign of a secret meeting).

Similarly, mass spying will change the nature of spying. All the data will be saved. It will all be searchable, and understandable, in bulk. Tell me who has talked about a particular topic in the past month, and how discussions about that topic have evolved. Person A did something; check if someone told them to do it. Find everyone who is plotting a crime, or spreading a rumor, or planning to attend a political protest.

There’s so much more. To uncover an organizational structure, look for someone who gives similar instructions to a group of people, then all the people they have relayed those instructions to. To find people’s confidants, look at whom they tell secrets to. You can track friendships and alliances as they form and break, in minute detail. In short, you can know everything about what everybody is talking about.

This spying is not limited to conversations on our phones or computers. Just as cameras everywhere fueled mass surveillance, microphones everywhere will fuel mass spying. Siri and Alexa and “Hey Google” are already always listening; the conversations just aren’t being saved yet.

Knowing that they are under constant surveillance changes how people behave. They conform. They self-censor, with the chilling effects that brings. Surveillance facilitates social control, and spying will only make this worse. Governments around the world already use mass surveillance; they will engage in mass spying as well.

Corporations will spy on people. Mass surveillance ushered in the era of personalized advertisements; mass spying will supercharge that industry. Information about what people are talking about, their moods, their secrets—it’s all catnip for marketers looking for an edge. The tech monopolies that are currently keeping us all under constant surveillance won’t be able to resist collecting and using all of that data.

In the early days of Gmail, Google talked about using people’s Gmail content to serve them personalized ads. The company stopped doing it, almost certainly because the keyword data it collected was so poor—and therefore not useful for marketing purposes. That will soon change. Maybe Google won’t be the first to spy on its users’ conversations, but once others start, they won’t be able to resist. Their true customers—their advertisers—will demand it.

We could limit this capability. We could prohibit mass spying. We could pass strong data-privacy rules. But we haven’t done anything to limit mass surveillance. Why would spying be any different?

This essay originally appeared in Slate.

Posted on December 5, 2023 at 7:10 AM31 Comments

Comments

Andy December 5, 2023 9:42 AM

Both governments and Big Internet Companies love the mass surveillance and mass spying models so they have no incentive to effective curtail it.

Dan Wolf December 5, 2023 10:12 AM

This is what many of us have been worried about for many years. We have several options that present themselves currently. We depend on Government and Big Tech to implement or embed ethical values into the A.I., not bloody likely. We restrict the use of A.I., depriving ourselves of the benefits which can be provided by A.I., again not likely. Or we develop and deploy means and methods of protecting individuals from collection & analysis of our personal data, and develop offensive techniques and applications to deal with overly intrusive A.I. In any case we face a paradigm shift in how we interact with our daily behaviors in this new society.

yet another bruce December 5, 2023 10:45 AM

Our own personal Jesus,
someone who hears you swear
someone who’s there

(apologies to Depeche Mode)

Time for service providers to all become data fiduciaries!

Al Sneed December 5, 2023 11:01 AM

very well-written, thank you. i’d like to emphasize that the data collected by companies is also resold to the government. in many cases, it might be illegal for the government to collect such data, but is legal to buy it.

i think we’re also about to re-authorize section 702.

A-A-Ron December 5, 2023 11:12 AM

Now what USA needs is a dictatorial-type government, extreme right or left does not matter, and they will have a lot more tools available than what Germany had in 1938.

R.E. December 5, 2023 11:20 AM

@AI Sneed
it might be illegal for the government to collect such data, but is legal to buy it

Also probably cheaper because tech companies already paid for the infrastructure and building the processes as part of “everyday operations”. Or like Google, for the sake of “advertising”. So the infrastructure used for surveillance is also part of a profit-making entity. As a result you can get the entire Data Analytical Services for some $6 million.

Felix December 5, 2023 11:41 AM

Constraining the institutions that have the resources and incentive to use mass spying takes leverage. Government is one source of leverage. But it’s pretty weak at representing people’s privacy interests.

We need other forms of collective representation. Consumers’ unions, for one, because mass spying and mass surveillance depend on consumer behaviors (buying and using networked computing devices).

postscript December 5, 2023 1:38 PM

Who is going to pay to store, back up and manage all that data and the ‘AI’ that will be grazing on it? I’m trying to imagine the Splunk fees.

TWT December 5, 2023 5:18 PM

@Bruce
Just as cameras everywhere fueled mass surveillance, microphones everywhere will fuel mass spying. Siri and Alexa and “Hey Google” are already always listening; the conversations just aren’t being saved yet.

This is where the real utility of 5G/6G/etc comes from. Not so much in making it possible for households to receive more data to various smart devices, but to send more data from those smart devices.

vas pup December 5, 2023 6:22 PM

The answer is ‘1984’ on steroids which we do already have today is profitable for corporations and loved by deep state bureaucracy: it is easy to pass many tasks to machine to avoid a lot of personal responsibility and put everybody in suspense state – Orwell had clear prediction even more than Nostradamus.
Goal is by both surveillance and spying make YOU pawn in THEIR games by manipulating your behavior for their profits and/or goals.

Clive Robinson December 5, 2023 7:27 PM

@ emily’s post, ALL,

Re : Remember doors alow passage both ways.

“Covert channels are going to become a thing.”

For those that think a little deeper there is Gus Simmon’s “Prisoner Problem” that gave rise to the issues of not being able to distinguish a covert channel, thus you have “subliminal channels”,

https://en.m.wikipedia.org/wiki/Subliminal_channel

Then consider a One Time Pad is not just secure against cryptanalysis via “perfect secrecy”.

If you do it right it gives you the first party –originator– in a communication deniability of message content if the second party –recipient– betrays to a third party –such as a government agency spy / agent– the KeyMat.

You simply need to generat from the ciphertext and innocuous message another key you claim was used.

As all messages are equiprobable, as are the keys all the messages are thus equiprobable.

All you have to do is,

1, Ensure secret message had no plaintext distinguishers.
2, Ensure secret message has no correlation to real life.

But if you take care to,

3, Ensure fake message has correlation to real life.
4, Give the fake message a plaintext distinguisher such as a simple “check sum”.

Then it makes the fakes you generate for deniability look real and the actual secret message look fake.

In the past I’ve also shown how using an OTP and code dictionary you can hide a secure message in simple plaintext you transmit.

Whilst the bandwidth is low you can use it to build subliminal authenticators.

What you need is redundancy, at a suitable level. For instance the opening salutation could be,

00, Hi
01, Hello
10, Dear colleague
11, Good day

Gives you two bits you can use for a covert / subliminal channel.

If you first encrypt the two bits of plaintext you want to send with two bits of OTP keytext you get two bits of ciphertext to use to select the salutation you put in the message you are “sending in the clear”.

The more “stock phrases” you have in your dictionary the more bits you can send, but the bandwidth is always going to be low, so the use you put them too needs to be important, like sending an authenticator.

An early use of a single bit was as a “Duress code” to let the recipient know you had been captured etc.

berburns December 5, 2023 10:39 PM

@ Mr. Peed Off,

I nominate smart vehicle license display for “Worst Idea of 2023”

Well, that person apparently brainstormed the pros and cons of things of “smart license plates” and decided to push an idea based entirely on the cons. Plus a bit of cluelessness. Having the plate display personal info for the police to read? They can already pull that up from a plate number, and could pull up renewal confirmation if we wanted to get rid of those old-timey stickers. (And having the text be small won’t protect against dashcams, which are common now.) Showing proof of insurance to keep cops from “harm’s way”? Is there some epidemic of cops dying while simply trying to confirm proper vehicle insurance? I think the drivers bear most of the risk, and couldn’t the DMV just require proof of insurance before renewing the plates anyway? I’m having trouble finding anything that could reasonably be called a benefit to the public.

But, in the spirit of George Constanza, doing the exact opposite of everything this person suggests might make smart plates be a good idea. Like showing less personal information. There’s no longer much need for a permanent identifier there. A QR code that changes every minute according to some cryptographic process could frustrate private tracking while still allowing authorized parties to look up the details. You’d have to take a picture to report someone, instead of trying to remember numbers, but that might be a reasonable trade-off for increased privacy.

Actually, isn’t the permanent identifier just about the only problem with license plates? I can’t think of another, excepting those botched Ontario plates of which one series peels horribly and its replacement isn’t readable in the dark. Kandel’s story certainly doesn’t mention any problem except for Kandel’s own boredom with this mostly-working century-old technology.

ces’t moi December 6, 2023 12:47 AM

@A-A-Ron, Aaron

Well the ex-Prez did supposedly re-post on Truth Social that he will make US of A to a dictatorship if elected

’https://www.washingtonpost.com/world/2023/12/04/trump-dictatorship-authoritarian-election-2024/

Chris Becke December 6, 2023 6:50 AM

@Andy

Both governments and Big Internet Companies love the mass surveillance and mass spying models so they have no incentive to effective curtail it.

I disagree. In so much as mass spying will have a chilling effect, Governments at least have an interest in protecting citizens from, at least, corporate mass spying, and downplaying their own mass spying capabilities.

People are lazy and careless. Governments that provide the illusion of protection from mass spying will doubtless have higher fidelity data collected by their secret programs that governments that simply compete with commercial spying.

Erdem Memisyazici December 6, 2023 10:55 AM

I wrote a lengthy and thoughtful comment based on my personal experience but it didn’t get posted so I’ll say the following.

I was drugged and interrogated by college kids who claimed to work security for the state University in town. I was locked inside a house, drugged and forced to talk for hours, an experience which gave me PTSD. I was asked about my political opinions about Edward Snowden as well as my life as a child in Turkey.

Afterwards I got all my privacy taken away from me because the same people believed sticking an implant in my jaw that works as conduction speakers would be a good idea and would make me forget about the event. I have been living with these speakers for almost a decade telling me to sit down and relax and make noise constantly and my life hasn’t been the same.

Had I not openly supported privacy and called Edward Snowden a hero my life would be different today.

Though one thing is for certain, I will not censor myself. You’d have to do that for me.

name.withheld.for.obvious.reasons December 6, 2023 3:00 PM

My concern is there are individuals and organizations seeking raw power for whom capture of such systems is beyond temptation. When the wishes of a stasi like automation comes true and meets with greed and control, a struggle will certainly insure.

ResearcherZero December 7, 2023 1:24 AM

Leverage.

Five Eyes partners focus on shared standards.

‘https://www.theregister.com/2023/12/07/australia_top_secret_cloud/

“metadata that reveals which apps a person has used, when they’ve received notifications, and the phone associated with a particular Google or Apple account.”

‘https://www.wired.com/story/apple-google-push-notification-surveillance/

“These companies should be permitted to generally reveal whether they have been compelled to facilitate this surveillance practice.”
https://www.wyden.senate.gov/news/press-releases/wyden-seeks-answers-from-justice-department-on-alleged-surveillance-of-apple-and-google-mobile-push-notifications

We will perhaps disclose such details now that this has been disclosed, says Apple.

‘https://www.reuters.com/technology/cybersecurity/governments-spying-apple-google-users-through-push-notifications-us-senator-2023-12-06/

Clive Robinson December 7, 2023 5:39 AM

@ ResearcherZero, ALL,

Re : Beware spook geeks bearing gifts.

“Five Eyes partners focus on shared standards.”

There is anold engineering joke about standards,

“Standards are like a toothbrush, every one agrees you should not just have but use them. But… nobody wants to use yours!”

There is a reason for this that engineers have learnt the hard way. Which is even the simplest of standards are necessarily incomplete thus edge and corner failure comes built in.

Thus two systems fully certified to the same standards won’t actually be compatible with each other. Hence “plug and play” getting called “Plug and Pray”.

For those who don’t get the issue and how much fun it can not be think back to the FTD driver debacle. FTD produced a popular USB to RS232 chip, that required a driver that Microsoft pushed out to all users. FTD got hit by the “grey market” of chips that were not theres but people thought they were. So literaly millions of USB devices were made with these chips all to often by quite reputable manufacturers (hit by “supply chain attacks”).

So FTD decided the loss of revenue was intolerable and found a way within their own “standard driver code” to “brick the grey chips” and pushed it out via Microsoft. The wailing was loud and immense FTD suffered more reputational damage than the grey market had ever caused them.

The point is though that there are always such edge and corner cases that can be exploited.

The other issue with standards is they cause a lack of “Hybrid vigor” and that kills fundemental inovation and also kills the domain.

To that end we have,

https://www.theregister.com/2023/12/07/memory_correction_five_eyes/

We’ve seen “We must all use the same programing language” before and when driven by a syandards view approach it causes problems… Any one here upto date on Ada – 2012?

Ada started life in the mid 1970’s and became a MIL-STD in the early 1980’s, and like the ghost of Christmas Past it still rattles on.

So it begs the question why are we seeing, every one should,

“Use Rust as a Must”…

It’s a case of,

“Every one should jump on the band waggon because the horse is pushing”

Any engineer can tell you why having a horse push rather than pull is a realy bad idea, if for no other reason it’s inherantly unstable and dangerous.

Speaking of which we have the UK Home Office, being pushed into unlawful acts by the Telco Industry and various UK Guard Labour entities,

https://www.theregister.com/2023/11/29/uk_sim_farm_ban/

Of course the Home Office rhetoric has nothing to do with Russia Attacking the Ukraine and using the “SIM Box” systems to provide communications in the Ukraine.

lurker December 7, 2023 7:28 PM

@Clive Robinson

The version I heard was:

Standards are like underwear, you should have plenty, and change them often.

Clive Robinson December 7, 2023 9:56 PM

@ lurker,

Re : The version I heard…

But you can not “wash rinse and repeate” with standards 😉

Dr No December 7, 2023 10:37 PM

@ Erdem Memisyazici – don’t worry as the implants have not been working for some time now due to a firmware update which went wrong and busted the sensitive electronics inside the implant.

ResearcherZero December 7, 2023 11:41 PM

@Clive Robinson

They are building the cloud system “very carefully”.

[I am not a tank, I am an ordinary sedan] (for future reference)

UK December 30, 2023 7:52 AM

…what was wrong, Mr. Schneier, with my text?
I just wondered about all the ad-systems uBlock blocked while reaching the page with the article on the other site.
Then I proposed to somehow make that much noise on the net, that the data for Mass Spying gets overwhelming, and I asked if somebody knows a tool which makes random links in the background on random sites (or perhaps a site list, so that everybody can choose his sites he wants the tool to connect to).
->this would produce that much data, that it would get more difficult to find the needle in the haystack.
Same procedure would I propose for CPUs. If it is idle it should do some random tasks to flush cashes by this and fill them with useless data to make side-channel attacks more difficult.

Leave a comment

Login

Allowed HTML <a href="URL"> • <em> <cite> <i> • <strong> <b> • <sub> <sup> • <ul> <ol> <li> • <blockquote> <pre> Markdown Extra syntax via https://michelf.ca/projects/php-markdown/extra/

Sidebar photo of Bruce Schneier by Joe MacInnis.