- cross-posted to:
- hackernews
- cross-posted to:
- hackernews
Nobody wants to use AI to bug our phones, or to build a sprawling nerve system to track our vitals, because our phones are already bugged. Everything we do on them is recorded a dozen times over, by our wireless carriers, by the websites we visit and the apps we use, by the vendors and ad networks those companies are sending their data to, and in the marketplaces that sell that data. We built the eyes of the Greco decades ago.
But that data has remained relatively secure—or maybe more precisely, its potential energy has remained relatively buried—largely because it’s tedious to work with. It’s messy; it’s scattered across different sources and in different formats; combining it together is a pain, and most of us are simply not interesting enough to investigate. Data analysts who work at shadowy government agencies have lives too, and they do not want to write 595-line SQL queries either.
But AI doesn’t mind. And that’s the boring danger of what happens next: Not of AI becoming a superintelligent Sherlock Holmes finding impossible patterns in its enormous mind palace, but of it being a million monkeys at a million typewriters, doing the grunt work no person wanted to do. Because when prying questions are a prompt away—rather than 24 hours of work away—who wouldn’t get tempted to pry?



I think the article fails to take several critical factors into consideration.
The complexity of dealing with such large amounts of information will keep increasing forever as the amount of information also grows
AI struggles with conflicting information and mistakes, which happen a lot especially when humans are involved, so eventually you will have lots of “garbage in garbage out” issues causing problems
The data one might be able to track will continuously be challenged or removed on legal/compliance bases over time, reducing its availability
For example: Yes the NSA might want our chatbot logs, but after enough people realize they might be/are getting them, people will stop feeding it as much, or introduce noise on purpose. It’s not a perfect vacuum of constant reliable information forever. We are already seeing that AI models learning from web results are getting caught up in their own slop making themselves dumber. And the sheer volume of information relative to the computing power necessary to process everything will also become a problem if they keep trying to process every single thing.
I think this comment is based on an extremely optimistic – bordering on fantastical – outlook.
The capacity and capability to handle the data will grow too.
This is what data analysis is though. Extracting patterns from noisy data. Ignoring outliers. I don’t think anybody is suggesting they’ll just dump a CSV of your web history into ChatGPT and ask it if you’re probably going to a protest this weekend (although does it sound so far fetched that that might actually work?), it’ll be used in combination with existing and constantly improving data mining techniques.
Are you implying data protection laws will not only not be inexorably eroded year upon year by increasingly surveillance-hungry governments, but will actually get a significantly better than their current milquetoast state? I’ve gotta say, that’s seeming increasingly unlikely to me; right now we’re seeing mandatory identity verification being legislated on more and more things by more and more governments.
This has to be a sarcastic reference to Snowden, right? The thing where the entire world found out about the how NSA absolutely is – not “might be” – monitoring your internet and conversation logs, and basically nobody did a fucking thing to change? That was 12 years ago.
Good thing they’re not doing anything crazy to get more computing power, like buying up practically the entire global supply of RAM or building data centres at an exponentially increasing rate.