I think the article fails to take several critical factors into consideration.
-
The complexity of dealing with such large amounts of information will keep increasing forever as the amount of information also grows
-
AI struggles with conflicting information and mistakes, which happen a lot especially when humans are involved, so eventually you will have lots of "garbage in garbage out" issues causing problems
-
The data one might be able to track will continuously be challenged or removed on legal/compliance bases over time, reducing its availability
For example: Yes the NSA might want our chatbot logs, but after enough people realize they might be/are getting them, people will stop feeding it as much, or introduce noise on purpose. It's not a perfect vacuum of constant reliable information forever. We are already seeing that AI models learning from web results are getting caught up in their own slop making themselves dumber. And the sheer volume of information relative to the computing power necessary to process everything will also become a problem if they keep trying to process every single thing.