this post was submitted on 14 Nov 2025
16 points (100.0% liked)

TechTakes

2295 readers
71 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS
top 2 comments
sorted by: hot top controversial new old
[–] o7___o7@awful.systems 5 points 5 days ago* (last edited 5 days ago)

Big "we caught a skid trying to use nmap" vibes

[–] scruiser@awful.systems 3 points 5 days ago

Even taking their story at face value:

  • It seems like they are hyping up LLM agents operating a bunch of scripts?

  • It indicates that their safety measures don't work

  • Anthropic will read your logs, so you don't have any privacy or confidentiality or security using their LLM, but, they will only find any problems months after the fact (this happened in June according to Anthropic but they didn't catch it until September),

If it’s a Chinese state actor … why are they using Claude Code? Why not Chinese chatbots like DeepSeek or Qwen? Those chatbots code just about as well as Claude. Anthropic do not address this really obvious question.

  • Exactly. There are also a bunch of open source models hackers could use for a marginal (if any) tradeoff in performance, with the benefit that they could run locally, so that their entire effort isn't dependent on hardware outsider of their control in the hands of someone that will shut them down if they check the logs.

You are not going to get a chatbot to reliably automate a long attack chain.

  • I don't actually find it that implausible someone managed to direct a bunch of scripts with an LLM? It won't be reliable, but if you can do a much greater volume of attacks maybe that makes up for the unreliability?

But yeah, the whole thing might be BS or at least bad exaggeration from Anthropic, they don't really precisely list what their sources and evidence are vs. what is inference (guesses) from that evidence. For instance, if a hacker tried to setup hacking LLM bots, and they mostly failed and wasted API calls and hallucinated a bunch of shit, if Anthropic just read the logs from their end and didn't do the legwork contacting people who had allegedly been hacked, they might "mistakenly' (a mistake that just so happens to hype up their product) think the logs represent successful hacks.