pkjqpg1h

joined 2 days ago
[–] pkjqpg1h@lemmy.zip 1 points 3 minutes ago

Tell them that and it ignores it.

You can't trust that it's impossible by it's architecture like if you tell it reset your memory... and it will simulate that it forgot, but it didn't and it will affect all prompts

This is way all models easily leak their system prompts.

[–] pkjqpg1h@lemmy.zip 1 points 10 minutes ago

could you explain little bit more

Any work a LLM can perform effectively is work no human should be performing.

[–] pkjqpg1h@lemmy.zip 2 points 12 minutes ago

why downvote this comment?

[–] pkjqpg1h@lemmy.zip 1 points 32 minutes ago

What a week! GLM-5, Minimax-2.5 and now Qwen-3.5 let's see

[–] pkjqpg1h@lemmy.zip 5 points 45 minutes ago

hmm, let's correct: never trust AI for advice

[–] pkjqpg1h@lemmy.zip 15 points 2 hours ago

"Anna's Archive is a non-profit project with two goals:

  1. Preservation: Backing up all knowledge and culture of humanity.

  2. Access: Making this knowledge and culture available to anyone in the world."

Thanks everyone who contributed this great project.

[–] pkjqpg1h@lemmy.zip 1 points 2 hours ago

I don't think so there is a mutual relationship with AI companies and the copyright's future is not bright

[–] pkjqpg1h@lemmy.zip 1 points 3 hours ago* (last edited 3 hours ago) (1 children)

I'm not a programmer, why are people so interested in Rust?

edit: typo

[–] pkjqpg1h@lemmy.zip 3 points 3 hours ago (1 children)

It's not about markdown and it wasn't accidently

"Improper neutralization of special elements used in a command" read

[–] pkjqpg1h@lemmy.zip 1 points 5 hours ago* (last edited 3 hours ago)

even chatgpt.com is not archivable, this is so stupid if we don't have snapshot of webpages how can we cite anything and trust that later?

edit: added screenshot

 

MiniMax-M2.5 is a SOTA large language model designed for real-world productivity. Trained in a diverse range of complex real-world digital working environments, M2.5 builds upon the coding expertise of M2.1 to extend into general office work, reaching fluency in generating and operating Word, Excel, and Powerpoint files, context switching between diverse software environments, and working across different agent and human teams. Scoring 80.2% on SWE-Bench Verified, 51.3% on Multi-SWE-Bench, and 76.3% on BrowseComp, M2.5 is also more token efficient than previous generations, having been trained to optimize its actions and output through planning.

 

Because they don't really search or index quality content (it's very expensive and hard to do) and their search implementation really sucks, they don't do any real improvement.

The process is like this:

  1. Take the user query and create 1-3 queries. For this process they use very stupid but fast and cheap models; because of that, sometimes they create very stupid search queries and, unlike a pro, they don't really know how to use search engines, like filtering, ranking, focusing...
  2. Combine these search results (it contains slop AI-generated summary pages, YouTube videos, maybe forums, maybe Wikipedia...).
  3. Use RAG with an LLM to find answers. LLMs will always try to find answers quickly, and instead of making a thinking loop in a long article they will use that slop page with a direct answer.

As you can see, there are many, many problems in this implementation:

  • The biggest problem is citation: they cite confidently but it's wrong.
  • They use low-quality data, like auto YouTube subtitles, improperly extracted tables and elements, content-farm sites, copycat sites, corporate blogs...
  • Their search results are low quality.
  • For the most important part (breaking down the user request) they use cheap, stupid models.
  • They handle all data in the same context instead of parallel requests (which is very expensive)

It's still strange to me: we always say "they have all the data, all the money, all the hardware..." but they still can't create a better AI search than random FOSS developers.

 

It says

A proprietary service that combines GPS, Wi-Fi, cellular networks etc. to determine your location more accurately. It may connect to Google servers.

But I don't have Google Play Services, I use microG and microG don't use Google's services for location they use beacondb.net. Can we change this?

view more: next ›