this post was submitted on 26 Oct 2025
218 points (99.5% liked)

science

22279 readers
141 users here now

A community to post scientific articles, news, and civil discussion.

rule #1: be kind

founded 2 years ago
MODERATORS
top 22 comments
sorted by: hot top controversial new old
[–] SaraTonin@lemmy.world 2 points 1 day ago

This is awesome, but things like how the paper score is calculated should probably be on the website itself, not some scrolling and a couple of clicks away on the GitHub page. Maybe a single link or pop up on the chart itself?

[–] Carrolade@lemmy.world 40 points 4 days ago (4 children)

Five maps so far. Is someone doing this by hand, the hard way? I figured it was an AI someone programmed, but if it's an individual or small team, big respect. Very neat project.

[–] porksnort@slrpnk.net 12 points 3 days ago

They provide links to their github that explains their whole methodology. This is a scientific effort and is as transparent and well-documented as a project can be. They provide the code so you can understand the exact mechanics at play or just fork the project if you want to take the work in a different direction.

It’s a great project and long overdue. I personally think scientific journals are incredibly outdated and haven’t been necessary for a couple of decades. Just put your work on a stable web site and cut out the parasites at the journals.

[–] obsoleteacct@lemmy.zip 4 points 3 days ago

I hope they are able to grow this without compromising the quality or vision, because I'm sure they'll have lots of people willing to get involved, but maybe not all for the most ethical reasons.

[–] oyzmo@lemmy.world 14 points 4 days ago

github links show 2x contributers. cool project

[–] CatsPajamas@lemmy.dbzer0.com -3 points 3 days ago (1 children)

AI would probably be pretty useful for this. You'd have to assume most of the "answers" are in the abstract, so you could just build one to scrape academic texts. Use an RAG so it doesn't hallucinate, maybe. Idk if that violates some T&C nonsense that doing it by hand doesn't though.

[–] entropicdrift@lemmy.sdf.org 8 points 3 days ago (1 children)

This is a bad idea. It's extremely likely to hallucinate at one point or another no matter how many tools you equip it with, and humans will eventually miss some fully made up citation or completely misrepresented conclusion.

[–] CatsPajamas@lemmy.dbzer0.com -4 points 3 days ago* (last edited 3 days ago) (2 children)

Google RAG

There are tons of AIs that are not auto regressive LLMs

[–] entropicdrift@lemmy.sdf.org 4 points 3 days ago

I'm a professional software engineer and I've used RAG. It doesn't prevent all hallucinations. Nothing can. The "hallucinations" are a fundamental part of the LLM architecture.

[–] obsoleteacct@lemmy.zip 3 points 3 days ago (1 children)

Are the down votes because people genuinely think this is an incorrect answer, or because they dislike anything remotely pro-AI?

[–] CatsPajamas@lemmy.dbzer0.com 2 points 3 days ago (1 children)

Both probably. Thought terminating cliches and all that. The most useful tool maybe ever. Wild.

[–] entropicdrift@lemmy.sdf.org 5 points 3 days ago* (last edited 3 days ago) (2 children)

I use LLMs daily as a professional software engineer. I didn't downvote you and I'm not disengaging my thinking here. RAGs don't solve everything, and it's better not to sacrifice scientific credibility to the altar of convenience.

It's always been easier to lie quickly than to dig for the truth. AIs are not consistent, regardless of the additional appendages you give them. They have no internal consistency by their very nature.

[–] porksnort@slrpnk.net 3 points 2 days ago (1 children)

And this isn’t even really a great application for RAG. Papermaps just goes off of references and citations. Perhaps a sentiment analysis would be marginally useful, but since you need a human to verify all LLM outputs it would be a dubious time savings.

The system scores review papers very favorably and the “yes/no/maybe” conclusion is right in the abstract, usually the last sentence or two of it. This is not a prime candidate for any LLM, it’s simple database operations on srtuctured data that already exists. There’s no use case here.

[–] entropicdrift@lemmy.sdf.org 3 points 2 days ago* (last edited 2 days ago)

Perhaps a sentiment analysis would be marginally useful, but since you need a human to verify all LLM outputs it would be a dubious time savings.

Thank you, yes. That's exactly my point. You'd need a human to verify all of the outputs anyways, and these are literally machines that exclusively make text that humans find believable, so you're likely adding to the problem of humans messing stuff up moreso than speeding anything up. Being wrong fast has always been easy, so it's no help here.

[–] CatsPajamas@lemmy.dbzer0.com 1 points 2 days ago

What would the failure rate on this be? What would the rate have to be to actually matter? Literally it would just poll the abstract and spit out yes no undecided. That is in the abstract. There is very little chance of there being any hallucinations that are meaningful at a degree large enough to vary literally anything.

Have you never had it organize things or analyze sentiments? I understand if that's not your use case but this is pretty fundamentally an easy application of AI.

[–] Jokulhlaups@lemmy.world 9 points 4 days ago (1 children)

Please add a section about nature! Global warming, deforestation, and other human effects on nature.

[–] porksnort@slrpnk.net 6 points 4 days ago

You can suggest new maps. They ask for links to papers, so if this is a thing you are passionate about and have some recent papers, especially review papers. Reviews seem to get more points in their schemes.

I love this project too and have a personal passion in neurobiology studies related to benefits of yoga. When I have a couple of hours, I will submit a map suggestion for that topic.

[–] frongt@lemmy.zip 7 points 4 days ago (4 children)

What does the circle size indicate?

[–] procesd@lemmy.world 9 points 4 days ago

From the docs in GitHub: "The size of the dots corresponds to the number of reviewed papers for literature reviews (non-reviews have the smallest size)..."

[–] PKscope@lemmy.world 2 points 4 days ago

I wondered the same. It doesn't seem to correlate to P-Size, citations, or participants. Maybe a combined factor of each that is calculated? I'm really not sure.

[–] 48954246@lemmy.world 2 points 4 days ago

Couldn't quite work that out either. I initially thought it might have been to do with the number of citations but that didn't pan out

[–] Quexotic 1 points 4 days ago

I think it might be the number of papers used to answer the question. Not totally certain though