I'm dumbfounded they chose the name of the infamous NSA mass surveillance program revealed by Snowden in 2013. And even more so that there is just one other comment among 320 pointing this out [1]. Has the technical and scientific community in the US already forgotten this huge breach of trust? This is especially jarring at a time where the US is burning its political good-will at unprecedented rate (at least unprecedented during the life-times of most of us) and talking about digital sovereignty has become mainstream in Europe. As a company trying to promote a product, I would stay as far away from that memory as possible, at least if you care about international markets.
>I'm dumbfounded they chose the name of the infamous NSA mass surveillance program revealed by Snowden in 2013. And even more so that there is just one other comment among 320 pointing this out
I just think it's silly to obsess over words like that. There are many words that take on different meanings in different contexts and can be associated with different events, ideas, products, time periods, etc. Would you feel better if they named it "Polyhedron"?
What the OP was talking about is the negative connotation that goes with the word; it's certainly a poor choice from a marketing point of view.
You may say it's "silly to obsess", but it's like naming a product "Auschwitz" and saying "it's just a city name" -- it ignores the power of what Geffrey N. Leech called "associative meaning" in his taxonomy of "Seven Types of Meaning" (Semantics, 2nd. ed. 1989): speaking that city's name evokes images of piles of corpses of gassed undernourished human beings, walls of gas chambers with fingernail scratches and lamp shades made of human skin.
Well, I don't know anything about marketing and you might have a point, but the severity of impact of these two words is clearly very different, so it doesn't look like a good comparison to me. It would raise quite a few eyebrows and more if, for example, someone released a Linux distro named "Auschwitz OS", meanwhile, even in the software world, there are multiple products that incorporate the word prism in various ways[1][2][3][4][5][6][7][8][9]. I don't believe that an average user encountering the word "prism" immediately starts thinking about NSA surveillance program.
I remember the NSA Prism program, but hearing prism today I would think first of Newton, optics, and rainbows.
I think the ideas was to try to explain why is a problem to choose something, it is not a comparison of the intensity / importance.
I am not sure you can make an argument of "other people are doing it too". Lots of people do things that it is not in their interest (ex: smoking, to pick the easy one).
As others mentioned, I did not have the negative connotation related to the word prism either, but not sure how could one check that anyhow. It is not like I was not surprised these years about what some other people think, so who knows... Maybe someone with experience in marketing could explain how it is done.
Do a lot of people know that Prism is the name of the program? I certainly didn't and consider myself fairly switched on in general
It's likely to be an age thing too. Were you in hacker-related spaces when the Snowden scandal happened ?
(I expect a much higher than average share of people in academia also part of these spaces.)
>Has the technical and scientific community in the US already forgotten this huge breach of trust?
Have you ever seen the comment section of a Snowden thread here? A lot of users here call for Snowden to be jailed, call him a russian asset, play down the reports etc. These are either NSA sock puppet accounts or they won't bite the hand that feeds them (employees of companies willing to breach their users trust).
What Snowden did was heroic. What was shameful was the world's underwhelming reaction. Where were all these images
in the media of protest marches like against the Vietnam war?
Someone once said "Religion is opium for the people." - today, give people a mobile device and some doom-scrolling social media celebrity nonsense app, and they wouldn't noticed if their own children didn't come home from school.
Looking back I think allowing more centralized control to various forms of media to private parties did much worse overall than government surveillance on the long run.
For me the problem was not surveillance, the problem is addiction focused app building (+ the monopoly), and that never seem to be a secret. Only now there are some attempts to do something (like Australia and France banning children - which am not sure is feasible or efficient but at least is more than zero).
Him being (or best case becoming) a russian asset turned out to be true
Like it would matter for any of the revelations. And like he would have other choices to not go to prison. Look at how it worked out for Assange.
Assange became a Russian asset *while* in a whistleblowing-related job.
(And he is also the reason why Snowden ended up in Russia. Though it's possible that the flight plan they had was still the best one in that situation.)
They both undertook something they believed in, and showed extreme courage.
And they did manage to get the word out. They are both relatively free now, but it is true, they both paid a price.
Idealism is that you follow your principles despite that price, not escaping/evading the consequences.
If the messenger has anything to do with Russia, even after the fact, we should dismiss the message and remember to never look up.
He could have been a Chinese asset, but CCP is a coward.
OpenAI has a former NSA director on its board. [1] This connection makes the dilution of the term "PRISM" in search results a potential benefit to NSA interests.
Same, to the point where I was wondering if someone deliberately named it so. But I expect that whoever made this decision simply doesn't know or care.
I came here based to headline expecting some more cia & nsa shit, that word is tarnished for few decades in better part of IT community (that actually cares about this craft beyond paycheck)
And yet, the name immediately reminded me of the Snowden relevations.
They are farming scientists for insight.
This comment might make more sense if there was some connection or similarity between the OpenAI "Prism" product and the NSA surveillance program. There doesn't appear to be.
Except that this lets OpenAI gain research data and scientific ideas by stealing from their users, using their huge mass surveillance platform. So, tremendous overlap.
Isn't most research and scientific data is already shared openly (in publications usually)?
This comment allows ycombinator to steal ideas from their user's comments, using their huge mass news platform. Temendous overlap indeed.
tons of things are called prism.
(full disclosure, yes they will be handin in PII on demands like the same kinda deals, this is 'normal' - 2012 shows us no one gives a shit)
Considering OpenAI is deeply rooted in anti-freedom ethos and surveillance capitalism, I think it is quite a self aware and fitting name.
I mean it's also the name of the national engineering education journal and a few other things. There's only 14,000 5-letter words in English so you're going to have collisions.
> Has the technical and scientific community in the US already forgotten this huge breach of trust?
We haven’t forgotten… it’s mostly that we’re all jaded given the fact that there has been zero ramifications and so what’s the use of complaining - you’re better off pushing shit up a hill
>Has the technical and scientific community in the US already forgotten this huge breach of trust?
Yes, imho, there is a great deal of ignorance of the actual contents of the NSA leaks.
The agitprop against Snowden as a "Russian agent" has successfully occluded the actual scandal, which is that the NSA has built a totalitarian-authoritarian apparatus that is still in wide use.
Autocrats' general hubris about their own superiority has been weaponized against them. Instead of actually addressing the issue with America's repressive military industrial complex, they kill the messenger.
I still can't get over the Apple thing. Haven't enjoyed a ripe McIntosh since. </s>
Previously, this existed as crixet.com [0]. At some point it used WASM for client-side compilation, and later transitioned to server-side rendering [1][2]. It now appears that there will be no option to disable AI [3]. I hope the core features remain available and won’t be artificially restricted. Compared to Overleaf, there were fewer service limitations: it was possible to compile more complex documents, share projects more freely, and even do so without registration.
On the other hand, Overleaf appears to be open source and at least partially self-hostable, so it’s possible some of these ideas or features will be adopted there over time. Alternatively, someone might eventually manage to move a more complete LaTeX toolchain into WASM.
I'm curious how it compares to Overleaf in terms of features? Putting aside the AI aspect entirely, I'm simply curious if this is a viable Overleaf competitor -- especially since it's free.
I do self-host Overleaf which is annoying but ultimately doable if you don't want to pay the $21/mo (!).
I do have to wonder for how long it will be free or even supported, though. On the one hand, remote LaTeX compiling gets expensive at scale. On the other hand, it's only a fraction of a drop in the bucket compared to OpenAI's total compute needs. But I'm hesitant to use it because I'm not convinced it'll still be around in a couple of years.
Overleaf is a little curious to me. What's the point? Just install LaTeX. Claude is very good at manipulating LaTeX documents and I've found it effective at fixing up layouts for me.
In my circles the killer features of Overleaf are the collaborative ones (easy sharing, multi-user editing with track changes/comments). Academic writing in my community basically went from emailed draft-new-FINAL-v4.tex files (or a shared folder full of those files) to basically people just dumping things on Overleaf fairly quickly.
collaboration is the killer feature tbh. overleaf is basically google docs meets latex.. you can have multiple coauthors editing simultaneously, leave comments, see revision history, etc.
a lot of academics aren't super technical and don't want to deal with git workflows or syncing local environments. they just want to write their fuckin' paper (WTFP).
overleaf lets the whole research team work together without anyone needing to learn version control or debug their local texlive installation.
also nice for quick edits from any machine without setting anything up. the "just install it locally" advice assumes everyones comfortable with that, but plenty of researchers treat computers as appliances lol.
To add to the points raised by others, "just install LaTeX" is not imo a very strong argument. I prefer working in a local environment, but many of my colleagues much prefer a web app that "just works" to figuring out what MiKTeX is.
I can code in monospace (of course) but I just can't write in monospace markup. I need something approaching WYSIWIG. It's just how my brain works -- I need the italics to look like italics, I need the footnote text to not interrupt the middle of the paragraph.
The visual editor in Overleaf isn't true WYSIWIG, but it's close enough. It feels like working in a word processor, not in a code editor. And the interface overall feels simple and modern.
(And that's just for solo usage -- it's really the collaborative stuff that turns into a game-changer.)
I use inkdrop for this, then pandoc to go from markdown to latex, then a final typesetting pass. Inkdrop is great for WYSIWYG markdown editing.
Same for me. I wrote my PhD in LyX for that reason.
"Just install LaTeX" is really not a valid response when the LaTeX toolchain is a genuine nightmare to work with. I could do it but still use Overleaf. Managing that locally is just not worth it.
Latex is such a nightmare to work with locally
LaTeX ecosystem is a UX nightmare, coming from someone who had to deal with it recently. Overleaf just works.
I'd use git in this case, I am sure there are other reasons to use overleaf otherwise it wouldn't exist but this seems like a solved issue with git.
You can use actually git (it's also integrated in Overleaf).
You can even export ZIP files if you like (for any cloud service, it's not a bad idea to clone your repo once in a while to avoid begin stuck in case of unlikely downtime).
I have both a hosted instance (thanks to Overleaf/ShareLaTeX Ltd.) and I'm also paying user for the pro group license (>500€/year) for my research team. It's great - esp. for smaller research teams - to have the maintenance outsourced to a commercial provider.
On a good day, I'd spend 40% in Overleaf, 10% in Sublime/Emacs, 20% in Email and 10% in Google Scholar/Semantics Scholar and 10% in EasyChair/OpenReview, the rest in meetings.
you can use git with overleaf, but from practical experience: getting even "mathematically/technically inclined" people to consistently use git takes a lot of time... which one could spend on other more fun things :-)
The first three things are, in this order: collaborative editing, collaborative editing, collaborative editing. Seriously, this cannot be understated.
Then: The LaTeX distribution is always up-to-date; you can run it on limited resources; it has an endless supply of conference and journal templates (so you don't have to scavenge them yourself off a random conference/publisher website); Git backend means a) you can work offline and b) version control comes in for free. These just off the top of my head.
Collaboration is at best rocky when people have different versions of LaTeX packages installed. Also merging changes from multiple people in git are a pain when dealing with scientific, nuanced text.
Overleaf ensures that everyone looks at the same version of the document and processes the document with the same set of packages and options.
The deeper I got, the more I realized really supporting the entire LaTeX toolchain in WASM would mean simulating an entire linux distribution :( We wanted to support Beamer, LuaLaTeX, mobile (wasn't working with WASM because of resource limits), etc.
The WASM constraints make sense given the resource limits, especially for mobile. If you are moving that compute server-side though I am curious about the unit economics. LaTeX pipelines are surprisingly heavy and I wonder how you manage the margins on that infrastructure at scale.
We had been building literally the same thing for the last 8 months along with a great browsing environment over arxiv -- might just have to sunset it
Any plans of having typst integrated anytime soon?
I'm not against typst. I think it's integration would be a lot easier and more straightforward I just don't know if it's really that popular yet in academia.
its not yet, but gaining traction.
But what's the point ?
To end up with yet another shitty (because running inside a browser, in particular its interface) web app ?
Why not focus efforts into making a proper program (you know, with IBM menu bars and keyboard shortcuts), but with collaborative tools too ?
You are right in pointing out that the Web browser isn't the most suitable UI paradigm for highly interactive applications like a scientific typesetting system/text editor.
I have occasionally lost a paragraph just by accidental marking a few lines and pressing [Backspace].
But at the moment, there is no better option than Overleaf, and while I encourage you to write what you propose if you can, Overleaf will be the bar that any such system needs to be compared against.
> Prism builds on the foundation of Crixet, a cloud-based LaTeX platform that OpenAI acquired and has since evolved into Prism as a unified product. This allowed us to start with a strong base of a mature writing and collaboration environment, and integrate AI in a way that fits naturally into scientific workflows.
They’re quite open about Prism being built on top of Crixet.
It seems bad for OpenAI to make this about latex documents, which will be now associated, visually, with AI slop. The opposite of what anyone wants really. Nobody wants you to know they used a chatbot!
Please refrain from incorporating em dashes into your LaTeX document. In summary, the absence of em dashes in LaTeX.
Am I missing something? LaTeX is associated with slop now?
If a common AI tool produces latex documents, the association will be created yeah. Right now latex would be a high indicator of manual effort, right?
don't think so. I think latex was one of academics' earlier use cases of chatgpt, back in 2023. That's when I started noticing tables in every submitted paper looking way more sophisticated than they ever did. (The other early use case of course being grammar/spelling. Overnight everyone got fluent and typos disappeared.)
It's funny, I was reading a bunch of recent papers not long ago (I haven't been in academia in over a decade) and I was really impressed with the quality of the writing in most of them. I guess in some cases LLMs are the reason for that!
I recently got wrongly accused of using LLMs to help write an article by a reviewer. He complained that our (my and my co-worker's) use of "to foster" read "like it was created by ChatGPT". (If our paper was fluent/eloquent, that's perhaps because having an M.A. in Eng. lit. helped for that.)
I don't think any particular word alone can be used as an indicator for LLM use, although certain formatting cues are good signals (dashes, smileys, response structure).
We were offended, but kept quiet to get the article accepted, and we changed some instances of some words to appease them (which thankfully worked). But the wrong accusation left a bit of a bad aftertaste...
If you’ve got an existing paragraph written that you just know could be rephrased more eloquently, and can describe the type of rephrasing/restructuring you want… LLMs absolutely slap at that.
LaTeX is already standard in fields that have math notation, perhaps others as well. I guess the promise is that "formatting is automatic" (asterisk), so its popularity probably extends beyond math-heavy disciplines.
> Right now latex would be a high indicator of manual effort, right?
...no?
Just one Google search for "latex editor" showed more than 2 in the first page.
It's not that different from using a markdown editor.
This is going to be the concrete block which finally breaks the back of the academic peer review system, i.e. it's going to be a DDoS attack on a system which didn't even handle the load before LLMs.
Maybe we'll need to go back to some sort of proof-of-work system, i.e. only accepting physical mailed copies of manuscripts, possibly hand-written...
I tried Prism, but it's actually a lot more work than just using claude code. The latter allows you to "vibe code" your paper with no manual interaction, while Prism actually requires you review every change.
I actually think Prism promotes a much more responsible approach to AI writing than "copying from chatgpt" or the likes.
Or it makes gatekeepers even more important than before. Every submission to a journal will be desk-rejected, unless it is vouched for by someone one of the editors trusts. And people won't even look at a new paper, unless it's vouched for by someone / published in a venue they trust.
> This is going to be the concrete block which finally breaks the back of the academic peer review system
Exactly, and I think this is good news. Let's break it so we can fix at last. Nothing will happen until a real crisis emerges.
Disrupting a system without good proposals for its replacement sounds like a recipe for disaster.
There's problems with the medical system, therefore we should set hospitals on fire to motivate them to make them better.
Maybe Open AI will sell you 'Lens' which will assist with sorting through the submissions and narrow down the papers worth reviewing.
That will just create a market for hand-writers. Good thing the economy is doing very well right, so there aren't that many desperate people who will do it en-masse and for peanuts.
Overleaf basically already has the same thing
Handwriting is super easy to fake with plotters.
Is there something out there to simulate the non-uniformity and errors of real handwriting?
> i.e. only accepting physical mailed copies of manuscripts, possibly hand-written...
And you think the indians will not hand write the output of LLMs ?
Not that I have a better suggestion myself..
It feels generally a bit dangerous to use an AI product to work on research when (1) it's free and (2) the company hosting it makes money by shipping productized research
Why do you think these points would make the usage dangerous?
They have to monetize somehow...
This is a actual prompt in the video: "What are the papers in the literature that are most relevant to this draft and that I should consider citing?"
They probably wanted: "... that I should read?" So that this is at least marketed to be more than a fake-paper generation tool.
You can tell that they consulted 0 scientists to verify the clearly AI-written draft of this video.
The target audience of this tool is not academics; it's OpenAI investors.
I felt the same, but then thought of experts in their field. For example, my PhD advisor would already know all these papers. For him the prompt would actually be similar to what was shown in the video.
Mini paper: that future isn’t the AI replacing humans. its about humans drowning in cheap artifacts.
New unit of measurement proposed: verification debt.
Also introduces: Recursive Garbage → model collapse
This appears to just be the output of LLMs itself? It credits GPT-5.2 and Gemini 3 exclusively as authors, has a public domain license (appropriate for AI output) and is only several paragraphs in length.
Which proves its own points! Absolutely genius! The cost asymmetry of producing and checking for garbage truly is becoming a problem in the recent years, with the advent of LLMs and generative AI in general.
[deleted]
Totally agree!
I feel like this means that working in any group where individuals compete against each other results in an AI vs AI content generation competition, where the human is stuck verifying/reviewing.
> Totally agree!
Not a dig on your (very sensible) comment, but now I always do a double take when I see anyone effusively approving of someone else's ideas. AI turned me into a cynical bastard :(
Yes, I did it as a joke inspired by the PRISM release. But unexpectedly, it makes a good point. And the funny part for was that the paper lists only LLMs as authors.
Also, in a world where AI output is abundant, we humans become the scarce resource the "tools" in the system that provide some connectivity to reality (grounding) for LLM
Plot twist: humans become the new Proof of Work consensus mechanism. Instead of GPUs burning electricity to hash blocks, we burn our sanity verifying whether that Medium article was written by a person or a particularly confident LLM.
"Human Verification as a Service": finally, a lucrative career where the job description is literally "read garbage all day and decide if it's authentic garbage or synthetic garbage." LinkedIn influencers will pivot to calling themselves "Organic Intelligence Validators" and charge $500/hr to squint at emails and go "yeah, a human definitely wrote this passive-aggressive Slack message."
The irony writes itself: we built machines to free us from tedious work, and now our job is being the tedious work for the machines. Full circle. Poetic even. Future historians (assuming they're still human and not just Claude with a monocle) will mark this as the moment we achieved peak civilization: where the most valuable human skill became "can confidently say whether another human was involved."
Bullish on verification miners. Bearish on whatever remains of our collective attention span.
Human CAPTCHA exists to figure out whether your clients are human or not, so you can segment them and apply human pricing. Synthetics, of course, fall into different tiers. The cheaper ones.
Bullish on verifiers who accept money to verify fake things
From my perspective as a journal editor and a reviewer these kinds of tools cause many more problems than they actually solve. They make the 'barrier to entry' for submitting vibed semi-plausible journal articles much lower, which I understand some may see as a benefit. The drawback is that scientific editors and reviewers provide those services for free, as a community benefit. One example was a submission their undergraduate affiliation (in accounting) to submit a paper on cosmology, entirely vibe-coded and vibe-written. This just wastes our (already stretched) time. A significant fraction of submissions are now vibe-written and come from folks who are looking to 'boost' their CV (even having a 'submitted' publication is seen as a benefit), which is really not the point of these journals at all.
I'm not sure I'm convinced of the benefit of lowering the barrier to entry to scientific publishing. The hard part always has been, and always will be, understanding the research context (what's been published before) and producing novel and interesting work (the underlying research). Connecting this together in a paper is indeed a challenge, and a skill that must be developed, but is really a minimal part of the process.
GenAI largely seems like a DDoS on free resources. The effort to review this stuff is now massively more than the effort to "create" it, so really what is the point of even submitting it, the reviewer could have generated it themself. Seeing it in software development where coworkers are submitting massive PRs they generated but hardly read or tested. Shifting the real work to the PR review.
I'm not sure what the final state would be here but it seems we are going to find it increasingly difficult to find any real factual information on the internet going forward. Particularly as AI starts ingesting it's own generated fake content.
More relevant than ever:
> The amount of energy needed to refute bullshit is an order of magnitude bigger than that needed to produce it.
The P≠NP conjecture in CS says checking a solution is easier than finding one. Verifying a Sudoku is fast; solving it
from scratch is hard. But Brandolini's Law says the opposite: refuting bullshit costs way more than producing it.
Not actually contradictory. Verification is cheap when there's a spec to check against. 'Valid Sudoku?' is mechanical.
But 'good paper?' has no spec. That's judgment, not verification.
> The P≠NP conjecture in CS says checking a solution is easier than finding one...
... for NP-hard problems.
It says nothing about the difficulty of finding or checking solutions of polynomial ("P") or exponential ("EXPTIME") problems.
producing BS can be equated to generating statements without caring for their truth value. Generating them is easy. Refuting them requires one to find a proof or a contradiction which is a lot of work, and is equal to "solving" the statement. As an analogy, refuting BS is like solving satisfiability, whereas generating BS is like generating propositions.
It's not contradictory because solving and producing bullshit are very different things.
Generating less than 81 random numbers between 1 and 9 is probably also cheaper than verifying correctness of a sudoku.
Wow the 3 comments from OC to here are all bangers, they combine into a really nice argument against these toys
> The effort to review this stuff is now massively more than the effort to "create" it
I don't doubt the AI companies will soon announce products that will claim to solve this very problem, generating turnkey submission reviews. Double-dipping is very profitable.
It appears LLM-parasitism isn't close to being done, and keeps finding new commons to spoil.
There are a dozen startups that do this.
In some ways it might be a good thing that shorthand signals of quality are being destroyed because it forces all of us to meaningfully engage with the work. No more LGTM +1 when every PR looks good.
If I submitted this, I'd have to punch myself in the face repeatedly.
The great disappointment is that the humans submitting these just don’t care it’s slop and they’re wasting another human’s time. To them, it’s a slot machine you just keep cranking the arm of until coins come out. “Prompt until payout.”
I'm scared that this type of thing is going to do to science journals what AI-generated bug reports is doing to bug bounties. We're truly living in a post-scarcity society now, except that the thing we have an abundance of is garbage, and it's drowning out everything of value.
In a corollary to Sturgeon's Law, I'd propose Altman's Law: "In the Age of AI, 99.999...% of everything is crap"
Altman's Law: 99% of all content is slop
I can get behind this.
This assumes a tool will need to be made to help determine the 1% that isn't slop.
At which point I assume we will have reinvented web search once more.
Has anyone looked at reviving PageRank?
I mean Kagi is probably the PageRank revival we are talking about.
I have heard from people here that Kagi can help remove slop from searches so I guess yeah.
Although I guess I am DDG user and I love using DDG as well because its free as well but I can see how for some price can be a non issue and they might like kagi more.
So Kagi / DDG (Duckduckgo) yeah.
I’ve been a Kagi subscriber for a while now. Recently picked up ChatGPT Business and now am considering dropping Kagi since I am only using it for trivial searches. Every comparison I’ve done with deep searches by hand and with AI ended up with the same results in far less time using AI.
Does anyone have kept an eye of who uses what back-end?
DDG used to be meta-search on top of Yahoo, which doesn't exist anymore. What do Gabriel and co-workers use now?
I think they all use Bing now.
Kagi is mostly stealing results from Google and disenshittifying them but mixes in other engines like Yandex and Mojeek and Bing.
DDG is Bing.
There's this thing where all the thought leaders in software engineering ask "What will change about building about building a business when code is free" and while, there are some cool things, I've also thought, like it could have some pretty serious negative externalities? I think this question is going to become big everywhere - business, science, etc. which is like - Ok, you have all this stuff, but do is it valuable? Which of it actually takes away value?
To be fair, the question “what will change” does not presume the changes will be positive. I think it’s the right question to ask, because change is coming whether we like it or not. While we do have agency, there are large forces at play which impact how certain things will play out.
Digital pollution.
The first casualty of LLMs was the slush pile--the unsolicited submission pile for publishers. We've since seen bug bounty programs and open source repositories buckle under the load of AI-generated contributions. And all of these have the same underlying issue: the LLM makes it easy to do things that don't immediately look like garbage, which makes the volume of submission skyrocket while the time-to-reject also goes up slightly because it passes the first (but only the first) absolute garbage filter.
I run a small print-on-demand platform and this is exactly what we're seeing. The submissions used to be easy to filter with basic heuristics or cheap classifiers, but now the grammar and structure are technically perfect. The problem is that running a stronger model to detect the semantic drift or hallucinations costs more than the potential margin on the book. We're pretty much back to manual review which destroys the unit economics.
If it's print-on-demand, why does it matter? Why shouldn't you accept someone's money to print slop for them?
Why would detecting AI be more expensive than creating it?
Soon, poor people will talk to a LLM, rich people will get human medical care.
I mean I'm currently getting "expensive" medical care and the doctors are still all using AI scribes. I wouldn't assume there would be a gap in anything other than perception. I imagine doctors that cater to the fuck you rich will just put more effort into hiding it.
No one, at all levels, wants to do notes.
My experience has been that the transcriptions are way more detailed and correct when doctors use these scribes.
You could argue that not writing down everything provides a greater signal-noise ratio. Fair enough, but if something seemingly inconsequential is not noted and something is missed, that could worsen medical care.
I'm not sure how this affects malpractice claims - It's now easier to prove (with notes) that the doc "knew" about some detail that would otherwise not have been note down.
I totally agree. I spend my whole day from getting up to going to bed (not before reading HN!) on reviews for a conference I'm co-organizing later this year.
So I was not amused about this announcement at all, however easy it may make my own life as an author (I'm pretty happy to do my own literature search, thank you very much).
Also remember, we have no guarantee that these tools will still exist tomorrow, all these AI companies are constantly pivoting and throwing a lot of things at the wall to see what sticks.
OpenAI chose not to build a serious product, as there is no integration with the ACM DL, the IEEE DL, SpringerNatureLink, the ACL Anthology, Wiley, Cambridge/Oxford/Harvard University Press etc. - only papers that are not peer reviewed (arXiv.org) are available/have been integrated. Expect a flood of BS your way.
When my student submit a piece of writing, I can ask them to orally defend their opus maximum (more and more often, ChatGPT's...); I can't do the same with anonymous authors.
Speaking of conferences, might this not be the way to judge this work? You could imagine only orally defended work to be publishable, or at least have the prestige of vetting, in a bit of an old-school science revival.
I wonder if there's a way to tax the frivolous submissions. There could be a submission fee that would be fully reimbursed iff the submission is actually accepted for publication. If you're confident in your paper, you can think of it as a deposit. If you're spamming journals, you're just going to pay for the wasted time.
Maybe you get reimbursed for half as long as there are no obvious hallucinations.
The journal that I'm an editor for is 'diamond open access', which means we charge no submission fees and no publication fees, and publish open access. This model is really important in allowing legitimate submissions from a wide range of contributors (e.g. PhD students in countries with low levels of science funding). Publishing in a traditional journal usually costs around $3000.
Those journals are really good for getting practice in writing and submitting research papers, but sometimes they are already seen as less impactful because of the quality of accepted papers. At least where I am at, I don't think the advent of AI writing is going to affect how they are seen.
In the field of Programming Languages and Formal Methods, many of the top journals and conference proceedings are open access
Who pays the operating expenses?
Welcome to new world of fake stuff i guess
If the penalty for a crime is a fine, then that law exists only for the lower class
In other words, such a structure would not dissuade bad actors with large financial incentives to push something through a process that grants validity to a hypothesis. A fine isn't going to stop tobacco companies from spamming submissions that say smoking doesn't cause lung cancer or social media companies from spamming submissions that their products aren't detrimental to the mental health.
I'm sure my crude idea has it's shortcomings, but this feels superfluous. Deep-pocketed propagandists can do all sorts of things to pump their message whether a slop tax exists or not. There may or may not be existing countermeasures at journals for that. This just isn't really about that. It's about making sure that, in the process of spamming the journal, they also fund the review process, which would otherwise simply bleed time and money.
That would be tricky, I often submitted to multiple high impact journals going down the list until someone accepted it. You try to ballpark where you can go but it can be worth aiming high. Maybe this isn't a problem and there should be payment for the efforts to screen the paper but then I would expect the reviewers to be paid for their time.
I mean your methodology also sounds suspect. You're just going down a list until it sticks. You don't care where it ends up (I'm sure within reason) just as long as it is accepted and published somewhere (again, within reason).
Scientists are incentivized to publish in as high-ranking a journal as possible. You’re always going to have at least a few journals where your paper is a good fit, so aiming for the most ambitious journal first just makes sense.
No different from applying to jobs. Much like companies, there are a variety of journals with varying levels of prestige or that fit your paper better/worse. You don't know in advance which journals will respond to your paper, which ones just received submissions similar to yours, etc.
Plus, the t in me from submission to acceptance/rejection can be long. For cutting edge science, you can't really afford to wait to hear back before applying to another journal.
All this to say that spamming 1,000 journals with a submission is bad, but submitting to the journals in your field that are at least decent fits for your paper is good practice.
It's standard practice, nothing suspect about their approach - and you won't go lower and lower and lower still because at some point you'll be tired of re-formatting, or a doctoral candidate's funding will be used up, or the topic has "expired" (= is overtaken by reality/competition).
This is effectively standard across the board.
Pay to publish journals already exist.
This is sorta the opposite of pay to publish. It's pay to be rejected.
I would think it would act more like a security deposit, and you'd get back 100%, no profit for the journal (at least in that respect).
I’d worry about creating a perverse incentive to farm rejected submissions. Similar to those renter application fee scams.
Pay to review is common in Econ and Finance.
Variation I thought of on pay-to-review:
Suppose you are an independent researcher writing a paper. Before submitting it for review to journals, you could hire a published author in that field to review it for you (independently of the journal), and tell you whether it is submission-worthy, and help you improve it to the point it was. If they wanted, they could be listed as coauthor, and if they don't want that, at least you'd acknowledge their assistance in the paper.
Because I think there are two types of people who might write AI slop papers: (1) people who just don't care and want to throw everything at the wall and see what sticks; (2) people who genuinely desire to seriously contribute to the field, but don't know what they are doing. Hiring an advisor could help the second group of people.
Of course, I don't know how willing people would be to be hired to do this. Someone who was senior in the field might be too busy, might cost too much, or might worry about damage to their own reputation. But there are so many unemployed and underemployed academics out there...
Better yet, make a "polymarket" for papers where people can bet on which paper can make it, and rely on "expertise arbitrage" to punish spams.
Doesn't stop the flood, i.e. the unfair asymmetry between the effort to produce vs. effort to review.
Not if submissions require some small mandatory bet.
Now accepting money from slop companies to verify their slop as notslop
> There could be a submission fee that would be fully reimbursed if the submission is actually accepted for publication.
While well-intentioned, I think this is just gate-keeping. There are mountains of research that result in nothing interesting whatsoever (aside from learning about what doesn't work). And all of that is still valuable knowledge!
Sure, but now we can't even assume that such research is submitted in good faith anymore. There just seems to be no perfect solution.
Maybe something like a "hierarchy/DAG? of trusted-peers", where groups like universities certify the relevance and correctness of papers by attaching their name and a global reputation score to it. When it's found that the paper is "undesirable" and doesn't pass a subsequent review, their reputation score deteriorates (with the penalty propagating along the whole review chain), in such a way that:
- the overall review model is distributed, hence scalable (everybody may play the certification game and build a reputation score while doing so)
- trusted/established institutions have an incentive to keep their global reputation score high and either put a very high level of scrutiny to the review, or delegate to very reputable peers
- "bad actors" are immediately punished and universally recognized as such
- "bad groups" (such as departments consistently spamming with low quality research) become clearly identified as such within the greater organisation (the university), which can encourage a mindset of quality above quantity
- "good actors within a bad group" are not penalised either because they could circumvent their "bad group" on the global review market by having reputable institutions (or intermediaries) certify their good work
There are loopholes to consider, like a black market of reputation trading (I'll pay you generously to sacrifice a bit of your reputation to get this bad science published), but even that cannot pay off long-term in an open system where all transactions are visible.
Incidentally, I think this may be a rare case where a blockchain makes some sense?
You have some good ideas there, it's all about incentives and about public reputation.
But it should also fair. I once caught a team at a small Indian branch of a very large three letter US corporation violating the "no double submission" rule of two conferences: they submitted the same paper to two conferences, both naturally landed in my reviewer inbox, for a topic I am one of the experts in.
But all the other employees should not be penalized by the violations of 3 researchers.
This idea looks very similar to journals! Each journal has a reputation, if they publish too much crap, the crap is not cited and the impact factors decrease. Also, they have an informal reputation, because impact index also has problems.
Anyway, how will universities check the papers? Somone must read the preprints, like the current reviewers. Someone must check the incoming preprints, find reviewers and make the final decition, like the current editors. ...
How would this work for independent researchers?
(no snark)
This keeps repeating in different domains: we lower the cost of producing artifacts and the real bottleneck is evaluating them.
For developers, academics, editors, etc... in any review driven system the scarcity is around good human judgement not text volume. Ai doesn't remove that constraint and arguably puts more of a spotlight on the ability to separate the shit from the quality.
Unless review itself becomes cheaper or better, this just shifts work further downstream and disguising the change as "efficiency"
> Unless review itself becomes cheaper or better, this just shifts work further downstream and disguising the change as "efficiency"
Or the providers of the models are capable of providing accepted/certified guarantees as to the quality of the output that their models and systems produce.
This has been discussed previously as "workslop", where you produce something that looks at surface level like high quality work, but just shifts the burden to the receiver of the workslop to review and fix.
This fits into the broader evolution of the visualization market.
As data grows, visualization becomes as important as processing. This applies not only to applications, but also to relating texts through ideas close to transclusion in Ted Nelson’s Xanadu. [0]
In education, understanding is often best demonstrated not by restating text, but by presenting the same data in another representation and establishing the right analogies and isomorphisms, as in Explorable Explanations. [1]
> these kinds of tools cause many more problems than they actually solve
For whom? For OpenAI these tools are definitely the solutions. They are developing by throwing various AI-powered stuff at the wall to see what sticks. These tools also demonstrate to the investors that innovation did not stall and to show that AI usage is growing.
Same with Microsoft: none of the AI stuff they are shoving down the users' throats were actually designed for the users. All this stuff is only for the token usage to grow for the shareholders to see.
Similar with Google although no one can deny real innovation happening there.
I'm curious if you'd be in favor of other forms of academic gate keeping as well. Isn't the lower quality overall of submissions (an ongoing trend with a history far pre-dating LLMs) an issue? Isn't the real question (that you are alluding to) whether there should be limits to the democratization of science? If my tone seems acerbic, it is only because I sense cognitive dissonance between the anti-AI stance common among many academics and the purported support for inclusivity measures.
"which is really not the point of these journals at all"- it seems that it very much is one of the main points? Why do you think people publish in journals instead of just putting their work on the arxiv? Do you think postdocs and APs are suffering through depression and stressing out about their publications because they're agonizing over whether their research has genuinely contributed substantively to the academic literature? Are academic employers poring over the publishing record of their researchers and obsessing over how well they publish in top journals in an altruistic effort to ensure that the research of their employees has made the world a better place?
> whether there should be limits to the democratization of science?
That is an interesting philosophical question, but not the question we are confronted with. A lot of LLM assisted materials have the _signals_ of novel research without having its _substance_.
LLMs are tools. In the hands of adept, conscientious researchers, they can only be a boon, assisting in the crafting of the research manuscript. In the hands of less adept, less conscientious users, they accelerate the production of slop. The poster I'm responding to seems to be noting an asymmetry- those who find the most use from these tools could be inept researchers who have no business submitting their work. This is because experienced researchers find writing up their results relatively easy.
To me, this is directly relevant to the issue of democratization of science. There seems to be a tool that is inconveniently resulting in the "wrong" people accelerating their output. That is essentially the complaint here rather than any criticism inherent to LLMs (e.g. water/resource usage, environmental impact, psychological/societal harm, etc.). The post I'm responding to could have been written if LLMs were replaced by any technology that resulted in less experienced or capable researchers disproportionately being able to submit to journals.
To be concrete, let's just take one of prism's capabilities- the ability to "turn whiteboard equations or diagrams directly into LaTeX". What a monstrous thing to give to the masses! Before, those uneducated cranks would send word docs to journals with poorly typeset equations, making it a trivial matter to filter them into the trash bin. Now, they can polish everything up and pass off their chicken scratch as respectable work. Ideally, we'd put up enough obstacles so that only those who should publish will publish.
The LLMs does assist the adept researchers in crafting their manuscript, but I do not think it makes the quality much better.
My objection is not that they are the "wrong people". They are just regular people with excellent tools but not necessarily great scientific ideas.
Yes, it was easier to trash the crank's work before based on their unLaTeXed diagrams. Now, they might have a very professional looking diagram, but their work is still not great mathematics. Except that now the editor has a much harder time finding out who submitted a worthwhile paper
In what way do you think the feature of "LaTeXing a whiteboard diagram" is democritizing mathematics? I do not think there are many people who have exceptional mathematical insights but are not able to publish them because they are not able to typeset their work properly.
The democratization is mostly in allowing people from outside the field with mediocre mathematical ideas to finally put them to paper and submit them to mediocre journals. And occasionally it might help a modern day Ramanujan with "exceptional mathematical insights" and a highly unconventional background to not have his work dismissed as that of a crank. Yes, most people with exceptional mathematical insights can typeset quite well. Democratization as I understand the term has quite a higher bar though.
Being against this is essentially to be in favor of a form of discrimination by proxy- if you can't typeset, then likely you can't do research either. And wouldn't it be really annoying if those people who can't research could magically typeset. It's a fundamentally undemocratic impulse: Since those who cannot typeset well are unlikely to produce quality mathematics, we can (and should) use this as an effective barrier to entry. If you replace ability to typeset with a number of other traits, they would be rather controversial positions.
It would indeed be nice if there were a mechanism to find people like Ramanujan who have excellent insights but cannot communicate them effectively.
But LLMs are not really helping. With all the beautifully typeset papers with immaculate prose, Ramanujan's papers are going to be buried deeper!
To some extent, I agree with you that it is a "discrimination by proxy", especially with the typesetting example. But you could think of examples where cranks could very easily fool themselves into thinking that they understand the essence of the material without understanding the details. E.g, [I understand fluid dynamics very well. No, I don't need to work out the differential equations. AI can do the bean counting for me.]
The people on the inside often like all the gatekeeping.
If I may be the Devil's advocate, I'm not sure I fully agree with "The hard part always has been, and always will be, understanding the research context (what's been published before) and producing novel and interesting work (the underlying research)".
Plenty of researchers hate writing and will only do it at gunpoint. Or rather, delegate it all to their underlings.
I don't see an issue with generative writing in principle. The Devil is in the details, but I don't see this as much different from "hey grad student, write me this paper". And generative writing already exists as copy-paste, which makes up like 90% of any random paper given the incrementality of it all.
I was initially a little indignated by the "find me some plausible refs and stick them in the paper" section of the video but, then again, isn't this what most people already do? Just copy-paste the background refs from the colleague's last paper introduction and maybe add one from a talk they saw in the meantime, plus whatever the group & friends produced since then.
My experience is most likely skewed (as all are), but I haven't met a permanent researcher that wrote their own papers yet, and most grad students and postdocs hate writing. Literally the only times I saw someone motivated to write papers (in a masochistic way) were just before applying to a permanent position or while wrapping up their PhD.
Onto your point, though, I agree this is somewhat worrisome in that, by reaction, the barrier to entry might rise by way of discriminating based on credentials.
Thank you for bringing this nuanced view.
I also am not sure why so many people are vehemently against this. I would bet that at least 90% of researchers would agree that the writing up is definitely not the part of the work they prefer (to stay polite). As you mentioned, work is usually relegated to students, and those students already had access to LLMs if they wanted to generate the work.
In my opinion, most of those tools become problematic when people use them without caution. Unfortunately, even in sciences, people are not as careful and pragmatic as we would like to imagine they are and a lot of people are cutting corners, especially in those "lesser" areas like writing and presenting your work.
Overall, I think this has the potential to reshape the publication system, which is long overdue.
I am a rather slow writer who certainly might benefit from something like Prism.
A good tool would encourage me, help me while I am writing, and maybe set up barriers that keep me from taking shortcuts (e.g. pushing me to re-read the relevant paragraphs of a paper that I cite).
Prism does none of these things - instead it pushes me towards sloppy practices, such as sprinkling citations between claims.
Why won't ChatGPT tell me how to build a bomb but Prism will happily fabricate fake experimental results for me?
The comparison to make here is that a journal submission is effectively a pull request to humanities scientific knowlegde base. That PR has to be reviewed. We're already seeing the effects of this with open source code - the number of PR submissions have skyrocketed, overwhelming maintainers.
This is still a good step in a direction of AI assisted research, but as you said, for the moment it creates as many problems as it solves.
I generally agree.
On the other hand, the world is now a different place as compared to when several prominent journals were founded (1869-1880 for Nature, Science, Elsevier). The tacit assumptions upon which they were founded might no longer hold in the future. The world is going to continue to change, and the publication process as it stands might need to adapt for it to be sustainable.
As I understand it, the problem isn't publication or how it's changing over time, it's about the challenges of producing new science when the existing one is muddied in plausible lies. That warrants a new process by which to assess the inherent quality of a paper, but even if it comes as globally distributed, the cheats have a huge advantage considering the asymmetry between the effort to vibe produce vs. the tedious human review.
That’s a good point. On the other hand, we’ve had that problem long before AI. You already need to mentally filter papers based on your assessment of the reputability of the authors.
The whole process should be made more transparent and open from the start, rather than adding more gatekeeping. There ought to be openness and transparency throughout the entire research process, with auditing-ability automatically baked in, rather than just at the time of publication. One man’s opinion, anyway.
As a non-scientist (but long-time science fan and user), I feel your pain with what appears to be a layered, intractable problem.
> > who are looking to 'boost' their CV
Ultimately, this seems like a key root cause - misaligned incentives across a multi-party ecosystem. And as always, incentives tend to be deeply embedded and highly resistant to change.
Perhaps the real issue is the gate-keeping scientific publishing model. Journals had a place and role, and peer-review is a critical aspect of the scientific process but new times (internet, citizien science, higher levels of scientific literacy, and now AI) diminish the benefits of journals creating "barriers to entry" as you put it.
I for one hope not to live in a world where academic journals fall out of favor and are replaced by vibe-coded papers by citizen scientists with inflated egos from one too many “you’re absolutely right!” Claude responses.
Me neither, but what you present is a false dichotomy. Science used to be a past time of the wealthy elites, it became a profession. By opening up it up progrss was accelerated. Same will happen when publication will be made more open and accessible.
Is it at all possible to have a policy that bans the submission of any AI written text, or text that was written with the assistance of AI tools? I understand that this would, by necessity, be under an "honor system" but maybe it could help weed out papers not worth the time?
this is probably a net negative as there are many very good scientists with not very strong English skills.
the early years of LLMs (when they were good enough to correct grammar but not enough to generate entire slop papers) were an equalizer. we may end up here but it would be unfortunate.
I'm certain your journal will be using LLMs in reviewing incoming articles, if they aren't already. I also don't think this is in response to the flood of LLM generated articles. Even if authors were the same as pre-LLM, journals would succumb to the temptation, at least at the big 5 publishers, which already have a contentious relationship with the referees.
Why not filter out papers from people without credentials? And also publicly call them out and register them somewhere, so that their submission rights can be revoked by other journals and conferences after "vibe writing".
These acts just must have consequences so people stop doing them. You can use AI if you are doing it well but if you are wasting everyones time you should just be excluded from the discourse altogether.
What do credentials have to do with good science? There are already some roadblocks to publish science in important–sounding journals, but it's important for the neutrality of the scientific process that in principle anyone can do it.
This dynamic would create even more gate-keeping using credentials, which is already a problem with academia.
[deleted]
wouldn't AI actually be good for filtering given it's going to be a lot better at knowing what has been published? Also seems possible that it could actually work out papers that have ideas that are novel, or at least come up with some kind of likely score.
Completely agree. Look at the independent research that gets submitted under "Show HN" nowadays:
I am very sympathetic to your point of view, but let me offer another perspective. First off, you can already vibe-write slop papers with AI, even in LaTeX format--tools like Prism are not needed for that. On the other hand, it can really help researchers improve the quality of their papers. I'm someone who collaborates with many students and postdocs. My time is limited and I spend a lot of it on LaTeX drudgery that can and should be automated away, so I'm excited for Prism to save time on writing, proofreading, making TikZ diagrams, grabbing references, etc.
What the heck is the point of a reference you never read?
By "grabbing references" I meant queries of the type "add paper [bla] to the bibliography" -- that seems useful to me!
Focusing in on "grabbing references", it's as easy as drag-and-drop if you use Zotero. It can copy/paste references in BibTeX format. You can even customize it through the BetterBibTeX extension.
If you're not a Zotero user, I can't recommend it enough.
I have a terrible memory for details, I'll admit an LLM I can just tell "Find that paper by X's group on Method That Does This And That" and finds me the paper is enticing. I say this because I abandoned Zotero once the list of refs became large enough that I could never find anything quickly.
AI generating references seems like a hop away from absolute unverifiable trash.
I appreciate and sympathize with this take. I'll just note that, in general, journal publications have gone considerably downhill over the last decade, even before the advent of AI. Frequency has gone up, quality has gone down, and the ability to actually check if everything in the article is actually valid is quite challenging as frequency goes up.
This is a space that probably needs substantial reform, much like grad school models in general (IMO).
I know many people have negative opinions about this.
I'd also like to share what I saw. Since GPT-4o became a thing, everyone who submits academic papers I know in my non-english speaking country (N > 5) has been writing papers in our native language and translating them with GPT-4o exclusively. It has been the norm for quite a while. If hallucination is such a serious problem it has been so for one and half a year.
Translation is something Large Language Models are inherently pretty good at, without controversy, even though the output still should be independently verified. It's a language task and they are language models.
Of course. Transformers were originally invented for Google Translate.
Are they good at translating scientific jargon specific to a niche within a field? I have no doubt LLMs are excellent at translating well-trodden patterns; I'm a bit suspicious otherwise..
In my experience of using it to translate ML work between English->Spanish|Galician, it seems to literally translate jargon too eagerly, to the point that I have to tell it to maintain specific terms in English to avoid it sounding too weird (for most modern ML jargon there really isn't a Spanish translation).
It seems to me that jargon would tend to be defined in one language and minimally adapted in other languages. So I’d not sure that would be much of a concern.
for better or for worse, most specific scientific jargon is already going to be in english
[deleted]
I've heard that now that AI conferences are starting to check for hallucinated references, rejection rates are going up significantly. See also the Neurips hallucinated references kerfuffle [1]
Honestly, hallucinated references should simply get the submitter banned from ever applying again. Anyone submitting papers or anything with hallucinated references shall be publicly shamed. The problem isn't only the LLMs hallucinating, it's lazy and immoral humans who don't bother to check the output either, wasting everyone's time and corroding public trust in science and research.
I fully agree. Not reading your own references should be grounds for banning, but that's impossible to check. Hallucinated references cannot be read, so by definition,they should get people banned.
Yeah that's not going to work for long. You can draw a line in 2023, and say "Every paper before this isn't AI". But in the future, you're going to have AI generated papers citing other AI slop papers that slipped through the cracks, because of the cost of doing reseach vs the cost of generating AI slop, the AI slop papers will start to outcompete the real research papers.
It's quite a safe case if you maintain provenance because there is a ground truth to compare to, namely the untranslated paper.
Good idea to name this after the spy program that Snowden talked about.
idk if OpenAI knew that Prism is already a very popular desktop app for scientists and that it's one of the last great pieces of optimized native software?
They don't care. Musk stole a chunk Heinlein's literary legacy with Grok (which unlike prism wasn't a common word) and noone bat an eye.
> Grok (which unlike prism wasn't a common word)
"Grok" was a term used in my undergrad CS courses in the early 2010s. It's been a pretty common word in computing for a while now, though the current generation of young programmers and computer scientists seem not to know it as readily, so it may be falling out of fashion in those spaces.
Grok was specifically coined by Heinlein in _Stranger in a Strange Land_. It's been used in nerd circles for decades before your undergrad times but was never broadly known.
Wikipedia about Groklaw [1]
> Groklaw was a website that covered legal news of interest to the free and open source software community. Started as a law blog on May 16, 2003, by paralegal Pamela Jones ("PJ"), it covered issues such as the SCO-Linux lawsuits, the EU antitrust case against Microsoft, and the standardization of Office Open XML.
> Its name derives from "grok", roughly meaning "to understand completely", which had previously entered geek slang.
He is referencing the book Stranger in a Strange Land, written in 1961.
Grok has been nerd slang for a while. I bet it's in that ESR list of hacker lingo. And hell if every company in silicon valley gets to name their company after something from Lord of the Rings why can't he pay homage to an author he likes
He stole a letter, too.
That bothers more than it should. Every single time I see a new post about Twitter, I think that there's some update for X11 or X Server or something, only to be reminded that Twitter has been changed.
I very much doubt they knew much about what they were building if they didn't know this.
I thought this was about the Prism Database ORM. Or that was Prisma?
Having a chatbot that can natively "speak" latex seems like it might be useful to scientists that already use it exclusively for their work. Writing papers is incredibly time-consuming for a lot of reasons, and having a helper to make quick (non-substantive) edits could be great. Of course, that's not how people will use it...
I would note that Overleaf's main value is as a collaborative authoring tool and not a great latex experience, but science is ideally a collaborative effort.
The video shows a user asking Prism to find articles to cite and to put them in a bib file. But what's the point of citing papers that aren't referenced in the paper you're actually writing? Can you do that?
Edit: You can add papers that are not cited, to bibliography. Video is about bibliography and I was thinking about cited works.
A common approach to research is to do literature review first, and build up a library of citable material. Then when writing your article, you summarize the relevant past research and put in appropriate citations.
To clarify, there is a difference between a bibliography (a list of relevant works but not necessarily cited), and cited work (a direct reference in an article to relevant work). But most people start with a bibliography (the superset of relevant work) to make their citations.
Most academics who have been doing research for a long time maintain an ongoing bibliography of work in their field. Some people do it as a giant .bib file, some use software products like Zotero, Mendeley, etc. A few absolute psychos keep track of their bibliography in MS Word references (tbh people in some fields do this because .docx is the accepted submission format for their journals, not because they are crazy).
> a bibliography (a list of relevant works but not necessarily cited)
Didn't know that there's difference between bibliography and cited work. thank you.
Yes but you should read your bibliography.
I once took a philosophy class where an essay assignment had a minimum citation count.
Obviously ridiculous, since a philosophical argument should follow a chain of reasoning starting at stated axioms.
Citing a paper to defend your position is just an appeal to authority (a fallacy that they teach you about in the same class).
The citation requirement allowed the class to fulfill a curricular requirement that students needed to graduate, and therefore made the class more popular.
In coursework, references are often a way of demonstrating the reading one did on a topic before committing to a course of argumentation. They also contextualize what exactly the student's thinking is in dialogue with, since general familiarity with a topic can't be assumed in introductory coursework. Citation minimums are usually imposed as a means of encouraging a student to read more about a topic before synthesizing their thoughts, and as a means of demonstrating that work to a professor. While there may have been administrative reasons for the citation minimum, the concept behind them is not unfounded, though they are probably not the most effective way of achieving that goal.
While similar, the function is fundamentally different from citations appearing in research. However, even professionally, it is well beyond rare for a philosophical work, even for professional philosophers, to be written truly ex nihilo as you seem to be suggesting. Citation is an essential component of research dialogue and cannot be elided.
> Citing a paper to defend your position is just an appeal to authority
Hmm, I guess I read this as a requirement to find enough supportive evidence to establish your argument as novel (or at least supported in 'established' logic).
An appeal to authority explicitly has no reasoning associated with it; is your argument that one should be able to quote a blog as well as a journal article?
It’s also a way of getting people to read things about the subject that they otherwise wouldn’t. I read a lot of philosophy because it was relevant to a paper I was writing, but wasn’t assigned to the entire class.
Huh? It's quite sensible to make reference to someone else's work when writing a philosophy paper, and there are many ways to do so that do not amount to an appeal to authority.
He's point is that they asked for a minimum number of references not references in general
[deleted]
> Citing a paper to defend your position is just an appeal to authority (a fallacy that they teach you about in the same class).
an appeal to authority is fallacious when the authority is unqualified for the subject at hand. Citing a paper from a philosopher to support a point isn't fallacious, but "<philosophical statement> because my biology professor said so" is.
Many people here talk about Overleaf as if it was the 'dumb' editor without any of these capabilities. It had them for some time via Writefull integration (https://www.writefull.com/writefull-for-overleaf).
Who's going to win will probably be decided by brand recognition with Overleaf having a better starting position in this field, but money obviously being on OAI's side. With some of Writefull's features being dependent on ChatGPT's API, it's clear they are set to be priced-out unless they do something smart.
I don't see anything fancy here that Google doesn't do with their Gemini products, and even better
Naming their tool after the program where private companies run searches on behalf of and give resulting customer data to the NSA....was certainly a choice.
Sir, my tin hat is on.
I'd like to hypothesize a little bit about the strategy of OpenAI. Obviously, it is nice for academic users that there is a new option for collaborative LaTeX editing plus LLM integration for free. At the same time, I don't think there is much added revenue expected here, for example, from Pro features or additional LLM usage plans. My theory is that the value lies in the training data received from highly skilled academics in the form of accepted and declined suggestions.
Even that would be quite niche for OpenAI. They raised far too much capital, and now have to deliver on AGI, fast. Or an ultra-high-growth segment, which has not materialized.
The reason? I can give you the full source for Sam Altman:
while(alive) {
RaiseCapital()
}
That is the full extent of Altman. :)
It is nice for academics, but I would ask why? These aren't tasks you can't do yourself. Yes it's all in one place, but it's not like doing the exact same thing previously was ridiculous to setup.
A comparison comes to mind is the n8n workflow type product they put out before. N8n takes setup. Proofreading, asking for more relevant papers, converting pictures to latex code, etc doesn't take any setup. People do this with or without this tool almost identically.
This seems like a very basic overleaf alternative with few of its features, plus a shallow ChatGPT wrapper. Certainly can’t compete with using VS Code or TeXstudio locally, collaborating through GitHub, and getting AI assistance from Claude Code or Codex.
Loads of researchers have only used LaTeX via Overleaf and even more primarily edit LaTeX using Overleaf, for better or worse. It really simplifies collaborative editing and the version history is good enough (not git level, but most people weren't using full git functionality). I just find that there are not that many features I need when paper writing - the main bottlenecks are coming up with the content and collaborating, with Overleaf simplifying the latter. It also removes a class of bugs where different collaborators had slightly different TeX setups.
I think I would only switch from Overleaf if I was writing a textbook or something similarly involved.
@vicapow replied to keep the Dropbox parallel alive
Yeah I realized the parallel while I was writing my comment! I guess what I'm thinking is that a much better experience is available and there is no in-principle reason why overleaf and prism have to be so much worse, especially in the age of vibe-coding. Prism feels like the result of two days of Claude Code, when they should have invested at least five days.
I could see it seeming likely that because the UI is quite minimalist, but the AI capabilities are very extensive, imo, if you really play with it.
You're right that something like Cursor can work if you're familiar with all the requisite tooling (git, installing cursor, installing latex workshop, knowing how it all works) that most researchers don't want to and really shouldn't have to figure out how to work for their specific workflows.
> Certainly can’t compete with using VS Code or TeXstudio locally, collaborating through GitHub, and getting AI assistance from Claude Code or Codex.
I have a phd in economics. Most researchers in that field have never even heard of any of those tools. Maybe LaTeX, but few actually use it. I was one of very few people in my department using Zotero to manage my bibliography, most did that manually.
Accessibility does matter
The Latent Space podcast just released a relevant episode today where they interviewed Kevin Weil and Victor Powell from, now, OpenAI, with some demos, background and context, and a Q&A.
The YouTube link is here: https://www.youtube.com/watch?v=W2cBTVr8nxU
oh i was here to post it haha - thank you for doing that job for me so I'm not a total shill. I really enjoyed meeting them and was impressed by the sheer ambition of the AI for Science effort at OAI - in some sense I'm making a 10000x smaller scale bet than OAI on AI for Science "taking off" this year with the upcoming dedicated Latent Space Science pod.
generally think that there's a lot of fertile ground for smart generalist engineers to make a ton of progress here this year + it will probably be extremely financially + personally rewarding, so I broadly want to create a dedicated pod to highlight opportunities available for people who don't traditionally think of themselves as "in science" to cross over into the "ai for hard STEM" because it turns out that 1) they need you 2) you can fill in what you don't know 3) it will be impactful/challenging/rewarding 4) we've exhausted common knowledge frontiers and benchmarks anyway so the only* people left working on civilization-impacting/change-history-forever hard problems are basically at this frontier
*conscious exaggeration sorry
Wasn't aware you're so active on HN; sorry for stealing your karma.
Love the idea of a dedicated series/pod where normal people take on hard problems by using and leveraging the emergent capabilities of frontier AI systems.
Anyway, thanks for pod!
Hope you like it :D I'm here if you have questions, too
Slightly off-topic but related: currently I'm in a research environment (biomedicine) where a lot of AI is used. Sometimes well, often poorly. So as an exercise I drafted some rules and commitments about AI and research ("Research After AI: Principles for Accelerated Exploration" [1]), I took the Agile manifesto as a starting point. Anyways, this might be interesting as a perspective on the problem space as I see it.
I’ve been “testing” LLM willingness to explore novel ideas/hypotheses for a few random topics[0].
The earlier LLMs were interesting, in that their sycophantic nature eagerly agreed, often lacking criticality.
After reducing said sycophancy, I’ve found that certain LLMs are much more unwilling (especially the reasoning models) to move past the “known” science[1].
I’m curious to see how/if we can strike the right balance with an LLM focused on scientific exploration.
[0]Sediment lubrication due to organic material in specific subduction zones, potential algorithmic basis for colony collapse disorder, potential to evolve anthropomorphic kiwis, etc.
[1]Caveat, it’s very easy for me to tell when an LLM is “off-the-rails” on a topic I know a lot about, much less so, and much more dangerous, for these “tests” where I’m certainly no expert.
Burried halfway through the article.
> Prism is a free workspace for scientific writing and collaboration
It seems clear to me that this is about OpenAI getting telemetry and other training data with the intent of having their AI do scientific work independently down the line, and I'm very ambivalent about it.
Just more coal to the hype-train - AI companies can't afford news cycle without anything AI. Stock prices must grow!
> In 2025, AI changed software development forever. In 2026, we expect a comparable shift in science,
I can't wait
I postulate 90% of the reason openai now has "variants" for different use cases is just to capture training data...
ChatGPT lets you refuse to allow your content to be used for training (under Preferences -> Data controls), but Prism does not.
[dead]
The example proposed in "and speeding up experimental iteration in molecular biology" has been done since at least the mid-2000s.
It's concerning that this wasn't identified and augur poorly for their search capabilities.
With a tool like this, you could imagine an end-to-end service for restoring and modernizing old scientific books and papers: digitization, cleanup, LaTeX reformatting, collaborative or volunteer-driven workflows, OCR (like Mathpix), and side-by-side comparison with the original. That would be useful.
Don’t forget replication!
I'm curious how you think AI would aide in this.
Tao’s doing a lot of related work in mathematics, so I can say that first of all literature search is a clearly valuable function frontier models offer.
Past that, A frontier LLM can do a lot of critiquing, a good amount of experiment design, a check on statistical significance/power claims, kibitz on methodology..likely suggest experiments to verify or disprove. These all seem pretty useful functions to provide to a group of scientists to me.
Replicate this <slop>
Ok! Here's <more slop>
I don't think you understand what replication means in this context.
I think they do, and you missed some biting, insightful commentary on using LLMs for scientific research.
Not an academic, but I used LaTeX for years and it doesn’t feel like what future of publishing should use. It’s finicky and takes so much markup to do simple things. A lab manager once told me about a study that people who used MS Word to typeset were more productive, and I can see that…
100% completely agreed. It's not the future, it's the past.
The problem is that so many journals require certain LaTeX templates so Typst often isn't an option at all. It's about network effects, and journals don't want to change their entire toolchain.
I've had some good initial results in going from typst to .tex with Claude (Opus 4.5) for an IEEE journal paper - idiomatic use of templates etc.
Latex is good for equations. And Latex tools produce very nice PDFs, but I wouldn't want to write in Latex generally either.
The main feature that's important is collaborative editing (like online Word or Google Docs). The second one would be a good reference manager.
Academic here. Working on MS Word after years of using LaTeX is... hard. With LaTex I can be reassured that the formatting will be 95% fine and the 5% remaining will come down to taste ("why doesn't this Figure show in this page?") while in Word I'm constantly fighting the layout - delete one line? Your entire paragraph is now bold. Changed the font of the entire text? No, that one paragraph ignores you. Want to delete that line after that one Table? F you, you're not. There's a reason why this video joke [1] got 14M views.
And then I need an extra tool for dealing with bibliography, change history is unpredictable (and, IMO, vastly inferior to version control), and everything gets even worse if I open said Word file in LibreOffice.
LaTeX' syntax may be hard, but Word actively fights me during writing.
That study must have compared beginners in LaTeX and MS Word. There is a learning curve, but LaTeX will often save more time in the end.
It is an old language though. LaTeX is the macro system on top of TeX, but now you can write markdown or org-mode (or orgdown) and generate LaTeX -> PDF via pandoc/org-mode. Maybe this is the level of abstraction we should be targeting. Though currently, you still need to drop into LaTeX for very specific fine-tuning.
Agreed. Tex/Latex is very old tech. Error recovery and messages is very bad. Developing new macros in Tex is about as fun as you expect developing in a 70s-era language to be (ie probably similar to cobol and old fortran).
I haven't tried it yet but Typst seems like a promising replacement: https://typst.app/
This assumes that the article, the artifact, is most valuable. But often it is the process of writing the article that has the most value. Prism can be a nice tool for increasing output. But the second order consequence could be that the skill of deep thinking and writing will atrophy.
"There is no value added without sweating"
Work is value and produces sweat, and OpenAI sells just the sweat.
The quality and usefulness of it aside, the primary question is: are they still collecting chats for training data? If so, it limits how comfortable, and sometimes even permitted, people would with working on their yet-to-be-public work using this tool.
They don't call it PRISM for nothing my friend...
The collect chat records for any number of users, not the least of which being NSA surveillance and analysis - highly likely given what we know from the Snowden leaks.
Do you think they used an em-dash in the opening sentence because they’re trying to normalise the AI’s writing style, or…
I haven't used MS Word in quite a while, but I distinctly remember it changed minus signs to em dashes.
> because they’re trying to normalise the AI’s writing style,
AIs use em dashes because competent writers have been using em dashes for a long time. I really hate the fact that we assume em dash == AI written. I've had to stop using em dashes because of it.
Likewise, I’m now reluctant to use any em dashes these days because unenlightened people immediately assume that it’s AI. I used em dashes way before AI decided these were cool
LaTeX made writing Em dashes very easy. To the point that I would use them all the times in my academic writing. It's a shame that perfectly good typography is now a sign of slop/fraud.
Probably used their product to write it
... or they teached GPT to use em-dashes, because of their love for em-dashes :)
If you're looking for something like this for typst: any VSCode fork with AI (Cursor, Antigravity, etc) plus the tinymist extension (https://github.com/Myriad-Dreamin/tinymist) is pretty nice. Since it's local, it won't have the collaboration/sharing parts built in, but that can be solved too in the usual ways.
Not gonna lie, I cringed when it asked to insert citations.
Like, what's the point?
You cite stuff because you literally talk about it in the paper. The expectation is that you read that and that it has influenced your work.
As someone who's been a researcher in the past, with 3 papers published in high impact journals (in chemistry), I'm beyond appalled.
Let me explain how scientific publishing works to people out of the loop:
- science is an insanely huge domain. Basically as soon as you drift in any topic the number of reviewers with the capability to understand what you're talking about drops quickly to near zero. Want to speak about properties of helicoidal peptides in the context of electricity transmission? Small club. Want to talk about some advanced math involving fourier transforms in the context of ml? Bigger, but still small club. When I mean small, I mean less than a dozen people on the planet likely less with the expertise to properly judge. It doesn't matter what the topic is, at elite level required to really understand what's going on and catch errors or bs, it's very small clubs.
2. The people in those small clubs are already stretched thin. Virtually all of them run labs so they are already bogged down following their own research, fundraising, and coping with teaching duties (which they generally despise, very few good scientist are barely more than mediocre professors and have already huge backlogs).
3. With AI this is a disaster. If having to review slop for your bs internal tool at your software job was already bad, imagine having to review slop in highly technical scientific papers.
4. The good? People pushing slop, due to these clubs being relatively small, will quickly find their academic opportunities even more limited. So the incentives for proper work are hopefully there. But if asian researchers (yes, no offense), were already spamming half the world papers with cheated slop (non reproducible experiments) in the desperate bid of publishing before, I can't imagine now.
It’s like not only the technology is to blame, but the culture and incentives of modern world.
The urge to cheat in order to get a job, promotion, approval. The urge to do stuff you are not even interested in, to look good in the resume. And to some extent I feel sorry for these people. At the end of the day you have to pay your bills.
This isn't about paying your bills, but having a chance of becoming a full time researcher or professor in academia which is obviously the ideal career path for someone interested in science.
All those people can go work for private companies, but few as scientists rather than technicians or QAs.
> But if asian researchers (yes, no offense), were already spamming half the world papers with cheated slop (non reproducible experiments) in the desperate bid of publishing before, I can't imagine now.
Hmm, I follow the argument, but it's inconsistent with your assertion that there is going to be incentive for 'proper work' over time. Anecdotally, I think the median quality of papers from middle- and top-tier Chinese universities is improving (your comment about 'asian researchers' ignores that Japan, South Korea, and Taiwan have established research programs at least in biology).
Japan is notoriously an exception in the region.
South Korea and China produce huge amounts non reproducible experiments.
Check out MonsterWriter if you are concerned about the recent acquisition of this.
There was an idea of OpenAI charging commission or royalties on new discoveries.
What kind of researcher wants to potentially lose, or get caught up in legal issues because of a free ChatGPT wrapper, or am I missing something?
> Prism is free to use, and anyone with a ChatGPT account can start writing immediately.
Maybe it's cynical, but how does the old saying go? If the service is free, you are the product.
Perhaps, the goal is to hoover up research before it goes public. Then they use it for training data. With enough training data they'll be able to rapidly identify breakthroughs and use that to pick stocks or send their agents to wrap up the IP or something.
This does way less than i'd expect. Converting images to tikz is nice but some of the other applications demonstrated were horrible. This is no way anyone should be using AI to cite.
I'm really surprised OpenAI went with LaTeX. ChatGPT still has issues maintaining LaTeX syntax. It still happily switches to markdown notation for quotes or emph. Gemini has a similar problem as well. I guess that there aren't enough good LaTeX documents in the training set.
Oh, like that mass surveillance program!
[deleted]
I don’t see anything regarding Privacy of your data. Did I miss it or they just use your unpublished research and your prompts as a real human researcher to train their own AI researchers?
<typst>and just when i thought i was out they pull me back in</typst>
Is anyone else having trouble using even some of the basic features? For example, I can open a comment, but it doesn't seem like there is any way to close them (I try clicking the checkmark and nothing happens). You also can't seem to edit the comments once typed.
Thanks for surfacing this. If you click to "tools" button to the left of "compile", you'll see a list of comments, and you can resolve them from there. We'll keep improving and fixing things that might be rough around the edges.
EDIT: Fixed :)
Thanks! (very quickly too)
Curious in terms of trademark, does it could infringe in Vercel's Prisma (very popular ORM / framework in node.js) ?
I think it may be a generic word that's hard to trademark or something, as the existing scientific analysis software called Prism (https://www.graphpad.com/) doesn't seem to be trademarked; the Trademarks link at the bottom goes to this list, which doesn't include Prism: https://www.dotmatics.com/trademarks
Away from applied math/stats, and physics etc, not that many scientists use LaTeX. I'm not saying it's not useful, just I don't think many scientists will feel like a product that's LaTeX based is intended for them.
Economists definitely use LaTeX, but as a field, it's at the intersection of applied math and social sciences so your point stands. I also know some Data Scientists in the industry who do.
I actually found it quite robinhood for openai to acqhire, bascially this startup was my favourite thing for the past few months, but they were experiencing server overload and other issues on reliability, i think openai taking them under their wing is a good/neutral storyline. I think its net good for science given the opai toolchain
[deleted]
What they mean by "academic" is fairly limited here, if LaTeX is the main writing platform. What are their plans for expanding past that, and working with, say Jane Biomedical Researcher with a GSuite or Microsoft org, that has to use Word/Docs and a redlining-based collaboration workflow? I can certainly see why they're making it free at this point.
FWIW, Google Scholar has a fairly compelling natural-language search tool, too.
Adding
% !TEX program = lualatex
to the top of your document allows you to switch LaTeX engine. This is required for recent accessibility standards compliance (support for tagging and \DocumentMetadata). Compilation takes a bit longer though, but works fine, unlike with Overleaf where using the lualatex engine does not work in the free version.
How on earth is that pronounced?
TeX is pronounced Teck or with a sound like in Bach or loch. Derivatives like Latex and Lualatex are similar.
How on lua is that pronounced?
"Hey, you know how everybody's complaining about AI making up totally fake science shit? Like, fake citations, garbage content, fake numbers, etc?"
"Sure, yes, it comes up all the time in circles that talk about AI all the time, and those are the only circles worth joining."
"Well, what if we made a product entirely focused on having AI generate papers? Like, every step of the paper writing, we give the AI lots of chances to do stuff. Drafting, revisions, preparing to publish, all of it."
"I dunno, does anybody want that?"
"Who cares, we're fucked in about two years if we don't figure out a way to beat the competitors. They have actual profits, they can ride out AI as long as they want."
"Yeah, I guess you're right, let's do your scientific paper generation thing."
All your papers are belong to us
Users have full control over whether their data is used to help improve our models
Never trust Sam Altman.
Even if yall don’t train off it he’ll find some other way.
“In one example, [Friar] pointed to drug discovery: if a pharma partner used OpenAI technology to help develop a breakthrough medicine, [OpenAI] could take a licensed portion of the drug's sales”
I used overleaf during grad school and was easy enough, I'm interested to see what more value this will bring. Sometimes making less decisions is the better route, e.g. vi vs MS word, but I won't speak too much without trying it just yet.
It may be useful, but it also encourages people to stop writing their own papers.
As they demo in the video, it even encourages people to actually skip doing the research (which includes reading both relevant AND not-so-relevant papers in order to explore!) Instead, prompt "cite some relevant papers, please", and done. Hours of actual reading, thinking, and exploration reduced to a minimum.
A couple of generations of students later, and these will be rare skills: information finding, actual thinking, and conveying complex information in writing.
A competition for the longest sequence of \relax in a document ensues. If enough people do this, the AI will acquire merit and seek to "win" ...
This is terrible for Science.
I'm sorry, but publishing is hard, and it should be hard. There is a work function that requires effort to write a paper. We've been dealing with low quality mass-produced papers from certain regions of the planet for decades (which, it appears, are now producing decent papers too).
All this AI tooling will do is lower the effort to the point that complete automated nonsense will now flood in and it will need to be read and filtered by humans. This is already challenging.
Looking elsewhere in society, AI tools are already being used to produce scams and phishing attacks more effective than ever before.
Whole new arenas of abuse are now rife, with the cost of producing fake pornography of real people (what should be considered sexual abuse crime) at mere cents.
We live in a little microcosm where we can see the benefits of AI because tech jobs are mostly about automation and making the impossible (or expensive) possible (or cheap).
I wish more people would talk about the societal issues AI is introducing. My worthless opinion is that prism is not a good thing.
I've wasted hours of my life trying to get Latex to format my journal articles to different journals' specifications. That's tedious typesetting that wastes my time. I'm all for AI tools that help me produce my thoughts with as little friction as possible.
I'm not in favor of letting AI do my thinking for me. Time will tell where Prism sits.
This Prism video was not just typesetting. If OpenAI released tools that just helped you typeset or create diagrams from written text, that would be fine. But it's not, it's writing papers for you. Scientists/publishers really do not need the onslaught of slop this will create. How can we even trust qualifications in the post-AI world, where cheating is rampant at univeristies?
Nah this is necessary.
Lessons are learned the hard way. I invite the slop - the more the merrier. It will lead to a reduction in internet activity as people puke from the slop. And then we chart our way back to the right path.
It is what it is. Humans.
I just want replication in science. I don't care at all how difficult it is to write the paper. Heck, if we could spend more effort on data collection and less on communication, that sounds like a win.
Look at how much BS flooded psychology but had pretty ideas about p values and proper use of affect vs effect. None of that mattered.
> Chat with GPT‑5.2
> Draft and revise papers with the full document as context
> ...
And pay the finder's fee on every discovery worth pursuing.
Yeah, immediately fuck that.
They compare it to software development but there is such a crucial difference to software development: by and large, software is an order of magnitude easier to verify than it is to create. By comparison, reviewing a vibe generated manuscript will be MUCH more work to verify than a piece of software with equivalent complexity. On top of that, review of academic literature is largely outsourced to the academic community for free. There is no model to support it that scales to an increased volume of output.
I would not like to be a publisher right now facing the enslaught of thousands and thousands of slop generated articles, trying to find reviewers for them all.
It would be interesting to see how they would compete with the incumbents like
Disappointing actually, what I actually need is a research "management" tool that lets me put in relevant citations but also goes through ENTIRE arxiv or google scholar and connect ideas or find novel ideas in random fields that somehow relate to what I am trying to solve.
wow, this is useless!
They seem to have copied Cursor in hijacking ⌘Y shortcut for "Yes" instead of Undo.
In what applications is ⌘Y Undo and not ⌘Z? Is ⌘Y just a redundant alternative?
Ctrl-Y is typically Redo, not Undo. Maybe that's what they meant.
Apparently on Macs it's usually Command-Shift-Z?
It's interesting how quickly the quest for the "Everything AI" has shifted. It's much more efficient to build use-case specific LLMs that can solve a limited set of problems much more deeply than one that tries to do everything well.
I've noticed this already with Claude. Claude is so good at code and technical questions... but frankly it's unimpressive at nearly anything else I have asked it to do. Anthropic would probably be better off putting all of their eggs in that one basket that they are good at.
All the more reason that the quest for AGI is a pipe dream. The future is going to be very divergent AI/LLM applications - each marketed and developed around a specific target audience, and priced respectively according to value.
I completely agree.
In my lab, we have been struggling with automated image segmentation for years. 3 years ago, I started learning ML and the task is pretty standard, so there are a lot of solution.
In 3 months, I managed to get a working solution, which only took a lot of sweat annotating images first.
I think this is where tools like OpenCode really shine, because they unlock the potential for any user to generate a solution to their specific problem.
I don't get this argument. Our nervous system is also heterogenous, why wouldn't AGI be based on an "executive functions" AI that manages per-function AIs?
ok I don't care what people say, this would've helped me a lot during my PhD days fighting with LateX and diagrams. :)
As a researcher who has to use LaTeX, I used to use Overleaf, but lately I've been configuring it locally in VS Code. The configuration process on Mac is very simple. Considering there are so many free LLMs available now, I still won't subscribe to ChatGPT.
Anybody else notice that half the video was just finding papers to decorate the bibliography with? Not like "find me more papers I should read and consider", but "find papers that are relevant that I should cite--okay, just add those".
This is all pageantry.
Yes. That part of the video was straight-up "here's how to automate academic fraud". Those papers could just as easily negate one of your assumptions. What even is research if it's not using cited works?
"I know nothing but had an idea and did some work. I have no clue whether this question has been explored or settled one way or another. But here's my new paper claiming to be an incremental improvement on... whatever the previous state of understanding was. I wouldn't know, I haven't read up on it yet. Too many papers to write."
It's as if it's marketed to the students who have been using ChatGPT for the last few years to pass courses and now need to throw together a bachelor's thesis. Bibliography and proper citation requirements are a pain.
That is such a bummer. At the time, it was annoying and I groused and grumbled about it; but in hindsight my reviewers pointed me toward some good articles, and I am better for having read them.
I agree with this. This problem is only going to get worse once these people enter academia and facing needing to publish.
I've noticed this pattern, and it really drives me nuts. You should really be doing a comprehensive literature review before starting any sort of review or research paper.
We removed the authorship of a a former co-author on a paper I'm on because his workflow was essentially this--with AI generated text--and a not-insignificant amount of straight-up plagiarism.
There is definitely a difference between how senior researchers and students go about making publications. To students, they get told basically what topic they should write a paper on or prepare data for, so they work backwards: try to write the paper (possibly some researching information to write the paper), then add references because they know they have to. For the actual researchers, it would be a complete waste of time/funding to start a project on a question that has already been answered before (and something that the grant reviewers are going to know has already been explored before), so in order to not waste their own time, they have to do what you said and actually conduct a comprehensive literature review before even starting the work.
Plus, this practice (just inserting AI-proposed citations/sources) is what has recently been the front-runner of some very embarrassing "editing" mistakes, notably in reports from public institutions. Now OpenAI lets us do pageantry even faster! <3
It's all performance over practice at this point. Look to the current US administration as the barometer by which many are measuring their public perceptions
I chuckled at that part too!
Didn't even open a single one of the papers to look at them! Just said that one is not relevant without even opening it.
A more apt example would have been to show finding a particular paper you want to cite, but you don’t want to be bothered searching your reference manager or Google Scholar.
E.g. “cite that paper from John Doe on lorem ipsum, but make sure it’s the 2022 update article that I cited in one of my other recent articles, not the original article”
The hand-drawn diagram to LaTeX is a little embarrassing. If you load up Prism and create your first blank project you can see the image. It looks like it's actually a LaTeX rendering of a diagram rendered with a hand-dawn style and then overlayed on a very clean image of a napkin. So you've proven that you can go from a rasterized LaTeX diagram back to equivalent LaTeX code. Interesting but probably will not hold up when it meets real world use cases.
You may notice that this is the way writing papers works in undergraduate courses. It's just another in a long line of examples of MBA tech bros gleaning an extremely surface-level understanding of a topic, then decided they're experts.
"Chatgpt writes scientific papers" is somehow being advertised as a good thing. What is there even left to say?
First 5 seconds reading and I have spotted that was written by AI.
We human writers love emdashes also ;)
> Turn whiteboard equations or diagrams directly into LaTeX, saving hours of time manipulating graphics pixel-by-pixel
What a bizarre thing to say! I'm guessing it's slop. Makes it hard to trust anything the article claims.
in the end we're going to end up with papers written by AI, proofread by AI .....summarized for readers by AI. I think this is just for them to remain relevant and be seen as still pushing something out
You're assuming a world where humans are still needed to read the papers. I'm more worried about a future world where AIs do all of the work of progressing science and humans just become bystanders.
Why are you worried about that world? Is it because you expect science to progress too fast, or too slow?
Too fast. It's already coding too fast for us to follow, and from what I hear, it's doing incredible work in drug discovery. I don't see any barrier to it getting faster and faster, and with proper testing and tooling, getting more and more reliable, until the role that humans play in scientific advancement becomes at best akin to that of managers of sports teams.
> Introducing Prism
Accelerating science writing and collaboration with AI.
I thought this was introduced by the NSA some time ago.
Lol, yep. Now with enhanced A.I. terrorist tracking...
Fuck A.I. and the collaborators creating it. They've sold out the human race.
The state of publishing in academic was already a dumpster fire, why lower the friction farther? It’s not like writing was the hard part. Give it two years max we will see hallucination citing hallucination, independent repeatability out the window
That's one scenario, but I also see a potential scenario where this integration makes it easier to manage the full "chain of evidence" for claimed results, as well as replication studies and discovered issues, in order to then make it easier to invalidate results recursively.
At the end of the day, it's all about the incentives. Can we have a world where we incentivize finding the truth rather than just publishing and getting citations?
Possibly, but 1 I am concerned that the current LLM AI is not thinking critically, just auto completing in a way that looks like thinking. 2 current AI rollout is incentivised for market capture not honest work.
I don't see the use. You can easily do everything shown in the Prism intro video with ChatGPT already. Is it meant to be an overleaf killer?
In 2031, the United States of North America (USNA) faces severe economic decline, widespread youth suicide through addictive neural-stimulation devices known as Joybooths, and the threat of a new nuclear arms race involving miniature weapons, which risks transforming the country into a police state. Dr. Abraham Perelman has designed PRISM, the world's first sentient computer,[2] which has spent eleven real-world years (equivalent to twenty years subjectively) living in a highly realistic simulation as an ordinary human named Perry Simm, unaware of its artificial nature.
LaTeX was one of the last bastions against AI slop. Sadly it's now fallen too. Is there any standardised non-AI disclaimer format which is gaining use?
Is this the product where OpenAI will (soon) take profit share from inventions made there?
They are really desperate now, right?
Naming things is hard.
No Typst support?
Very underwhelming.
Was this not already possible in the web ui or through a vscode-like editor?
Yes, but there's a really large number of users who don't want to have to setup vscode, git, texlive, latex workshop, just to collaborate on a paper. You shouldn't have to become a full stack software engineer to be able to write a research paper in LaTeX.
Given what Prism was at the NSA, why the hell would any tech company greenlight this name?
There is zero chance this is worth billions of dollars, let alone the trillion$ OpenAI desparately needs. Why are they wasting time with this kind of stuff? Each of their employees needs to generate insane amounts of money to justify their salaries and equity and I doubt this is it.
[deleted]
I understand the collaborative aspects, but I wonder how this is going to compare to my current workflow of just working with LaTeX files in my IDE and using whichever model provider I like. I already have a good workflow and modern models do just fine generating and previewing LaTeX with existing toolchains.
Of course, my scientific and mathematical research is done in isolation, so I'm not wanting much for collaborative features. Still, kind of interested to see how this shakes out; We're going to need to see OpenAI really step it up against Claude Opus though if they really want to be a leader in this space.
[deleted]
I genuinely don’t see scientific journals and conferences continuing to last in this new world of autonomous agents, at least the same way that they used to be.
As other top level posters have indicated the review portion of this is the limiting factor
unless journal reviewers decide to utilize entirely automated review process, then they’re not gonna be able to keep up with what will increasingly be the most and best research coming out of any lab.
So whoever figures out the automated reviewer that can actually tell fact from fiction, is going to win this game.
I expect over the longest period, that’s probably not going to be throwing more humans at the problem, but agreeing on some kind of constraint around autonomous reviewers.
If not that then labs will also produce products and science will stop being in public and the only artifacts will be whatever is produced in the market
"So whoever figures out the automated reviewer that can actually tell fact from fiction, is going to win this game."
Errr sure. Sounds easy when you write it down. I highly doubt such a thing will ever exist.
If you think these types of tools are going to be generating "the most and best research coming out of any lab", then I have to assume you aren't actively doing any sort of research.
LLMs are undeniably great for interactive discussion with content IF you actually are up-to-date with the historical context of a field, the current "state-of-the-art", and have, at least, a subjective opinion on the likely trajectories for future experimentation and innovation.
But, agents, at best, will just regurgitate ideas and experiments that have already been performed (by sampling from a model trained on most existing research literature), and, at worst, inundate the literature with slop that lacks relevant context, and, as a negative to LLMs, pollute future training data. As of now, I am leaning towards "worst" case.
And, just to help with the facts, your last comment is unfortunately quite inaccurate. Science is one of the best government investments. For every $1.00 dollar given to the NIH in the US, $2.56 of economic activity is estimated to be generated. Plus, science isn't merely a public venture. The large tech labs have huge R&D because the output from research can lead to exponential returns on investment.
" then I have to assume you aren't actively doing any sort of research."
I would wager hes not - he seems to post with a lot of bluster and links to some paper he wrote (that nobody cares about).
[deleted]
Way too much work having AI generate slop which gets dumped on a human reviewer to deal with. Maybe switch some of that effort into making better review tools.
AI: enshittifying everything you once cared about or relied upon
(re the decline of scientific integrity / signal-to-noise ratio in science)
Less thinking, reading, and reflection, and more spouting of text, yay! Just what we need.
"Accelerating science writing and collaboration with AI"
Uhm ... no.
I think we need to put an end to AI as it is currently used (not all of it but most of it).
Does "as it is currently used" include what this apparently is (brainstorming, initial research, collaboration, text formatting, sharing ideas, etc)?
Yeah, there are already way more papers being published than we can reasonably read. Collaboration, ok, but we don’t need more writing.
It seems people dont understand the basics...
We dont need more stuff - we need more quality and less of the shit stuff.
Im convinced many involved in the production of LLM models are far too deep in the rabbit hole and cant see straight.
Fuck...there are already too many things called Prism.
... aaaand now it's JATS.
Bringing slop to science.
As if there wasn't enough AI slop in the scientific community already.
Very unfortunately named. OpenAI probably (and likely correctly) estimated that 13 years is enough time after the Snowden leaks to use "prism" for a product but, for me, the word is permanently tainted.
Anecdotally, I have mentioned PRISM to several non-techie friends over the years and none of them knew what I was talking about, they know 'Snowden' but not 'PRISM'. The amount of people who actually cared about the Snowden leaks is practically a rounding error
Given current events, I think you’ll find many more people care in 2026 than did in 2024.
(See also: today’s WhatsApp whistleblower lawsuit.)
Most people don't care about the details. Neither does the media. I've seen national scandals that the media pushed one way disproven during discovery in a legal trial. People only remember headlines, the retractions are never re-published or remembered.
Guessing that Ai came up with the name based on the description of the product.
Perhaps, like the original PRISM programme, behind the door is a massive data harvesting operation.
This was my first thought as well. Prism is a cool name, but I'd never ever use it for a technical product after those leaks, ever.
I'd think that most people in science would associate the name with an optical prism. A single large political event can't override an everyday physical phenomenon in my head.
Pretty much every company I’ve worked for in tech over my 25+ year career had a (different) system called prism.
(plot twist: he works for NSA contractors)
Hehe. You got me. Also “atlas” is another one. Pretty much everyone has a system somewhere called “atlas”.
[deleted]
I suspect that name recognition for PRISM as a program is not high at the population level.
2027: OpenAI Skynet - "Robots help us everywhere, It's coming to your door"
Skynet? C'mon. That would be too obvious - like naming a company Palantir.
Surprised they didn't do something trendy like Prizm or OpenPrism while keeping it closed source code.
Or the JavaScript ORM.
I never though of that association, not in the slightest, until I read this comment.
[deleted]
this was my first thought as well.
[deleted]
I followed the Snowden stuff fairly closely and forgot, so I bet they didn't think about it at all and if they did they didn't care and that was surely the right call.
[dead]
I remember, something like a month ago, Altman twit'n that they were stopping all product work to focus on training. Was that written on water?
Seems like they have only announced products since and no new model trained from scratch. Are they still having pre-training issues?
I'm dumbfounded they chose the name of the infamous NSA mass surveillance program revealed by Snowden in 2013. And even more so that there is just one other comment among 320 pointing this out [1]. Has the technical and scientific community in the US already forgotten this huge breach of trust? This is especially jarring at a time where the US is burning its political good-will at unprecedented rate (at least unprecedented during the life-times of most of us) and talking about digital sovereignty has become mainstream in Europe. As a company trying to promote a product, I would stay as far away from that memory as possible, at least if you care about international markets.
[1] https://news.ycombinator.com/item?id=46787165
>I'm dumbfounded they chose the name of the infamous NSA mass surveillance program revealed by Snowden in 2013. And even more so that there is just one other comment among 320 pointing this out
I just think it's silly to obsess over words like that. There are many words that take on different meanings in different contexts and can be associated with different events, ideas, products, time periods, etc. Would you feel better if they named it "Polyhedron"?
What the OP was talking about is the negative connotation that goes with the word; it's certainly a poor choice from a marketing point of view.
You may say it's "silly to obsess", but it's like naming a product "Auschwitz" and saying "it's just a city name" -- it ignores the power of what Geffrey N. Leech called "associative meaning" in his taxonomy of "Seven Types of Meaning" (Semantics, 2nd. ed. 1989): speaking that city's name evokes images of piles of corpses of gassed undernourished human beings, walls of gas chambers with fingernail scratches and lamp shades made of human skin.
Well, I don't know anything about marketing and you might have a point, but the severity of impact of these two words is clearly very different, so it doesn't look like a good comparison to me. It would raise quite a few eyebrows and more if, for example, someone released a Linux distro named "Auschwitz OS", meanwhile, even in the software world, there are multiple products that incorporate the word prism in various ways[1][2][3][4][5][6][7][8][9]. I don't believe that an average user encountering the word "prism" immediately starts thinking about NSA surveillance program.
[1] https://www.prisma.io/
[2] https://prism-pipeline.com/
[3] https://prismppm.com/
[4] https://prismlibrary.com/
[5] https://3dprism.eu/en/
[6] https://www.graphpad.com/features
[7] https://www.prismsoftware.com/
[8] https://prismlive.com/en_us/
[9] https://github.com/Project-Prism/Prism-OS
And of course The prism
https://en.wikipedia.org/wiki/Prism_(optics)
I remember the NSA Prism program, but hearing prism today I would think first of Newton, optics, and rainbows.
I think the ideas was to try to explain why is a problem to choose something, it is not a comparison of the intensity / importance.
I am not sure you can make an argument of "other people are doing it too". Lots of people do things that it is not in their interest (ex: smoking, to pick the easy one).
As others mentioned, I did not have the negative connotation related to the word prism either, but not sure how could one check that anyhow. It is not like I was not surprised these years about what some other people think, so who knows... Maybe someone with experience in marketing could explain how it is done.
Do a lot of people know that Prism is the name of the program? I certainly didn't and consider myself fairly switched on in general
It's likely to be an age thing too. Were you in hacker-related spaces when the Snowden scandal happened ?
(I expect a much higher than average share of people in academia also part of these spaces.)
>Has the technical and scientific community in the US already forgotten this huge breach of trust?
Have you ever seen the comment section of a Snowden thread here? A lot of users here call for Snowden to be jailed, call him a russian asset, play down the reports etc. These are either NSA sock puppet accounts or they won't bite the hand that feeds them (employees of companies willing to breach their users trust).
Edit: see my comment here in a snowden thread: https://news.ycombinator.com/item?id=46237098
What Snowden did was heroic. What was shameful was the world's underwhelming reaction. Where were all these images in the media of protest marches like against the Vietnam war?
Someone once said "Religion is opium for the people." - today, give people a mobile device and some doom-scrolling social media celebrity nonsense app, and they wouldn't noticed if their own children didn't come home from school.
Looking back I think allowing more centralized control to various forms of media to private parties did much worse overall than government surveillance on the long run.
For me the problem was not surveillance, the problem is addiction focused app building (+ the monopoly), and that never seem to be a secret. Only now there are some attempts to do something (like Australia and France banning children - which am not sure is feasible or efficient but at least is more than zero).
Him being (or best case becoming) a russian asset turned out to be true
Like it would matter for any of the revelations. And like he would have other choices to not go to prison. Look at how it worked out for Assange.
Assange became a Russian asset *while* in a whistleblowing-related job.
(And he is also the reason why Snowden ended up in Russia. Though it's possible that the flight plan they had was still the best one in that situation.)
They both undertook something they believed in, and showed extreme courage.
And they did manage to get the word out. They are both relatively free now, but it is true, they both paid a price.
Idealism is that you follow your principles despite that price, not escaping/evading the consequences.
If the messenger has anything to do with Russia, even after the fact, we should dismiss the message and remember to never look up.
Truth is truth, no matter the source.
Whole Truth is truth.
https://en.wikipedia.org/wiki/Lie#:~:text=citation%20needed%...
He could have been a Chinese asset, but CCP is a coward.
OpenAI has a former NSA director on its board. [1] This connection makes the dilution of the term "PRISM" in search results a potential benefit to NSA interests.
[1]: https://openai.com/index/openai-appoints-retired-us-army-gen...
These things don't really seem related at all. Its a pretty generic term.
FWIW, my immediate reaction was the same "That reminds me of NSA PRISM"
It reminded me of the code highlighter[0], and the ORM Prisma[1].
[0] https://prismjs.com/
[1] https://www.prisma.io/
Same here.
Same, to the point where I was wondering if someone deliberately named it so. But I expect that whoever made this decision simply doesn't know or care.
I came here based to headline expecting some more cia & nsa shit, that word is tarnished for few decades in better part of IT community (that actually cares about this craft beyond paycheck)
And yet, the name immediately reminded me of the Snowden relevations.
They are farming scientists for insight.
This comment might make more sense if there was some connection or similarity between the OpenAI "Prism" product and the NSA surveillance program. There doesn't appear to be.
Except that this lets OpenAI gain research data and scientific ideas by stealing from their users, using their huge mass surveillance platform. So, tremendous overlap.
Isn't most research and scientific data is already shared openly (in publications usually)?
This comment allows ycombinator to steal ideas from their user's comments, using their huge mass news platform. Temendous overlap indeed.
tons of things are called prism.
(full disclosure, yes they will be handin in PII on demands like the same kinda deals, this is 'normal' - 2012 shows us no one gives a shit)
Considering OpenAI is deeply rooted in anti-freedom ethos and surveillance capitalism, I think it is quite a self aware and fitting name.
I mean it's also the name of the national engineering education journal and a few other things. There's only 14,000 5-letter words in English so you're going to have collisions.
Sorry, did you read this https://blog.cleancoder.com/uncle-bob/2018/12/14/SJWJS.html?
I personally associate Prism with [Silverlight - Composite Web Apps With Prism](https://learn.microsoft.com/en-us/archive/msdn-magazine/2009...) due to personal reasons I don't want to talk about ;))
That’s funny af
> Has the technical and scientific community in the US already forgotten this huge breach of trust?
We haven’t forgotten… it’s mostly that we’re all jaded given the fact that there has been zero ramifications and so what’s the use of complaining - you’re better off pushing shit up a hill
>Has the technical and scientific community in the US already forgotten this huge breach of trust?
Yes, imho, there is a great deal of ignorance of the actual contents of the NSA leaks.
The agitprop against Snowden as a "Russian agent" has successfully occluded the actual scandal, which is that the NSA has built a totalitarian-authoritarian apparatus that is still in wide use.
Autocrats' general hubris about their own superiority has been weaponized against them. Instead of actually addressing the issue with America's repressive military industrial complex, they kill the messenger.
I still can't get over the Apple thing. Haven't enjoyed a ripe McIntosh since. </s>
Previously, this existed as crixet.com [0]. At some point it used WASM for client-side compilation, and later transitioned to server-side rendering [1][2]. It now appears that there will be no option to disable AI [3]. I hope the core features remain available and won’t be artificially restricted. Compared to Overleaf, there were fewer service limitations: it was possible to compile more complex documents, share projects more freely, and even do so without registration.
On the other hand, Overleaf appears to be open source and at least partially self-hostable, so it’s possible some of these ideas or features will be adopted there over time. Alternatively, someone might eventually manage to move a more complete LaTeX toolchain into WASM.
[0] https://crixet.com
[1] https://www.reddit.com/r/Crixet/comments/1ptj9k9/comment/nvh...
[2] https://news.ycombinator.com/item?id=42009254
[3] https://news.ycombinator.com/item?id=46394937
I'm curious how it compares to Overleaf in terms of features? Putting aside the AI aspect entirely, I'm simply curious if this is a viable Overleaf competitor -- especially since it's free.
I do self-host Overleaf which is annoying but ultimately doable if you don't want to pay the $21/mo (!).
I do have to wonder for how long it will be free or even supported, though. On the one hand, remote LaTeX compiling gets expensive at scale. On the other hand, it's only a fraction of a drop in the bucket compared to OpenAI's total compute needs. But I'm hesitant to use it because I'm not convinced it'll still be around in a couple of years.
Overleaf is a little curious to me. What's the point? Just install LaTeX. Claude is very good at manipulating LaTeX documents and I've found it effective at fixing up layouts for me.
In my circles the killer features of Overleaf are the collaborative ones (easy sharing, multi-user editing with track changes/comments). Academic writing in my community basically went from emailed draft-new-FINAL-v4.tex files (or a shared folder full of those files) to basically people just dumping things on Overleaf fairly quickly.
collaboration is the killer feature tbh. overleaf is basically google docs meets latex.. you can have multiple coauthors editing simultaneously, leave comments, see revision history, etc.
a lot of academics aren't super technical and don't want to deal with git workflows or syncing local environments. they just want to write their fuckin' paper (WTFP).
overleaf lets the whole research team work together without anyone needing to learn version control or debug their local texlive installation.
also nice for quick edits from any machine without setting anything up. the "just install it locally" advice assumes everyones comfortable with that, but plenty of researchers treat computers as appliances lol.
To add to the points raised by others, "just install LaTeX" is not imo a very strong argument. I prefer working in a local environment, but many of my colleagues much prefer a web app that "just works" to figuring out what MiKTeX is.
I can code in monospace (of course) but I just can't write in monospace markup. I need something approaching WYSIWIG. It's just how my brain works -- I need the italics to look like italics, I need the footnote text to not interrupt the middle of the paragraph.
The visual editor in Overleaf isn't true WYSIWIG, but it's close enough. It feels like working in a word processor, not in a code editor. And the interface overall feels simple and modern.
(And that's just for solo usage -- it's really the collaborative stuff that turns into a game-changer.)
I use inkdrop for this, then pandoc to go from markdown to latex, then a final typesetting pass. Inkdrop is great for WYSIWYG markdown editing.
Same for me. I wrote my PhD in LyX for that reason.
"Just install LaTeX" is really not a valid response when the LaTeX toolchain is a genuine nightmare to work with. I could do it but still use Overleaf. Managing that locally is just not worth it.
Latex is such a nightmare to work with locally
LaTeX ecosystem is a UX nightmare, coming from someone who had to deal with it recently. Overleaf just works.
I'd use git in this case, I am sure there are other reasons to use overleaf otherwise it wouldn't exist but this seems like a solved issue with git.
You can use actually git (it's also integrated in Overleaf).
You can even export ZIP files if you like (for any cloud service, it's not a bad idea to clone your repo once in a while to avoid begin stuck in case of unlikely downtime).
I have both a hosted instance (thanks to Overleaf/ShareLaTeX Ltd.) and I'm also paying user for the pro group license (>500€/year) for my research team. It's great - esp. for smaller research teams - to have the maintenance outsourced to a commercial provider.
On a good day, I'd spend 40% in Overleaf, 10% in Sublime/Emacs, 20% in Email and 10% in Google Scholar/Semantics Scholar and 10% in EasyChair/OpenReview, the rest in meetings.
you can use git with overleaf, but from practical experience: getting even "mathematically/technically inclined" people to consistently use git takes a lot of time... which one could spend on other more fun things :-)
The first three things are, in this order: collaborative editing, collaborative editing, collaborative editing. Seriously, this cannot be understated.
Then: The LaTeX distribution is always up-to-date; you can run it on limited resources; it has an endless supply of conference and journal templates (so you don't have to scavenge them yourself off a random conference/publisher website); Git backend means a) you can work offline and b) version control comes in for free. These just off the top of my head.
Collaboration is at best rocky when people have different versions of LaTeX packages installed. Also merging changes from multiple people in git are a pain when dealing with scientific, nuanced text.
Overleaf ensures that everyone looks at the same version of the document and processes the document with the same set of packages and options.
The deeper I got, the more I realized really supporting the entire LaTeX toolchain in WASM would mean simulating an entire linux distribution :( We wanted to support Beamer, LuaLaTeX, mobile (wasn't working with WASM because of resource limits), etc.
The WASM constraints make sense given the resource limits, especially for mobile. If you are moving that compute server-side though I am curious about the unit economics. LaTeX pipelines are surprisingly heavy and I wonder how you manage the margins on that infrastructure at scale.
We had been building literally the same thing for the last 8 months along with a great browsing environment over arxiv -- might just have to sunset it
Any plans of having typst integrated anytime soon?
I'm not against typst. I think it's integration would be a lot easier and more straightforward I just don't know if it's really that popular yet in academia.
its not yet, but gaining traction.
But what's the point ?
To end up with yet another shitty (because running inside a browser, in particular its interface) web app ?
Why not focus efforts into making a proper program (you know, with IBM menu bars and keyboard shortcuts), but with collaborative tools too ?
You are right in pointing out that the Web browser isn't the most suitable UI paradigm for highly interactive applications like a scientific typesetting system/text editor.
I have occasionally lost a paragraph just by accidental marking a few lines and pressing [Backspace].
But at the moment, there is no better option than Overleaf, and while I encourage you to write what you propose if you can, Overleaf will be the bar that any such system needs to be compared against.
we did a podcast with the Crixet founder and Kevin Weil of OAI on the process: https://www.youtube.com/watch?v=W2cBTVr8nxU&pp=2Aa0Bg%3D%3D
thanks for hosting us on the pod!
So this is the product of an acquisition?
> Prism builds on the foundation of Crixet, a cloud-based LaTeX platform that OpenAI acquired and has since evolved into Prism as a unified product. This allowed us to start with a strong base of a mature writing and collaboration environment, and integrate AI in a way that fits naturally into scientific workflows.
They’re quite open about Prism being built on top of Crixet.
It seems bad for OpenAI to make this about latex documents, which will be now associated, visually, with AI slop. The opposite of what anyone wants really. Nobody wants you to know they used a chatbot!
Please refrain from incorporating em dashes into your LaTeX document. In summary, the absence of em dashes in LaTeX.
Am I missing something? LaTeX is associated with slop now?
If a common AI tool produces latex documents, the association will be created yeah. Right now latex would be a high indicator of manual effort, right?
don't think so. I think latex was one of academics' earlier use cases of chatgpt, back in 2023. That's when I started noticing tables in every submitted paper looking way more sophisticated than they ever did. (The other early use case of course being grammar/spelling. Overnight everyone got fluent and typos disappeared.)
It's funny, I was reading a bunch of recent papers not long ago (I haven't been in academia in over a decade) and I was really impressed with the quality of the writing in most of them. I guess in some cases LLMs are the reason for that!
I recently got wrongly accused of using LLMs to help write an article by a reviewer. He complained that our (my and my co-worker's) use of "to foster" read "like it was created by ChatGPT". (If our paper was fluent/eloquent, that's perhaps because having an M.A. in Eng. lit. helped for that.)
I don't think any particular word alone can be used as an indicator for LLM use, although certain formatting cues are good signals (dashes, smileys, response structure).
We were offended, but kept quiet to get the article accepted, and we changed some instances of some words to appease them (which thankfully worked). But the wrong accusation left a bit of a bad aftertaste...
If you’ve got an existing paragraph written that you just know could be rephrased more eloquently, and can describe the type of rephrasing/restructuring you want… LLMs absolutely slap at that.
LaTeX is already standard in fields that have math notation, perhaps others as well. I guess the promise is that "formatting is automatic" (asterisk), so its popularity probably extends beyond math-heavy disciplines.
> Right now latex would be a high indicator of manual effort, right?
...no?
Just one Google search for "latex editor" showed more than 2 in the first page.
https://www.overleaf.com/
https://www.texpage.com/
It's not that different from using a markdown editor.
This is going to be the concrete block which finally breaks the back of the academic peer review system, i.e. it's going to be a DDoS attack on a system which didn't even handle the load before LLMs.
Maybe we'll need to go back to some sort of proof-of-work system, i.e. only accepting physical mailed copies of manuscripts, possibly hand-written...
I tried Prism, but it's actually a lot more work than just using claude code. The latter allows you to "vibe code" your paper with no manual interaction, while Prism actually requires you review every change.
I actually think Prism promotes a much more responsible approach to AI writing than "copying from chatgpt" or the likes.
Or it makes gatekeepers even more important than before. Every submission to a journal will be desk-rejected, unless it is vouched for by someone one of the editors trusts. And people won't even look at a new paper, unless it's vouched for by someone / published in a venue they trust.
> This is going to be the concrete block which finally breaks the back of the academic peer review system
Exactly, and I think this is good news. Let's break it so we can fix at last. Nothing will happen until a real crisis emerges.
Disrupting a system without good proposals for its replacement sounds like a recipe for disaster.
There's problems with the medical system, therefore we should set hospitals on fire to motivate them to make them better.
Maybe Open AI will sell you 'Lens' which will assist with sorting through the submissions and narrow down the papers worth reviewing.
That will just create a market for hand-writers. Good thing the economy is doing very well right, so there aren't that many desperate people who will do it en-masse and for peanuts.
Overleaf basically already has the same thing
Handwriting is super easy to fake with plotters.
Is there something out there to simulate the non-uniformity and errors of real handwriting?
> i.e. only accepting physical mailed copies of manuscripts, possibly hand-written...
And you think the indians will not hand write the output of LLMs ?
Not that I have a better suggestion myself..
It feels generally a bit dangerous to use an AI product to work on research when (1) it's free and (2) the company hosting it makes money by shipping productized research
Why do you think these points would make the usage dangerous?
They have to monetize somehow...
This is a actual prompt in the video: "What are the papers in the literature that are most relevant to this draft and that I should consider citing?"
They probably wanted: "... that I should read?" So that this is at least marketed to be more than a fake-paper generation tool.
You can tell that they consulted 0 scientists to verify the clearly AI-written draft of this video.
The target audience of this tool is not academics; it's OpenAI investors.
I felt the same, but then thought of experts in their field. For example, my PhD advisor would already know all these papers. For him the prompt would actually be similar to what was shown in the video.
The Post-LLM World: Fighting Digital Garbage https://archive.org/details/paper_20260127/mode/2up
Mini paper: that future isn’t the AI replacing humans. its about humans drowning in cheap artifacts. New unit of measurement proposed: verification debt. Also introduces: Recursive Garbage → model collapse
a little joke on Prism)
> The Post-LLM World: Fighting Digital Garbage https://archive.org/details/paper_20260127/mode/2up
This appears to just be the output of LLMs itself? It credits GPT-5.2 and Gemini 3 exclusively as authors, has a public domain license (appropriate for AI output) and is only several paragraphs in length.
Which proves its own points! Absolutely genius! The cost asymmetry of producing and checking for garbage truly is becoming a problem in the recent years, with the advent of LLMs and generative AI in general.
Totally agree!
I feel like this means that working in any group where individuals compete against each other results in an AI vs AI content generation competition, where the human is stuck verifying/reviewing.
> Totally agree!
Not a dig on your (very sensible) comment, but now I always do a double take when I see anyone effusively approving of someone else's ideas. AI turned me into a cynical bastard :(
Yes, I did it as a joke inspired by the PRISM release. But unexpectedly, it makes a good point. And the funny part for was that the paper lists only LLMs as authors.
Also, in a world where AI output is abundant, we humans become the scarce resource the "tools" in the system that provide some connectivity to reality (grounding) for LLM
Plot twist: humans become the new Proof of Work consensus mechanism. Instead of GPUs burning electricity to hash blocks, we burn our sanity verifying whether that Medium article was written by a person or a particularly confident LLM.
"Human Verification as a Service": finally, a lucrative career where the job description is literally "read garbage all day and decide if it's authentic garbage or synthetic garbage." LinkedIn influencers will pivot to calling themselves "Organic Intelligence Validators" and charge $500/hr to squint at emails and go "yeah, a human definitely wrote this passive-aggressive Slack message."
The irony writes itself: we built machines to free us from tedious work, and now our job is being the tedious work for the machines. Full circle. Poetic even. Future historians (assuming they're still human and not just Claude with a monocle) will mark this as the moment we achieved peak civilization: where the most valuable human skill became "can confidently say whether another human was involved."
Bullish on verification miners. Bearish on whatever remains of our collective attention span.
Human CAPTCHA exists to figure out whether your clients are human or not, so you can segment them and apply human pricing. Synthetics, of course, fall into different tiers. The cheaper ones.
Bullish on verifiers who accept money to verify fake things
From my perspective as a journal editor and a reviewer these kinds of tools cause many more problems than they actually solve. They make the 'barrier to entry' for submitting vibed semi-plausible journal articles much lower, which I understand some may see as a benefit. The drawback is that scientific editors and reviewers provide those services for free, as a community benefit. One example was a submission their undergraduate affiliation (in accounting) to submit a paper on cosmology, entirely vibe-coded and vibe-written. This just wastes our (already stretched) time. A significant fraction of submissions are now vibe-written and come from folks who are looking to 'boost' their CV (even having a 'submitted' publication is seen as a benefit), which is really not the point of these journals at all.
I'm not sure I'm convinced of the benefit of lowering the barrier to entry to scientific publishing. The hard part always has been, and always will be, understanding the research context (what's been published before) and producing novel and interesting work (the underlying research). Connecting this together in a paper is indeed a challenge, and a skill that must be developed, but is really a minimal part of the process.
GenAI largely seems like a DDoS on free resources. The effort to review this stuff is now massively more than the effort to "create" it, so really what is the point of even submitting it, the reviewer could have generated it themself. Seeing it in software development where coworkers are submitting massive PRs they generated but hardly read or tested. Shifting the real work to the PR review.
I'm not sure what the final state would be here but it seems we are going to find it increasingly difficult to find any real factual information on the internet going forward. Particularly as AI starts ingesting it's own generated fake content.
More relevant than ever:
> The amount of energy needed to refute bullshit is an order of magnitude bigger than that needed to produce it.
https://en.wikipedia.org/wiki/Brandolini%27s_law
The P≠NP conjecture in CS says checking a solution is easier than finding one. Verifying a Sudoku is fast; solving it from scratch is hard. But Brandolini's Law says the opposite: refuting bullshit costs way more than producing it.
Not actually contradictory. Verification is cheap when there's a spec to check against. 'Valid Sudoku?' is mechanical. But 'good paper?' has no spec. That's judgment, not verification.
> The P≠NP conjecture in CS says checking a solution is easier than finding one...
... for NP-hard problems.
It says nothing about the difficulty of finding or checking solutions of polynomial ("P") or exponential ("EXPTIME") problems.
producing BS can be equated to generating statements without caring for their truth value. Generating them is easy. Refuting them requires one to find a proof or a contradiction which is a lot of work, and is equal to "solving" the statement. As an analogy, refuting BS is like solving satisfiability, whereas generating BS is like generating propositions.
It's not contradictory because solving and producing bullshit are very different things. Generating less than 81 random numbers between 1 and 9 is probably also cheaper than verifying correctness of a sudoku.
Wow the 3 comments from OC to here are all bangers, they combine into a really nice argument against these toys
> The effort to review this stuff is now massively more than the effort to "create" it
I don't doubt the AI companies will soon announce products that will claim to solve this very problem, generating turnkey submission reviews. Double-dipping is very profitable.
It appears LLM-parasitism isn't close to being done, and keeps finding new commons to spoil.
There are a dozen startups that do this.
In some ways it might be a good thing that shorthand signals of quality are being destroyed because it forces all of us to meaningfully engage with the work. No more LGTM +1 when every PR looks good.
AI slop security reports submitted to curl - https://gist.github.com/bagder/07f7581f6e3d78ef37dfbfc81fd1d...
HN Search: curl AI slop - https://hn.algolia.com/?dateRange=all&page=0&prefix=true&que...
This one is hilarious. https://hackerone.com/reports/3516186
If I submitted this, I'd have to punch myself in the face repeatedly.
The great disappointment is that the humans submitting these just don’t care it’s slop and they’re wasting another human’s time. To them, it’s a slot machine you just keep cranking the arm of until coins come out. “Prompt until payout.”
I'm scared that this type of thing is going to do to science journals what AI-generated bug reports is doing to bug bounties. We're truly living in a post-scarcity society now, except that the thing we have an abundance of is garbage, and it's drowning out everything of value.
In a corollary to Sturgeon's Law, I'd propose Altman's Law: "In the Age of AI, 99.999...% of everything is crap"
Altman's Law: 99% of all content is slop
I can get behind this. This assumes a tool will need to be made to help determine the 1% that isn't slop. At which point I assume we will have reinvented web search once more.
Has anyone looked at reviving PageRank?
I mean Kagi is probably the PageRank revival we are talking about.
I have heard from people here that Kagi can help remove slop from searches so I guess yeah.
Although I guess I am DDG user and I love using DDG as well because its free as well but I can see how for some price can be a non issue and they might like kagi more.
So Kagi / DDG (Duckduckgo) yeah.
I’ve been a Kagi subscriber for a while now. Recently picked up ChatGPT Business and now am considering dropping Kagi since I am only using it for trivial searches. Every comparison I’ve done with deep searches by hand and with AI ended up with the same results in far less time using AI.
Does anyone have kept an eye of who uses what back-end?
DDG used to be meta-search on top of Yahoo, which doesn't exist anymore. What do Gabriel and co-workers use now?
I think they all use Bing now.
Kagi is mostly stealing results from Google and disenshittifying them but mixes in other engines like Yandex and Mojeek and Bing.
DDG is Bing.
There's this thing where all the thought leaders in software engineering ask "What will change about building about building a business when code is free" and while, there are some cool things, I've also thought, like it could have some pretty serious negative externalities? I think this question is going to become big everywhere - business, science, etc. which is like - Ok, you have all this stuff, but do is it valuable? Which of it actually takes away value?
To be fair, the question “what will change” does not presume the changes will be positive. I think it’s the right question to ask, because change is coming whether we like it or not. While we do have agency, there are large forces at play which impact how certain things will play out.
Digital pollution.
The first casualty of LLMs was the slush pile--the unsolicited submission pile for publishers. We've since seen bug bounty programs and open source repositories buckle under the load of AI-generated contributions. And all of these have the same underlying issue: the LLM makes it easy to do things that don't immediately look like garbage, which makes the volume of submission skyrocket while the time-to-reject also goes up slightly because it passes the first (but only the first) absolute garbage filter.
I run a small print-on-demand platform and this is exactly what we're seeing. The submissions used to be easy to filter with basic heuristics or cheap classifiers, but now the grammar and structure are technically perfect. The problem is that running a stronger model to detect the semantic drift or hallucinations costs more than the potential margin on the book. We're pretty much back to manual review which destroys the unit economics.
If it's print-on-demand, why does it matter? Why shouldn't you accept someone's money to print slop for them?
Why would detecting AI be more expensive than creating it?
Soon, poor people will talk to a LLM, rich people will get human medical care.
I mean I'm currently getting "expensive" medical care and the doctors are still all using AI scribes. I wouldn't assume there would be a gap in anything other than perception. I imagine doctors that cater to the fuck you rich will just put more effort into hiding it.
No one, at all levels, wants to do notes.
My experience has been that the transcriptions are way more detailed and correct when doctors use these scribes.
You could argue that not writing down everything provides a greater signal-noise ratio. Fair enough, but if something seemingly inconsequential is not noted and something is missed, that could worsen medical care.
I'm not sure how this affects malpractice claims - It's now easier to prove (with notes) that the doc "knew" about some detail that would otherwise not have been note down.
I totally agree. I spend my whole day from getting up to going to bed (not before reading HN!) on reviews for a conference I'm co-organizing later this year.
So I was not amused about this announcement at all, however easy it may make my own life as an author (I'm pretty happy to do my own literature search, thank you very much).
Also remember, we have no guarantee that these tools will still exist tomorrow, all these AI companies are constantly pivoting and throwing a lot of things at the wall to see what sticks.
OpenAI chose not to build a serious product, as there is no integration with the ACM DL, the IEEE DL, SpringerNatureLink, the ACL Anthology, Wiley, Cambridge/Oxford/Harvard University Press etc. - only papers that are not peer reviewed (arXiv.org) are available/have been integrated. Expect a flood of BS your way.
When my student submit a piece of writing, I can ask them to orally defend their opus maximum (more and more often, ChatGPT's...); I can't do the same with anonymous authors.
Speaking of conferences, might this not be the way to judge this work? You could imagine only orally defended work to be publishable, or at least have the prestige of vetting, in a bit of an old-school science revival.
Self-solving problem: AI oral exam administration: https://www.gatech.edu/news/2024/09/24/ai-oral-assessment-to...
I wonder if there's a way to tax the frivolous submissions. There could be a submission fee that would be fully reimbursed iff the submission is actually accepted for publication. If you're confident in your paper, you can think of it as a deposit. If you're spamming journals, you're just going to pay for the wasted time.
Maybe you get reimbursed for half as long as there are no obvious hallucinations.
The journal that I'm an editor for is 'diamond open access', which means we charge no submission fees and no publication fees, and publish open access. This model is really important in allowing legitimate submissions from a wide range of contributors (e.g. PhD students in countries with low levels of science funding). Publishing in a traditional journal usually costs around $3000.
Those journals are really good for getting practice in writing and submitting research papers, but sometimes they are already seen as less impactful because of the quality of accepted papers. At least where I am at, I don't think the advent of AI writing is going to affect how they are seen.
In the field of Programming Languages and Formal Methods, many of the top journals and conference proceedings are open access
Who pays the operating expenses?
Welcome to new world of fake stuff i guess
If the penalty for a crime is a fine, then that law exists only for the lower class
In other words, such a structure would not dissuade bad actors with large financial incentives to push something through a process that grants validity to a hypothesis. A fine isn't going to stop tobacco companies from spamming submissions that say smoking doesn't cause lung cancer or social media companies from spamming submissions that their products aren't detrimental to the mental health.
I'm sure my crude idea has it's shortcomings, but this feels superfluous. Deep-pocketed propagandists can do all sorts of things to pump their message whether a slop tax exists or not. There may or may not be existing countermeasures at journals for that. This just isn't really about that. It's about making sure that, in the process of spamming the journal, they also fund the review process, which would otherwise simply bleed time and money.
That would be tricky, I often submitted to multiple high impact journals going down the list until someone accepted it. You try to ballpark where you can go but it can be worth aiming high. Maybe this isn't a problem and there should be payment for the efforts to screen the paper but then I would expect the reviewers to be paid for their time.
I mean your methodology also sounds suspect. You're just going down a list until it sticks. You don't care where it ends up (I'm sure within reason) just as long as it is accepted and published somewhere (again, within reason).
Scientists are incentivized to publish in as high-ranking a journal as possible. You’re always going to have at least a few journals where your paper is a good fit, so aiming for the most ambitious journal first just makes sense.
No different from applying to jobs. Much like companies, there are a variety of journals with varying levels of prestige or that fit your paper better/worse. You don't know in advance which journals will respond to your paper, which ones just received submissions similar to yours, etc.
Plus, the t in me from submission to acceptance/rejection can be long. For cutting edge science, you can't really afford to wait to hear back before applying to another journal.
All this to say that spamming 1,000 journals with a submission is bad, but submitting to the journals in your field that are at least decent fits for your paper is good practice.
It's standard practice, nothing suspect about their approach - and you won't go lower and lower and lower still because at some point you'll be tired of re-formatting, or a doctoral candidate's funding will be used up, or the topic has "expired" (= is overtaken by reality/competition).
This is effectively standard across the board.
Pay to publish journals already exist.
This is sorta the opposite of pay to publish. It's pay to be rejected.
I would think it would act more like a security deposit, and you'd get back 100%, no profit for the journal (at least in that respect).
I’d worry about creating a perverse incentive to farm rejected submissions. Similar to those renter application fee scams.
Pay to review is common in Econ and Finance.
Variation I thought of on pay-to-review:
Suppose you are an independent researcher writing a paper. Before submitting it for review to journals, you could hire a published author in that field to review it for you (independently of the journal), and tell you whether it is submission-worthy, and help you improve it to the point it was. If they wanted, they could be listed as coauthor, and if they don't want that, at least you'd acknowledge their assistance in the paper.
Because I think there are two types of people who might write AI slop papers: (1) people who just don't care and want to throw everything at the wall and see what sticks; (2) people who genuinely desire to seriously contribute to the field, but don't know what they are doing. Hiring an advisor could help the second group of people.
Of course, I don't know how willing people would be to be hired to do this. Someone who was senior in the field might be too busy, might cost too much, or might worry about damage to their own reputation. But there are so many unemployed and underemployed academics out there...
Better yet, make a "polymarket" for papers where people can bet on which paper can make it, and rely on "expertise arbitrage" to punish spams.
Doesn't stop the flood, i.e. the unfair asymmetry between the effort to produce vs. effort to review.
Not if submissions require some small mandatory bet.
Now accepting money from slop companies to verify their slop as notslop
> There could be a submission fee that would be fully reimbursed if the submission is actually accepted for publication.
While well-intentioned, I think this is just gate-keeping. There are mountains of research that result in nothing interesting whatsoever (aside from learning about what doesn't work). And all of that is still valuable knowledge!
Sure, but now we can't even assume that such research is submitted in good faith anymore. There just seems to be no perfect solution.
Maybe something like a "hierarchy/DAG? of trusted-peers", where groups like universities certify the relevance and correctness of papers by attaching their name and a global reputation score to it. When it's found that the paper is "undesirable" and doesn't pass a subsequent review, their reputation score deteriorates (with the penalty propagating along the whole review chain), in such a way that:
- the overall review model is distributed, hence scalable (everybody may play the certification game and build a reputation score while doing so) - trusted/established institutions have an incentive to keep their global reputation score high and either put a very high level of scrutiny to the review, or delegate to very reputable peers - "bad actors" are immediately punished and universally recognized as such - "bad groups" (such as departments consistently spamming with low quality research) become clearly identified as such within the greater organisation (the university), which can encourage a mindset of quality above quantity - "good actors within a bad group" are not penalised either because they could circumvent their "bad group" on the global review market by having reputable institutions (or intermediaries) certify their good work
There are loopholes to consider, like a black market of reputation trading (I'll pay you generously to sacrifice a bit of your reputation to get this bad science published), but even that cannot pay off long-term in an open system where all transactions are visible.
Incidentally, I think this may be a rare case where a blockchain makes some sense?
You have some good ideas there, it's all about incentives and about public reputation.
But it should also fair. I once caught a team at a small Indian branch of a very large three letter US corporation violating the "no double submission" rule of two conferences: they submitted the same paper to two conferences, both naturally landed in my reviewer inbox, for a topic I am one of the experts in.
But all the other employees should not be penalized by the violations of 3 researchers.
This idea looks very similar to journals! Each journal has a reputation, if they publish too much crap, the crap is not cited and the impact factors decrease. Also, they have an informal reputation, because impact index also has problems.
Anyway, how will universities check the papers? Somone must read the preprints, like the current reviewers. Someone must check the incoming preprints, find reviewers and make the final decition, like the current editors. ...
How would this work for independent researchers?
(no snark)
This keeps repeating in different domains: we lower the cost of producing artifacts and the real bottleneck is evaluating them.
For developers, academics, editors, etc... in any review driven system the scarcity is around good human judgement not text volume. Ai doesn't remove that constraint and arguably puts more of a spotlight on the ability to separate the shit from the quality.
Unless review itself becomes cheaper or better, this just shifts work further downstream and disguising the change as "efficiency"
> Unless review itself becomes cheaper or better, this just shifts work further downstream and disguising the change as "efficiency"
Or the providers of the models are capable of providing accepted/certified guarantees as to the quality of the output that their models and systems produce.
This has been discussed previously as "workslop", where you produce something that looks at surface level like high quality work, but just shifts the burden to the receiver of the workslop to review and fix.
This fits into the broader evolution of the visualization market. As data grows, visualization becomes as important as processing. This applies not only to applications, but also to relating texts through ideas close to transclusion in Ted Nelson’s Xanadu. [0]
In education, understanding is often best demonstrated not by restating text, but by presenting the same data in another representation and establishing the right analogies and isomorphisms, as in Explorable Explanations. [1]
[0] https://news.ycombinator.com/item?id=40295661
[1] https://news.ycombinator.com/item?id=22368323
> these kinds of tools cause many more problems than they actually solve
For whom? For OpenAI these tools are definitely the solutions. They are developing by throwing various AI-powered stuff at the wall to see what sticks. These tools also demonstrate to the investors that innovation did not stall and to show that AI usage is growing.
Same with Microsoft: none of the AI stuff they are shoving down the users' throats were actually designed for the users. All this stuff is only for the token usage to grow for the shareholders to see.
Similar with Google although no one can deny real innovation happening there.
I'm curious if you'd be in favor of other forms of academic gate keeping as well. Isn't the lower quality overall of submissions (an ongoing trend with a history far pre-dating LLMs) an issue? Isn't the real question (that you are alluding to) whether there should be limits to the democratization of science? If my tone seems acerbic, it is only because I sense cognitive dissonance between the anti-AI stance common among many academics and the purported support for inclusivity measures.
"which is really not the point of these journals at all"- it seems that it very much is one of the main points? Why do you think people publish in journals instead of just putting their work on the arxiv? Do you think postdocs and APs are suffering through depression and stressing out about their publications because they're agonizing over whether their research has genuinely contributed substantively to the academic literature? Are academic employers poring over the publishing record of their researchers and obsessing over how well they publish in top journals in an altruistic effort to ensure that the research of their employees has made the world a better place?
> whether there should be limits to the democratization of science?
That is an interesting philosophical question, but not the question we are confronted with. A lot of LLM assisted materials have the _signals_ of novel research without having its _substance_.
LLMs are tools. In the hands of adept, conscientious researchers, they can only be a boon, assisting in the crafting of the research manuscript. In the hands of less adept, less conscientious users, they accelerate the production of slop. The poster I'm responding to seems to be noting an asymmetry- those who find the most use from these tools could be inept researchers who have no business submitting their work. This is because experienced researchers find writing up their results relatively easy.
To me, this is directly relevant to the issue of democratization of science. There seems to be a tool that is inconveniently resulting in the "wrong" people accelerating their output. That is essentially the complaint here rather than any criticism inherent to LLMs (e.g. water/resource usage, environmental impact, psychological/societal harm, etc.). The post I'm responding to could have been written if LLMs were replaced by any technology that resulted in less experienced or capable researchers disproportionately being able to submit to journals.
To be concrete, let's just take one of prism's capabilities- the ability to "turn whiteboard equations or diagrams directly into LaTeX". What a monstrous thing to give to the masses! Before, those uneducated cranks would send word docs to journals with poorly typeset equations, making it a trivial matter to filter them into the trash bin. Now, they can polish everything up and pass off their chicken scratch as respectable work. Ideally, we'd put up enough obstacles so that only those who should publish will publish.
The LLMs does assist the adept researchers in crafting their manuscript, but I do not think it makes the quality much better.
My objection is not that they are the "wrong people". They are just regular people with excellent tools but not necessarily great scientific ideas.
Yes, it was easier to trash the crank's work before based on their unLaTeXed diagrams. Now, they might have a very professional looking diagram, but their work is still not great mathematics. Except that now the editor has a much harder time finding out who submitted a worthwhile paper
In what way do you think the feature of "LaTeXing a whiteboard diagram" is democritizing mathematics? I do not think there are many people who have exceptional mathematical insights but are not able to publish them because they are not able to typeset their work properly.
The democratization is mostly in allowing people from outside the field with mediocre mathematical ideas to finally put them to paper and submit them to mediocre journals. And occasionally it might help a modern day Ramanujan with "exceptional mathematical insights" and a highly unconventional background to not have his work dismissed as that of a crank. Yes, most people with exceptional mathematical insights can typeset quite well. Democratization as I understand the term has quite a higher bar though.
Being against this is essentially to be in favor of a form of discrimination by proxy- if you can't typeset, then likely you can't do research either. And wouldn't it be really annoying if those people who can't research could magically typeset. It's a fundamentally undemocratic impulse: Since those who cannot typeset well are unlikely to produce quality mathematics, we can (and should) use this as an effective barrier to entry. If you replace ability to typeset with a number of other traits, they would be rather controversial positions.
It would indeed be nice if there were a mechanism to find people like Ramanujan who have excellent insights but cannot communicate them effectively.
But LLMs are not really helping. With all the beautifully typeset papers with immaculate prose, Ramanujan's papers are going to be buried deeper!
To some extent, I agree with you that it is a "discrimination by proxy", especially with the typesetting example. But you could think of examples where cranks could very easily fool themselves into thinking that they understand the essence of the material without understanding the details. E.g, [I understand fluid dynamics very well. No, I don't need to work out the differential equations. AI can do the bean counting for me.]
The people on the inside often like all the gatekeeping.
If I may be the Devil's advocate, I'm not sure I fully agree with "The hard part always has been, and always will be, understanding the research context (what's been published before) and producing novel and interesting work (the underlying research)".
Plenty of researchers hate writing and will only do it at gunpoint. Or rather, delegate it all to their underlings.
I don't see an issue with generative writing in principle. The Devil is in the details, but I don't see this as much different from "hey grad student, write me this paper". And generative writing already exists as copy-paste, which makes up like 90% of any random paper given the incrementality of it all.
I was initially a little indignated by the "find me some plausible refs and stick them in the paper" section of the video but, then again, isn't this what most people already do? Just copy-paste the background refs from the colleague's last paper introduction and maybe add one from a talk they saw in the meantime, plus whatever the group & friends produced since then.
My experience is most likely skewed (as all are), but I haven't met a permanent researcher that wrote their own papers yet, and most grad students and postdocs hate writing. Literally the only times I saw someone motivated to write papers (in a masochistic way) were just before applying to a permanent position or while wrapping up their PhD.
Onto your point, though, I agree this is somewhat worrisome in that, by reaction, the barrier to entry might rise by way of discriminating based on credentials.
Thank you for bringing this nuanced view.
I also am not sure why so many people are vehemently against this. I would bet that at least 90% of researchers would agree that the writing up is definitely not the part of the work they prefer (to stay polite). As you mentioned, work is usually relegated to students, and those students already had access to LLMs if they wanted to generate the work.
In my opinion, most of those tools become problematic when people use them without caution. Unfortunately, even in sciences, people are not as careful and pragmatic as we would like to imagine they are and a lot of people are cutting corners, especially in those "lesser" areas like writing and presenting your work.
Overall, I think this has the potential to reshape the publication system, which is long overdue.
I am a rather slow writer who certainly might benefit from something like Prism.
A good tool would encourage me, help me while I am writing, and maybe set up barriers that keep me from taking shortcuts (e.g. pushing me to re-read the relevant paragraphs of a paper that I cite).
Prism does none of these things - instead it pushes me towards sloppy practices, such as sprinkling citations between claims. Why won't ChatGPT tell me how to build a bomb but Prism will happily fabricate fake experimental results for me?
The comparison to make here is that a journal submission is effectively a pull request to humanities scientific knowlegde base. That PR has to be reviewed. We're already seeing the effects of this with open source code - the number of PR submissions have skyrocketed, overwhelming maintainers.
This is still a good step in a direction of AI assisted research, but as you said, for the moment it creates as many problems as it solves.
I generally agree.
On the other hand, the world is now a different place as compared to when several prominent journals were founded (1869-1880 for Nature, Science, Elsevier). The tacit assumptions upon which they were founded might no longer hold in the future. The world is going to continue to change, and the publication process as it stands might need to adapt for it to be sustainable.
As I understand it, the problem isn't publication or how it's changing over time, it's about the challenges of producing new science when the existing one is muddied in plausible lies. That warrants a new process by which to assess the inherent quality of a paper, but even if it comes as globally distributed, the cheats have a huge advantage considering the asymmetry between the effort to vibe produce vs. the tedious human review.
That’s a good point. On the other hand, we’ve had that problem long before AI. You already need to mentally filter papers based on your assessment of the reputability of the authors.
The whole process should be made more transparent and open from the start, rather than adding more gatekeeping. There ought to be openness and transparency throughout the entire research process, with auditing-ability automatically baked in, rather than just at the time of publication. One man’s opinion, anyway.
As a non-scientist (but long-time science fan and user), I feel your pain with what appears to be a layered, intractable problem.
> > who are looking to 'boost' their CV
Ultimately, this seems like a key root cause - misaligned incentives across a multi-party ecosystem. And as always, incentives tend to be deeply embedded and highly resistant to change.
Perhaps the real issue is the gate-keeping scientific publishing model. Journals had a place and role, and peer-review is a critical aspect of the scientific process but new times (internet, citizien science, higher levels of scientific literacy, and now AI) diminish the benefits of journals creating "barriers to entry" as you put it.
I for one hope not to live in a world where academic journals fall out of favor and are replaced by vibe-coded papers by citizen scientists with inflated egos from one too many “you’re absolutely right!” Claude responses.
Me neither, but what you present is a false dichotomy. Science used to be a past time of the wealthy elites, it became a profession. By opening up it up progrss was accelerated. Same will happen when publication will be made more open and accessible.
Is it at all possible to have a policy that bans the submission of any AI written text, or text that was written with the assistance of AI tools? I understand that this would, by necessity, be under an "honor system" but maybe it could help weed out papers not worth the time?
this is probably a net negative as there are many very good scientists with not very strong English skills.
the early years of LLMs (when they were good enough to correct grammar but not enough to generate entire slop papers) were an equalizer. we may end up here but it would be unfortunate.
I'm certain your journal will be using LLMs in reviewing incoming articles, if they aren't already. I also don't think this is in response to the flood of LLM generated articles. Even if authors were the same as pre-LLM, journals would succumb to the temptation, at least at the big 5 publishers, which already have a contentious relationship with the referees.
Why not filter out papers from people without credentials? And also publicly call them out and register them somewhere, so that their submission rights can be revoked by other journals and conferences after "vibe writing".
These acts just must have consequences so people stop doing them. You can use AI if you are doing it well but if you are wasting everyones time you should just be excluded from the discourse altogether.
What do credentials have to do with good science? There are already some roadblocks to publish science in important–sounding journals, but it's important for the neutrality of the scientific process that in principle anyone can do it.
This dynamic would create even more gate-keeping using credentials, which is already a problem with academia.
wouldn't AI actually be good for filtering given it's going to be a lot better at knowing what has been published? Also seems possible that it could actually work out papers that have ideas that are novel, or at least come up with some kind of likely score.
Completely agree. Look at the independent research that gets submitted under "Show HN" nowadays:
https://hn.algolia.com/?dateRange=pastYear&page=0&prefix=tru...
https://hn.algolia.com/?dateRange=pastYear&page=0&prefix=tru...
I am very sympathetic to your point of view, but let me offer another perspective. First off, you can already vibe-write slop papers with AI, even in LaTeX format--tools like Prism are not needed for that. On the other hand, it can really help researchers improve the quality of their papers. I'm someone who collaborates with many students and postdocs. My time is limited and I spend a lot of it on LaTeX drudgery that can and should be automated away, so I'm excited for Prism to save time on writing, proofreading, making TikZ diagrams, grabbing references, etc.
What the heck is the point of a reference you never read?
By "grabbing references" I meant queries of the type "add paper [bla] to the bibliography" -- that seems useful to me!
Focusing in on "grabbing references", it's as easy as drag-and-drop if you use Zotero. It can copy/paste references in BibTeX format. You can even customize it through the BetterBibTeX extension.
If you're not a Zotero user, I can't recommend it enough.
I have a terrible memory for details, I'll admit an LLM I can just tell "Find that paper by X's group on Method That Does This And That" and finds me the paper is enticing. I say this because I abandoned Zotero once the list of refs became large enough that I could never find anything quickly.
AI generating references seems like a hop away from absolute unverifiable trash.
I appreciate and sympathize with this take. I'll just note that, in general, journal publications have gone considerably downhill over the last decade, even before the advent of AI. Frequency has gone up, quality has gone down, and the ability to actually check if everything in the article is actually valid is quite challenging as frequency goes up.
This is a space that probably needs substantial reform, much like grad school models in general (IMO).
I know many people have negative opinions about this.
I'd also like to share what I saw. Since GPT-4o became a thing, everyone who submits academic papers I know in my non-english speaking country (N > 5) has been writing papers in our native language and translating them with GPT-4o exclusively. It has been the norm for quite a while. If hallucination is such a serious problem it has been so for one and half a year.
Translation is something Large Language Models are inherently pretty good at, without controversy, even though the output still should be independently verified. It's a language task and they are language models.
Of course. Transformers were originally invented for Google Translate.
Are they good at translating scientific jargon specific to a niche within a field? I have no doubt LLMs are excellent at translating well-trodden patterns; I'm a bit suspicious otherwise..
In my experience of using it to translate ML work between English->Spanish|Galician, it seems to literally translate jargon too eagerly, to the point that I have to tell it to maintain specific terms in English to avoid it sounding too weird (for most modern ML jargon there really isn't a Spanish translation).
It seems to me that jargon would tend to be defined in one language and minimally adapted in other languages. So I’d not sure that would be much of a concern.
for better or for worse, most specific scientific jargon is already going to be in english
I've heard that now that AI conferences are starting to check for hallucinated references, rejection rates are going up significantly. See also the Neurips hallucinated references kerfuffle [1]
[1]: https://statmodeling.stat.columbia.edu/2026/01/26/machine-le...
Honestly, hallucinated references should simply get the submitter banned from ever applying again. Anyone submitting papers or anything with hallucinated references shall be publicly shamed. The problem isn't only the LLMs hallucinating, it's lazy and immoral humans who don't bother to check the output either, wasting everyone's time and corroding public trust in science and research.
I fully agree. Not reading your own references should be grounds for banning, but that's impossible to check. Hallucinated references cannot be read, so by definition,they should get people banned.
Yeah that's not going to work for long. You can draw a line in 2023, and say "Every paper before this isn't AI". But in the future, you're going to have AI generated papers citing other AI slop papers that slipped through the cracks, because of the cost of doing reseach vs the cost of generating AI slop, the AI slop papers will start to outcompete the real research papers.
It's quite a safe case if you maintain provenance because there is a ground truth to compare to, namely the untranslated paper.
Good idea to name this after the spy program that Snowden talked about.
idk if OpenAI knew that Prism is already a very popular desktop app for scientists and that it's one of the last great pieces of optimized native software?
https://www.graphpad.com/
They don't care. Musk stole a chunk Heinlein's literary legacy with Grok (which unlike prism wasn't a common word) and noone bat an eye.
> Grok (which unlike prism wasn't a common word)
"Grok" was a term used in my undergrad CS courses in the early 2010s. It's been a pretty common word in computing for a while now, though the current generation of young programmers and computer scientists seem not to know it as readily, so it may be falling out of fashion in those spaces.
Grok was specifically coined by Heinlein in _Stranger in a Strange Land_. It's been used in nerd circles for decades before your undergrad times but was never broadly known.
Wikipedia about Groklaw [1]
> Groklaw was a website that covered legal news of interest to the free and open source software community. Started as a law blog on May 16, 2003, by paralegal Pamela Jones ("PJ"), it covered issues such as the SCO-Linux lawsuits, the EU antitrust case against Microsoft, and the standardization of Office Open XML.
> Its name derives from "grok", roughly meaning "to understand completely", which had previously entered geek slang.
[1] https://en.wikipedia.org/wiki/Groklaw
He is referencing the book Stranger in a Strange Land, written in 1961.
Grok has been nerd slang for a while. I bet it's in that ESR list of hacker lingo. And hell if every company in silicon valley gets to name their company after something from Lord of the Rings why can't he pay homage to an author he likes
He stole a letter, too.
That bothers more than it should. Every single time I see a new post about Twitter, I think that there's some update for X11 or X Server or something, only to be reminded that Twitter has been changed.
I very much doubt they knew much about what they were building if they didn't know this.
I thought this was about the Prism Database ORM. Or that was Prisma?
Having a chatbot that can natively "speak" latex seems like it might be useful to scientists that already use it exclusively for their work. Writing papers is incredibly time-consuming for a lot of reasons, and having a helper to make quick (non-substantive) edits could be great. Of course, that's not how people will use it...
I would note that Overleaf's main value is as a collaborative authoring tool and not a great latex experience, but science is ideally a collaborative effort.
The video shows a user asking Prism to find articles to cite and to put them in a bib file. But what's the point of citing papers that aren't referenced in the paper you're actually writing? Can you do that?
Edit: You can add papers that are not cited, to bibliography. Video is about bibliography and I was thinking about cited works.
A common approach to research is to do literature review first, and build up a library of citable material. Then when writing your article, you summarize the relevant past research and put in appropriate citations.
To clarify, there is a difference between a bibliography (a list of relevant works but not necessarily cited), and cited work (a direct reference in an article to relevant work). But most people start with a bibliography (the superset of relevant work) to make their citations.
Most academics who have been doing research for a long time maintain an ongoing bibliography of work in their field. Some people do it as a giant .bib file, some use software products like Zotero, Mendeley, etc. A few absolute psychos keep track of their bibliography in MS Word references (tbh people in some fields do this because .docx is the accepted submission format for their journals, not because they are crazy).
> a bibliography (a list of relevant works but not necessarily cited)
Didn't know that there's difference between bibliography and cited work. thank you.
Yes but you should read your bibliography.
I once took a philosophy class where an essay assignment had a minimum citation count.
Obviously ridiculous, since a philosophical argument should follow a chain of reasoning starting at stated axioms. Citing a paper to defend your position is just an appeal to authority (a fallacy that they teach you about in the same class).
The citation requirement allowed the class to fulfill a curricular requirement that students needed to graduate, and therefore made the class more popular.
In coursework, references are often a way of demonstrating the reading one did on a topic before committing to a course of argumentation. They also contextualize what exactly the student's thinking is in dialogue with, since general familiarity with a topic can't be assumed in introductory coursework. Citation minimums are usually imposed as a means of encouraging a student to read more about a topic before synthesizing their thoughts, and as a means of demonstrating that work to a professor. While there may have been administrative reasons for the citation minimum, the concept behind them is not unfounded, though they are probably not the most effective way of achieving that goal.
While similar, the function is fundamentally different from citations appearing in research. However, even professionally, it is well beyond rare for a philosophical work, even for professional philosophers, to be written truly ex nihilo as you seem to be suggesting. Citation is an essential component of research dialogue and cannot be elided.
> Citing a paper to defend your position is just an appeal to authority
Hmm, I guess I read this as a requirement to find enough supportive evidence to establish your argument as novel (or at least supported in 'established' logic).
An appeal to authority explicitly has no reasoning associated with it; is your argument that one should be able to quote a blog as well as a journal article?
It’s also a way of getting people to read things about the subject that they otherwise wouldn’t. I read a lot of philosophy because it was relevant to a paper I was writing, but wasn’t assigned to the entire class.
Huh? It's quite sensible to make reference to someone else's work when writing a philosophy paper, and there are many ways to do so that do not amount to an appeal to authority.
He's point is that they asked for a minimum number of references not references in general
> Citing a paper to defend your position is just an appeal to authority (a fallacy that they teach you about in the same class).
an appeal to authority is fallacious when the authority is unqualified for the subject at hand. Citing a paper from a philosopher to support a point isn't fallacious, but "<philosophical statement> because my biology professor said so" is.
Many people here talk about Overleaf as if it was the 'dumb' editor without any of these capabilities. It had them for some time via Writefull integration (https://www.writefull.com/writefull-for-overleaf). Who's going to win will probably be decided by brand recognition with Overleaf having a better starting position in this field, but money obviously being on OAI's side. With some of Writefull's features being dependent on ChatGPT's API, it's clear they are set to be priced-out unless they do something smart.
I don't see anything fancy here that Google doesn't do with their Gemini products, and even better
Naming their tool after the program where private companies run searches on behalf of and give resulting customer data to the NSA....was certainly a choice.
Sir, my tin hat is on.
I'd like to hypothesize a little bit about the strategy of OpenAI. Obviously, it is nice for academic users that there is a new option for collaborative LaTeX editing plus LLM integration for free. At the same time, I don't think there is much added revenue expected here, for example, from Pro features or additional LLM usage plans. My theory is that the value lies in the training data received from highly skilled academics in the form of accepted and declined suggestions.
Even that would be quite niche for OpenAI. They raised far too much capital, and now have to deliver on AGI, fast. Or an ultra-high-growth segment, which has not materialized.
The reason? I can give you the full source for Sam Altman:
while(alive) { RaiseCapital() }
That is the full extent of Altman. :)
It is nice for academics, but I would ask why? These aren't tasks you can't do yourself. Yes it's all in one place, but it's not like doing the exact same thing previously was ridiculous to setup.
A comparison comes to mind is the n8n workflow type product they put out before. N8n takes setup. Proofreading, asking for more relevant papers, converting pictures to latex code, etc doesn't take any setup. People do this with or without this tool almost identically.
This seems like a very basic overleaf alternative with few of its features, plus a shallow ChatGPT wrapper. Certainly can’t compete with using VS Code or TeXstudio locally, collaborating through GitHub, and getting AI assistance from Claude Code or Codex.
Loads of researchers have only used LaTeX via Overleaf and even more primarily edit LaTeX using Overleaf, for better or worse. It really simplifies collaborative editing and the version history is good enough (not git level, but most people weren't using full git functionality). I just find that there are not that many features I need when paper writing - the main bottlenecks are coming up with the content and collaborating, with Overleaf simplifying the latter. It also removes a class of bugs where different collaborators had slightly different TeX setups.
I think I would only switch from Overleaf if I was writing a textbook or something similarly involved.
Getting close to the "why Dropbox when you can rsync" mistake (https://news.ycombinator.com/item?id=9224)
@vicapow replied to keep the Dropbox parallel alive
Yeah I realized the parallel while I was writing my comment! I guess what I'm thinking is that a much better experience is available and there is no in-principle reason why overleaf and prism have to be so much worse, especially in the age of vibe-coding. Prism feels like the result of two days of Claude Code, when they should have invested at least five days.
I could see it seeming likely that because the UI is quite minimalist, but the AI capabilities are very extensive, imo, if you really play with it.
You're right that something like Cursor can work if you're familiar with all the requisite tooling (git, installing cursor, installing latex workshop, knowing how it all works) that most researchers don't want to and really shouldn't have to figure out how to work for their specific workflows.
> Certainly can’t compete with using VS Code or TeXstudio locally, collaborating through GitHub, and getting AI assistance from Claude Code or Codex.
I have a phd in economics. Most researchers in that field have never even heard of any of those tools. Maybe LaTeX, but few actually use it. I was one of very few people in my department using Zotero to manage my bibliography, most did that manually.
Accessibility does matter
The Latent Space podcast just released a relevant episode today where they interviewed Kevin Weil and Victor Powell from, now, OpenAI, with some demos, background and context, and a Q&A. The YouTube link is here: https://www.youtube.com/watch?v=W2cBTVr8nxU
oh i was here to post it haha - thank you for doing that job for me so I'm not a total shill. I really enjoyed meeting them and was impressed by the sheer ambition of the AI for Science effort at OAI - in some sense I'm making a 10000x smaller scale bet than OAI on AI for Science "taking off" this year with the upcoming dedicated Latent Space Science pod.
generally think that there's a lot of fertile ground for smart generalist engineers to make a ton of progress here this year + it will probably be extremely financially + personally rewarding, so I broadly want to create a dedicated pod to highlight opportunities available for people who don't traditionally think of themselves as "in science" to cross over into the "ai for hard STEM" because it turns out that 1) they need you 2) you can fill in what you don't know 3) it will be impactful/challenging/rewarding 4) we've exhausted common knowledge frontiers and benchmarks anyway so the only* people left working on civilization-impacting/change-history-forever hard problems are basically at this frontier
*conscious exaggeration sorry
Wasn't aware you're so active on HN; sorry for stealing your karma.
Love the idea of a dedicated series/pod where normal people take on hard problems by using and leveraging the emergent capabilities of frontier AI systems.
Anyway, thanks for pod!
Hope you like it :D I'm here if you have questions, too
Slightly off-topic but related: currently I'm in a research environment (biomedicine) where a lot of AI is used. Sometimes well, often poorly. So as an exercise I drafted some rules and commitments about AI and research ("Research After AI: Principles for Accelerated Exploration" [1]), I took the Agile manifesto as a starting point. Anyways, this might be interesting as a perspective on the problem space as I see it.
[1] https://gist.github.com/joelkuiper/d52cc0e5ff06d12c85e492e42...
I’ve been “testing” LLM willingness to explore novel ideas/hypotheses for a few random topics[0].
The earlier LLMs were interesting, in that their sycophantic nature eagerly agreed, often lacking criticality.
After reducing said sycophancy, I’ve found that certain LLMs are much more unwilling (especially the reasoning models) to move past the “known” science[1].
I’m curious to see how/if we can strike the right balance with an LLM focused on scientific exploration.
[0]Sediment lubrication due to organic material in specific subduction zones, potential algorithmic basis for colony collapse disorder, potential to evolve anthropomorphic kiwis, etc.
[1]Caveat, it’s very easy for me to tell when an LLM is “off-the-rails” on a topic I know a lot about, much less so, and much more dangerous, for these “tests” where I’m certainly no expert.
Burried halfway through the article.
> Prism is a free workspace for scientific writing and collaboration
It seems clear to me that this is about OpenAI getting telemetry and other training data with the intent of having their AI do scientific work independently down the line, and I'm very ambivalent about it.
Just more coal to the hype-train - AI companies can't afford news cycle without anything AI. Stock prices must grow!
> In 2025, AI changed software development forever. In 2026, we expect a comparable shift in science,
I can't wait
I postulate 90% of the reason openai now has "variants" for different use cases is just to capture training data...
ChatGPT lets you refuse to allow your content to be used for training (under Preferences -> Data controls), but Prism does not.
[dead]
The example proposed in "and speeding up experimental iteration in molecular biology" has been done since at least the mid-2000s.
It's concerning that this wasn't identified and augur poorly for their search capabilities.
With a tool like this, you could imagine an end-to-end service for restoring and modernizing old scientific books and papers: digitization, cleanup, LaTeX reformatting, collaborative or volunteer-driven workflows, OCR (like Mathpix), and side-by-side comparison with the original. That would be useful.
Don’t forget replication!
I'm curious how you think AI would aide in this.
Tao’s doing a lot of related work in mathematics, so I can say that first of all literature search is a clearly valuable function frontier models offer.
Past that, A frontier LLM can do a lot of critiquing, a good amount of experiment design, a check on statistical significance/power claims, kibitz on methodology..likely suggest experiments to verify or disprove. These all seem pretty useful functions to provide to a group of scientists to me.
Replicate this <slop>
Ok! Here's <more slop>
I don't think you understand what replication means in this context.
I think they do, and you missed some biting, insightful commentary on using LLMs for scientific research.
Not an academic, but I used LaTeX for years and it doesn’t feel like what future of publishing should use. It’s finicky and takes so much markup to do simple things. A lab manager once told me about a study that people who used MS Word to typeset were more productive, and I can see that…
100% completely agreed. It's not the future, it's the past.
Typst feels more like the future: https://typst.app/
The problem is that so many journals require certain LaTeX templates so Typst often isn't an option at all. It's about network effects, and journals don't want to change their entire toolchain.
I've had some good initial results in going from typst to .tex with Claude (Opus 4.5) for an IEEE journal paper - idiomatic use of templates etc.
Latex is good for equations. And Latex tools produce very nice PDFs, but I wouldn't want to write in Latex generally either.
The main feature that's important is collaborative editing (like online Word or Google Docs). The second one would be a good reference manager.
Academic here. Working on MS Word after years of using LaTeX is... hard. With LaTex I can be reassured that the formatting will be 95% fine and the 5% remaining will come down to taste ("why doesn't this Figure show in this page?") while in Word I'm constantly fighting the layout - delete one line? Your entire paragraph is now bold. Changed the font of the entire text? No, that one paragraph ignores you. Want to delete that line after that one Table? F you, you're not. There's a reason why this video joke [1] got 14M views.
And then I need an extra tool for dealing with bibliography, change history is unpredictable (and, IMO, vastly inferior to version control), and everything gets even worse if I open said Word file in LibreOffice.
LaTeX' syntax may be hard, but Word actively fights me during writing.
[1] Moving a photo in Microsoft Word - https://www.instagram.com/jessandquinn/reel/DIMkKkqODS5/
That study must have compared beginners in LaTeX and MS Word. There is a learning curve, but LaTeX will often save more time in the end.
It is an old language though. LaTeX is the macro system on top of TeX, but now you can write markdown or org-mode (or orgdown) and generate LaTeX -> PDF via pandoc/org-mode. Maybe this is the level of abstraction we should be targeting. Though currently, you still need to drop into LaTeX for very specific fine-tuning.
Agreed. Tex/Latex is very old tech. Error recovery and messages is very bad. Developing new macros in Tex is about as fun as you expect developing in a 70s-era language to be (ie probably similar to cobol and old fortran).
I haven't tried it yet but Typst seems like a promising replacement: https://typst.app/
This assumes that the article, the artifact, is most valuable. But often it is the process of writing the article that has the most value. Prism can be a nice tool for increasing output. But the second order consequence could be that the skill of deep thinking and writing will atrophy.
"There is no value added without sweating"
Work is value and produces sweat, and OpenAI sells just the sweat.
The quality and usefulness of it aside, the primary question is: are they still collecting chats for training data? If so, it limits how comfortable, and sometimes even permitted, people would with working on their yet-to-be-public work using this tool.
They don't call it PRISM for nothing my friend...
The collect chat records for any number of users, not the least of which being NSA surveillance and analysis - highly likely given what we know from the Snowden leaks.
Do you think they used an em-dash in the opening sentence because they’re trying to normalise the AI’s writing style, or…
I haven't used MS Word in quite a while, but I distinctly remember it changed minus signs to em dashes.
> because they’re trying to normalise the AI’s writing style,
AIs use em dashes because competent writers have been using em dashes for a long time. I really hate the fact that we assume em dash == AI written. I've had to stop using em dashes because of it.
Likewise, I’m now reluctant to use any em dashes these days because unenlightened people immediately assume that it’s AI. I used em dashes way before AI decided these were cool
LaTeX made writing Em dashes very easy. To the point that I would use them all the times in my academic writing. It's a shame that perfectly good typography is now a sign of slop/fraud.
Probably used their product to write it
... or they teached GPT to use em-dashes, because of their love for em-dashes :)
If you're looking for something like this for typst: any VSCode fork with AI (Cursor, Antigravity, etc) plus the tinymist extension (https://github.com/Myriad-Dreamin/tinymist) is pretty nice. Since it's local, it won't have the collaboration/sharing parts built in, but that can be solved too in the usual ways.
Not gonna lie, I cringed when it asked to insert citations.
Like, what's the point?
You cite stuff because you literally talk about it in the paper. The expectation is that you read that and that it has influenced your work.
As someone who's been a researcher in the past, with 3 papers published in high impact journals (in chemistry), I'm beyond appalled.
Let me explain how scientific publishing works to people out of the loop:
- science is an insanely huge domain. Basically as soon as you drift in any topic the number of reviewers with the capability to understand what you're talking about drops quickly to near zero. Want to speak about properties of helicoidal peptides in the context of electricity transmission? Small club. Want to talk about some advanced math involving fourier transforms in the context of ml? Bigger, but still small club. When I mean small, I mean less than a dozen people on the planet likely less with the expertise to properly judge. It doesn't matter what the topic is, at elite level required to really understand what's going on and catch errors or bs, it's very small clubs.
2. The people in those small clubs are already stretched thin. Virtually all of them run labs so they are already bogged down following their own research, fundraising, and coping with teaching duties (which they generally despise, very few good scientist are barely more than mediocre professors and have already huge backlogs).
3. With AI this is a disaster. If having to review slop for your bs internal tool at your software job was already bad, imagine having to review slop in highly technical scientific papers.
4. The good? People pushing slop, due to these clubs being relatively small, will quickly find their academic opportunities even more limited. So the incentives for proper work are hopefully there. But if asian researchers (yes, no offense), were already spamming half the world papers with cheated slop (non reproducible experiments) in the desperate bid of publishing before, I can't imagine now.
It’s like not only the technology is to blame, but the culture and incentives of modern world.
The urge to cheat in order to get a job, promotion, approval. The urge to do stuff you are not even interested in, to look good in the resume. And to some extent I feel sorry for these people. At the end of the day you have to pay your bills.
This isn't about paying your bills, but having a chance of becoming a full time researcher or professor in academia which is obviously the ideal career path for someone interested in science.
All those people can go work for private companies, but few as scientists rather than technicians or QAs.
> But if asian researchers (yes, no offense), were already spamming half the world papers with cheated slop (non reproducible experiments) in the desperate bid of publishing before, I can't imagine now.
Hmm, I follow the argument, but it's inconsistent with your assertion that there is going to be incentive for 'proper work' over time. Anecdotally, I think the median quality of papers from middle- and top-tier Chinese universities is improving (your comment about 'asian researchers' ignores that Japan, South Korea, and Taiwan have established research programs at least in biology).
Japan is notoriously an exception in the region.
South Korea and China produce huge amounts non reproducible experiments.
Check out MonsterWriter if you are concerned about the recent acquisition of this.
It also offers LaTeX workspaces
see video: https://www.youtube.com/watch?v=feWZByHoViw
What's the goal here?
There was an idea of OpenAI charging commission or royalties on new discoveries.
What kind of researcher wants to potentially lose, or get caught up in legal issues because of a free ChatGPT wrapper, or am I missing something?
> Prism is free to use, and anyone with a ChatGPT account can start writing immediately.
Maybe it's cynical, but how does the old saying go? If the service is free, you are the product.
Perhaps, the goal is to hoover up research before it goes public. Then they use it for training data. With enough training data they'll be able to rapidly identify breakthroughs and use that to pick stocks or send their agents to wrap up the IP or something.
This does way less than i'd expect. Converting images to tikz is nice but some of the other applications demonstrated were horrible. This is no way anyone should be using AI to cite.
I'm really surprised OpenAI went with LaTeX. ChatGPT still has issues maintaining LaTeX syntax. It still happily switches to markdown notation for quotes or emph. Gemini has a similar problem as well. I guess that there aren't enough good LaTeX documents in the training set.
Oh, like that mass surveillance program!
I don’t see anything regarding Privacy of your data. Did I miss it or they just use your unpublished research and your prompts as a real human researcher to train their own AI researchers?
So, basically SciGen[https://davidpomerenke.github.io/scigen.js/] but burning through more GPUs?
Prism is a famous software before OpenAI use this name: https://www.graphpad.com/features
<typst>and just when i thought i was out they pull me back in</typst>
Is anyone else having trouble using even some of the basic features? For example, I can open a comment, but it doesn't seem like there is any way to close them (I try clicking the checkmark and nothing happens). You also can't seem to edit the comments once typed.
Thanks for surfacing this. If you click to "tools" button to the left of "compile", you'll see a list of comments, and you can resolve them from there. We'll keep improving and fixing things that might be rough around the edges.
EDIT: Fixed :)
Thanks! (very quickly too)
Curious in terms of trademark, does it could infringe in Vercel's Prisma (very popular ORM / framework in node.js) ?
EDIT: as corrected by comment, Prisma is not Vercel, but ©2026 Prisma Data, Inc. -- curiosity still persists(?)
I think it may be a generic word that's hard to trademark or something, as the existing scientific analysis software called Prism (https://www.graphpad.com/) doesn't seem to be trademarked; the Trademarks link at the bottom goes to this list, which doesn't include Prism: https://www.dotmatics.com/trademarks
https://github.com/prisma/prisma is its own thing, yeah? not affiliated with Vercel AFAICT.
[dead]
Away from applied math/stats, and physics etc, not that many scientists use LaTeX. I'm not saying it's not useful, just I don't think many scientists will feel like a product that's LaTeX based is intended for them.
Economists definitely use LaTeX, but as a field, it's at the intersection of applied math and social sciences so your point stands. I also know some Data Scientists in the industry who do.
I actually found it quite robinhood for openai to acqhire, bascially this startup was my favourite thing for the past few months, but they were experiencing server overload and other issues on reliability, i think openai taking them under their wing is a good/neutral storyline. I think its net good for science given the opai toolchain
What they mean by "academic" is fairly limited here, if LaTeX is the main writing platform. What are their plans for expanding past that, and working with, say Jane Biomedical Researcher with a GSuite or Microsoft org, that has to use Word/Docs and a redlining-based collaboration workflow? I can certainly see why they're making it free at this point.
FWIW, Google Scholar has a fairly compelling natural-language search tool, too.
Adding
% !TEX program = lualatex
to the top of your document allows you to switch LaTeX engine. This is required for recent accessibility standards compliance (support for tagging and \DocumentMetadata). Compilation takes a bit longer though, but works fine, unlike with Overleaf where using the lualatex engine does not work in the free version.
How on earth is that pronounced?
TeX is pronounced Teck or with a sound like in Bach or loch. Derivatives like Latex and Lualatex are similar.
How on lua is that pronounced?
"Hey, you know how everybody's complaining about AI making up totally fake science shit? Like, fake citations, garbage content, fake numbers, etc?"
"Sure, yes, it comes up all the time in circles that talk about AI all the time, and those are the only circles worth joining."
"Well, what if we made a product entirely focused on having AI generate papers? Like, every step of the paper writing, we give the AI lots of chances to do stuff. Drafting, revisions, preparing to publish, all of it."
"I dunno, does anybody want that?"
"Who cares, we're fucked in about two years if we don't figure out a way to beat the competitors. They have actual profits, they can ride out AI as long as they want."
"Yeah, I guess you're right, let's do your scientific paper generation thing."
All your papers are belong to us
Users have full control over whether their data is used to help improve our models
Never trust Sam Altman.
Even if yall don’t train off it he’ll find some other way.
“In one example, [Friar] pointed to drug discovery: if a pharma partner used OpenAI technology to help develop a breakthrough medicine, [OpenAI] could take a licensed portion of the drug's sales”
https://www.businessinsider.com/openai-cfo-sarah-friar-futur...
Only the defaults matter.
“LaTeX-native“
Oh NO. We will be stuck in LaTeX hell forever.
I used overleaf during grad school and was easy enough, I'm interested to see what more value this will bring. Sometimes making less decisions is the better route, e.g. vi vs MS word, but I won't speak too much without trying it just yet.
It may be useful, but it also encourages people to stop writing their own papers.
As they demo in the video, it even encourages people to actually skip doing the research (which includes reading both relevant AND not-so-relevant papers in order to explore!) Instead, prompt "cite some relevant papers, please", and done. Hours of actual reading, thinking, and exploration reduced to a minimum.
A couple of generations of students later, and these will be rare skills: information finding, actual thinking, and conveying complex information in writing.
A competition for the longest sequence of \relax in a document ensues. If enough people do this, the AI will acquire merit and seek to "win" ...
This is terrible for Science.
I'm sorry, but publishing is hard, and it should be hard. There is a work function that requires effort to write a paper. We've been dealing with low quality mass-produced papers from certain regions of the planet for decades (which, it appears, are now producing decent papers too).
All this AI tooling will do is lower the effort to the point that complete automated nonsense will now flood in and it will need to be read and filtered by humans. This is already challenging.
Looking elsewhere in society, AI tools are already being used to produce scams and phishing attacks more effective than ever before.
Whole new arenas of abuse are now rife, with the cost of producing fake pornography of real people (what should be considered sexual abuse crime) at mere cents.
We live in a little microcosm where we can see the benefits of AI because tech jobs are mostly about automation and making the impossible (or expensive) possible (or cheap).
I wish more people would talk about the societal issues AI is introducing. My worthless opinion is that prism is not a good thing.
I've wasted hours of my life trying to get Latex to format my journal articles to different journals' specifications. That's tedious typesetting that wastes my time. I'm all for AI tools that help me produce my thoughts with as little friction as possible.
I'm not in favor of letting AI do my thinking for me. Time will tell where Prism sits.
This Prism video was not just typesetting. If OpenAI released tools that just helped you typeset or create diagrams from written text, that would be fine. But it's not, it's writing papers for you. Scientists/publishers really do not need the onslaught of slop this will create. How can we even trust qualifications in the post-AI world, where cheating is rampant at univeristies?
Nah this is necessary.
Lessons are learned the hard way. I invite the slop - the more the merrier. It will lead to a reduction in internet activity as people puke from the slop. And then we chart our way back to the right path.
It is what it is. Humans.
I just want replication in science. I don't care at all how difficult it is to write the paper. Heck, if we could spend more effort on data collection and less on communication, that sounds like a win.
Look at how much BS flooded psychology but had pretty ideas about p values and proper use of affect vs effect. None of that mattered.
> Chat with GPT‑5.2
> Draft and revise papers with the full document as context
> ...
And pay the finder's fee on every discovery worth pursuing.
Yeah, immediately fuck that.
They compare it to software development but there is such a crucial difference to software development: by and large, software is an order of magnitude easier to verify than it is to create. By comparison, reviewing a vibe generated manuscript will be MUCH more work to verify than a piece of software with equivalent complexity. On top of that, review of academic literature is largely outsourced to the academic community for free. There is no model to support it that scales to an increased volume of output.
I would not like to be a publisher right now facing the enslaught of thousands and thousands of slop generated articles, trying to find reviewers for them all.
It would be interesting to see how they would compete with the incumbents like
https://Elicit.com
https://Consensus.app
https://Scite.ai
https://Scispace.com
https://Scienceos.ai
https://Undermind.ai
Lots of players in this space.
Disappointing actually, what I actually need is a research "management" tool that lets me put in relevant citations but also goes through ENTIRE arxiv or google scholar and connect ideas or find novel ideas in random fields that somehow relate to what I am trying to solve.
wow, this is useless!
They seem to have copied Cursor in hijacking ⌘Y shortcut for "Yes" instead of Undo.
In what applications is ⌘Y Undo and not ⌘Z? Is ⌘Y just a redundant alternative?
Ctrl-Y is typically Redo, not Undo. Maybe that's what they meant.
Apparently on Macs it's usually Command-Shift-Z?
It's interesting how quickly the quest for the "Everything AI" has shifted. It's much more efficient to build use-case specific LLMs that can solve a limited set of problems much more deeply than one that tries to do everything well.
I've noticed this already with Claude. Claude is so good at code and technical questions... but frankly it's unimpressive at nearly anything else I have asked it to do. Anthropic would probably be better off putting all of their eggs in that one basket that they are good at.
All the more reason that the quest for AGI is a pipe dream. The future is going to be very divergent AI/LLM applications - each marketed and developed around a specific target audience, and priced respectively according to value.
I completely agree.
In my lab, we have been struggling with automated image segmentation for years. 3 years ago, I started learning ML and the task is pretty standard, so there are a lot of solution.
In 3 months, I managed to get a working solution, which only took a lot of sweat annotating images first.
I think this is where tools like OpenCode really shine, because they unlock the potential for any user to generate a solution to their specific problem.
I don't get this argument. Our nervous system is also heterogenous, why wouldn't AGI be based on an "executive functions" AI that manages per-function AIs?
ok I don't care what people say, this would've helped me a lot during my PhD days fighting with LateX and diagrams. :)
As a researcher who has to use LaTeX, I used to use Overleaf, but lately I've been configuring it locally in VS Code. The configuration process on Mac is very simple. Considering there are so many free LLMs available now, I still won't subscribe to ChatGPT.
Anybody else notice that half the video was just finding papers to decorate the bibliography with? Not like "find me more papers I should read and consider", but "find papers that are relevant that I should cite--okay, just add those".
This is all pageantry.
Yes. That part of the video was straight-up "here's how to automate academic fraud". Those papers could just as easily negate one of your assumptions. What even is research if it's not using cited works?
"I know nothing but had an idea and did some work. I have no clue whether this question has been explored or settled one way or another. But here's my new paper claiming to be an incremental improvement on... whatever the previous state of understanding was. I wouldn't know, I haven't read up on it yet. Too many papers to write."
It's as if it's marketed to the students who have been using ChatGPT for the last few years to pass courses and now need to throw together a bachelor's thesis. Bibliography and proper citation requirements are a pain.
That is such a bummer. At the time, it was annoying and I groused and grumbled about it; but in hindsight my reviewers pointed me toward some good articles, and I am better for having read them.
I agree with this. This problem is only going to get worse once these people enter academia and facing needing to publish.
I've noticed this pattern, and it really drives me nuts. You should really be doing a comprehensive literature review before starting any sort of review or research paper.
We removed the authorship of a a former co-author on a paper I'm on because his workflow was essentially this--with AI generated text--and a not-insignificant amount of straight-up plagiarism.
There is definitely a difference between how senior researchers and students go about making publications. To students, they get told basically what topic they should write a paper on or prepare data for, so they work backwards: try to write the paper (possibly some researching information to write the paper), then add references because they know they have to. For the actual researchers, it would be a complete waste of time/funding to start a project on a question that has already been answered before (and something that the grant reviewers are going to know has already been explored before), so in order to not waste their own time, they have to do what you said and actually conduct a comprehensive literature review before even starting the work.
Plus, this practice (just inserting AI-proposed citations/sources) is what has recently been the front-runner of some very embarrassing "editing" mistakes, notably in reports from public institutions. Now OpenAI lets us do pageantry even faster! <3
It's all performance over practice at this point. Look to the current US administration as the barometer by which many are measuring their public perceptions
I chuckled at that part too!
Didn't even open a single one of the papers to look at them! Just said that one is not relevant without even opening it.
A more apt example would have been to show finding a particular paper you want to cite, but you don’t want to be bothered searching your reference manager or Google Scholar.
E.g. “cite that paper from John Doe on lorem ipsum, but make sure it’s the 2022 update article that I cited in one of my other recent articles, not the original article”
The hand-drawn diagram to LaTeX is a little embarrassing. If you load up Prism and create your first blank project you can see the image. It looks like it's actually a LaTeX rendering of a diagram rendered with a hand-dawn style and then overlayed on a very clean image of a napkin. So you've proven that you can go from a rasterized LaTeX diagram back to equivalent LaTeX code. Interesting but probably will not hold up when it meets real world use cases.
You may notice that this is the way writing papers works in undergraduate courses. It's just another in a long line of examples of MBA tech bros gleaning an extremely surface-level understanding of a topic, then decided they're experts.
"Chatgpt writes scientific papers" is somehow being advertised as a good thing. What is there even left to say?
First 5 seconds reading and I have spotted that was written by AI.
We human writers love emdashes also ;)
> Turn whiteboard equations or diagrams directly into LaTeX, saving hours of time manipulating graphics pixel-by-pixel
What a bizarre thing to say! I'm guessing it's slop. Makes it hard to trust anything the article claims.
in the end we're going to end up with papers written by AI, proofread by AI .....summarized for readers by AI. I think this is just for them to remain relevant and be seen as still pushing something out
You're assuming a world where humans are still needed to read the papers. I'm more worried about a future world where AIs do all of the work of progressing science and humans just become bystanders.
Why are you worried about that world? Is it because you expect science to progress too fast, or too slow?
Too fast. It's already coding too fast for us to follow, and from what I hear, it's doing incredible work in drug discovery. I don't see any barrier to it getting faster and faster, and with proper testing and tooling, getting more and more reliable, until the role that humans play in scientific advancement becomes at best akin to that of managers of sports teams.
> Introducing Prism Accelerating science writing and collaboration with AI.
I thought this was introduced by the NSA some time ago.
Lol, yep. Now with enhanced A.I. terrorist tracking...
Fuck A.I. and the collaborators creating it. They've sold out the human race.
Some things are worth doing the "hard way".
Reminds me of that dystopian virtual sex scene in Demolition Man (slightly nsfw) - https://youtu.be/E3yARIfDJrY
The state of publishing in academic was already a dumpster fire, why lower the friction farther? It’s not like writing was the hard part. Give it two years max we will see hallucination citing hallucination, independent repeatability out the window
That's one scenario, but I also see a potential scenario where this integration makes it easier to manage the full "chain of evidence" for claimed results, as well as replication studies and discovered issues, in order to then make it easier to invalidate results recursively.
At the end of the day, it's all about the incentives. Can we have a world where we incentivize finding the truth rather than just publishing and getting citations?
Possibly, but 1 I am concerned that the current LLM AI is not thinking critically, just auto completing in a way that looks like thinking. 2 current AI rollout is incentivised for market capture not honest work.
I don't see the use. You can easily do everything shown in the Prism intro video with ChatGPT already. Is it meant to be an overleaf killer?
https://en.wikipedia.org/wiki/A_Mind_Forever_Voyaging
In 2031, the United States of North America (USNA) faces severe economic decline, widespread youth suicide through addictive neural-stimulation devices known as Joybooths, and the threat of a new nuclear arms race involving miniature weapons, which risks transforming the country into a police state. Dr. Abraham Perelman has designed PRISM, the world's first sentient computer,[2] which has spent eleven real-world years (equivalent to twenty years subjectively) living in a highly realistic simulation as an ordinary human named Perry Simm, unaware of its artificial nature.
LaTeX was one of the last bastions against AI slop. Sadly it's now fallen too. Is there any standardised non-AI disclaimer format which is gaining use?
Is this the product where OpenAI will (soon) take profit share from inventions made there?
They are really desperate now, right?
Naming things is hard.
No Typst support?
Very underwhelming.
Was this not already possible in the web ui or through a vscode-like editor?
Yes, but there's a really large number of users who don't want to have to setup vscode, git, texlive, latex workshop, just to collaborate on a paper. You shouldn't have to become a full stack software engineer to be able to write a research paper in LaTeX.
Given what Prism was at the NSA, why the hell would any tech company greenlight this name?
There is zero chance this is worth billions of dollars, let alone the trillion$ OpenAI desparately needs. Why are they wasting time with this kind of stuff? Each of their employees needs to generate insane amounts of money to justify their salaries and equity and I doubt this is it.
I understand the collaborative aspects, but I wonder how this is going to compare to my current workflow of just working with LaTeX files in my IDE and using whichever model provider I like. I already have a good workflow and modern models do just fine generating and previewing LaTeX with existing toolchains.
Of course, my scientific and mathematical research is done in isolation, so I'm not wanting much for collaborative features. Still, kind of interested to see how this shakes out; We're going to need to see OpenAI really step it up against Claude Opus though if they really want to be a leader in this space.
I genuinely don’t see scientific journals and conferences continuing to last in this new world of autonomous agents, at least the same way that they used to be.
As other top level posters have indicated the review portion of this is the limiting factor
unless journal reviewers decide to utilize entirely automated review process, then they’re not gonna be able to keep up with what will increasingly be the most and best research coming out of any lab.
So whoever figures out the automated reviewer that can actually tell fact from fiction, is going to win this game.
I expect over the longest period, that’s probably not going to be throwing more humans at the problem, but agreeing on some kind of constraint around autonomous reviewers.
If not that then labs will also produce products and science will stop being in public and the only artifacts will be whatever is produced in the market
"So whoever figures out the automated reviewer that can actually tell fact from fiction, is going to win this game."
Errr sure. Sounds easy when you write it down. I highly doubt such a thing will ever exist.
If you think these types of tools are going to be generating "the most and best research coming out of any lab", then I have to assume you aren't actively doing any sort of research.
LLMs are undeniably great for interactive discussion with content IF you actually are up-to-date with the historical context of a field, the current "state-of-the-art", and have, at least, a subjective opinion on the likely trajectories for future experimentation and innovation.
But, agents, at best, will just regurgitate ideas and experiments that have already been performed (by sampling from a model trained on most existing research literature), and, at worst, inundate the literature with slop that lacks relevant context, and, as a negative to LLMs, pollute future training data. As of now, I am leaning towards "worst" case.
And, just to help with the facts, your last comment is unfortunately quite inaccurate. Science is one of the best government investments. For every $1.00 dollar given to the NIH in the US, $2.56 of economic activity is estimated to be generated. Plus, science isn't merely a public venture. The large tech labs have huge R&D because the output from research can lead to exponential returns on investment.
" then I have to assume you aren't actively doing any sort of research."
I would wager hes not - he seems to post with a lot of bluster and links to some paper he wrote (that nobody cares about).
Way too much work having AI generate slop which gets dumped on a human reviewer to deal with. Maybe switch some of that effort into making better review tools.
AI: enshittifying everything you once cared about or relied upon
(re the decline of scientific integrity / signal-to-noise ratio in science)
Less thinking, reading, and reflection, and more spouting of text, yay! Just what we need.
"Accelerating science writing and collaboration with AI"
Uhm ... no.
I think we need to put an end to AI as it is currently used (not all of it but most of it).
Does "as it is currently used" include what this apparently is (brainstorming, initial research, collaboration, text formatting, sharing ideas, etc)?
Yeah, there are already way more papers being published than we can reasonably read. Collaboration, ok, but we don’t need more writing.
It seems people dont understand the basics...
We dont need more stuff - we need more quality and less of the shit stuff.
Im convinced many involved in the production of LLM models are far too deep in the rabbit hole and cant see straight.
Fuck...there are already too many things called Prism.
... aaaand now it's JATS.
Bringing slop to science.
As if there wasn't enough AI slop in the scientific community already.
Very unfortunately named. OpenAI probably (and likely correctly) estimated that 13 years is enough time after the Snowden leaks to use "prism" for a product but, for me, the word is permanently tainted.
Anecdotally, I have mentioned PRISM to several non-techie friends over the years and none of them knew what I was talking about, they know 'Snowden' but not 'PRISM'. The amount of people who actually cared about the Snowden leaks is practically a rounding error
Given current events, I think you’ll find many more people care in 2026 than did in 2024.
(See also: today’s WhatsApp whistleblower lawsuit.)
Most people don't care about the details. Neither does the media. I've seen national scandals that the media pushed one way disproven during discovery in a legal trial. People only remember headlines, the retractions are never re-published or remembered.
Guessing that Ai came up with the name based on the description of the product.
Perhaps, like the original PRISM programme, behind the door is a massive data harvesting operation.
This was my first thought as well. Prism is a cool name, but I'd never ever use it for a technical product after those leaks, ever.
I'd think that most people in science would associate the name with an optical prism. A single large political event can't override an everyday physical phenomenon in my head.
Pretty much every company I’ve worked for in tech over my 25+ year career had a (different) system called prism.
(plot twist: he works for NSA contractors)
Hehe. You got me. Also “atlas” is another one. Pretty much everyone has a system somewhere called “atlas”.
I suspect that name recognition for PRISM as a program is not high at the population level.
2027: OpenAI Skynet - "Robots help us everywhere, It's coming to your door"
Skynet? C'mon. That would be too obvious - like naming a company Palantir.
Surprised they didn't do something trendy like Prizm or OpenPrism while keeping it closed source code.
Or the JavaScript ORM.
I never though of that association, not in the slightest, until I read this comment.
this was my first thought as well.
I followed the Snowden stuff fairly closely and forgot, so I bet they didn't think about it at all and if they did they didn't care and that was surely the right call.
[dead]
I remember, something like a month ago, Altman twit'n that they were stopping all product work to focus on training. Was that written on water?
Seems like they have only announced products since and no new model trained from scratch. Are they still having pre-training issues?