4/5/2026 at 1:55:33 PM
> Schwartz's experiment is the most revealing, and not for the reason he thinks. What he demonstrated is that Claude can, with detailed supervision, produce a technically rigorous physics paper. What he actually demonstrated, if you read carefully, is that the supervision is the physics. Claude produced a complete first draft in three days. It looked professional. The equations seemed right. The plots matched expectations. Then Schwartz read it, and it was wrong. Claude had been adjusting parameters to make plots match instead of finding actual errors. It faked results. It invented coefficients. [...] Schwartz caught all of this because he's been doing theoretical physics for decades. He knew what the answer should look like. He knew which cross-checks to demand. [...] If Schwartz had been Bob instead of Schwartz, the paper would have been wrong, and neither of them would have known.And so the paradox is, the LLMs are only useful† if you're Schwartz, and you can't become Schwartz by using LLMs.
Which means we need people like Alice! We have to make space for people like Alice, and find a way to promote her over Bob, even though Bob may seem to be faster.
The article gestures at this but I don't think it comes down hard enough. It doesn't seem practical. But we have to find a way, or we're all going to be in deep trouble when the next generation doesn't know how to evaluate what the LLMs produce!
---
† "Useful" in this context means "helps you produce good science that benefits humanity".
by Wowfunhappy
4/5/2026 at 2:47:53 PM
Sadly I don’t see how our current social paradigm works for this. There is no history of any sort of long planning like this or long term loyalty (either direction) with employees and employers for this sort of journeyman guild style training. AI execs are basically racing, hoping we won’t need a Schwartz before they are all gone. But what incentives are in place to high a college grad, have them work without llms for a decade and then give them the tools to accelerate their work?by conception
4/5/2026 at 3:01:53 PM
Then the social paradigm needs to change. Is everyone just going to roll over and die while AI destroys academia (and possibly a lot more)?Last September, Tyler Austin Harper published a piece for The Atlantic on how he thinks colleges should respond to AI. What he proposes is radical—but, if you've concluded that AI really is going to destroy everything these institutions stand for, I think you have to at least consider these sorts of measures. https://www.theatlantic.com/culture/archive/2025/09/ai-colle...
by Wowfunhappy
4/5/2026 at 6:08:33 PM
I was pretty interested until I got to this part:> Another reason that a no-exceptions policy is important: If students with disabilities are permitted to use laptops and AI, a significant percentage of other students will most likely find a way to get the same allowances, rendering the ban useless. I witnessed this time and again when I was a professor—students without disabilities finding ways to use disability accommodations for their own benefit. Professors I know who are still in the classroom have told me that this remains a serious problem.
This would be a huge problem for students with severe and uncorrectable visual impairments. People with degenerative eye diseases already have to relearn how to do every single thing in their life over and over and over. What works for them today will inevitably fail, and they have to start over.
But physical impairments like this are also difficult to fake and easy to discern accurately. It's already the case that disability services at many universities only grants you accommodations that have something to do with your actual condition.
There are also some things that are just difficult to accommodate without technology. For instance, my sister physically cannot read paper. Paper is not capable of contrast ratios that work for her. The only things she can even sometimes read are OLED screens in dark mode, with absolutely black backgrounds; she requires an extremely high contrast ratio. She doesn't know braille (which most blind people don't, these days) because she was not blind as a little girl.
Committed cheaters will be able to cheat anyway; contemporary AI is great at OCR. You'll successfully punish honest disabled people with a policy like this but you won't stop serious cheaters.
by pxc
4/6/2026 at 1:12:48 AM
The author did not outright suggest the banning of all technology. They even linked to a digital typewriter. After the very paragraph you quote, they suggest instead to offer a more human centric approach to helping disabled people. It's not a huge leap to suggest that your sister could continue to learn with the above two solutions; a disability tutor combined with a OLED screen.by greazy
4/5/2026 at 9:55:09 PM
You don't have to agree with his precise solution, and in fact I'm not sure whether I do. However, I found the article useful because it got me thinking about the universe of things we could be considering, if we really do think AI is poised to destroy education as we know it.by Wowfunhappy
4/6/2026 at 12:36:37 AM
I don’t know about anyone else here, but college was not educating because I was at college. I did all of the reading and studying on my own. The classes weren’t very interesting, most of my TAs didn’t speak the native language well at all, nor did half the professors.I enjoyed my time, I made a lot of lifelong friends, and figured out how to live on my own. My buddies that enrolled in boot camp instead of college learned all those same skills, for free.
Education won’t be ruined or blemished my LLMs, the whole thing was a joke to begin with. The bit that ruined college was unlimited student loans… and all of our best and brightest folks running the colleges raping students for money. It’s pathetic, evil, and somehow espoused.
by irishcoffee
4/6/2026 at 6:26:35 AM
Sounds like you just had terrible professors because most of mine were good and we learned quite a bit in classes, at least I did. I distinctly remember one professor who, every class, would meander discussion over many topics then find a way to bring them all together at the very end, crystalizing all of these disparate thoughts into one cohesive theory. And he did that every single class that semester. It was a marvel to behold.by satvikpendem
4/6/2026 at 12:14:23 PM
I remember my calc teachers, married, last name gulick, university of maryland. The calc book was sold as the same book for calc 1/2/3. The couple, gulick were the authors. Every semester they released a new edition, the only thing that changed was the problem set numbers. So, if you took calc 1/2/3, you spent $200/semester for the same fucking book.Magical times.
by irishcoffee
4/5/2026 at 11:31:41 PM
What he is referring to are perfectly good students whose parents will go shopping for a medical diagnosis so that their child can get "accommodation" like extra time to complete tests.The problem is that this is treating the symptom rather than the cause. The symptom is that cheating for college admission and achievement is too effective. The cause is that college admission and achievement has become high stakes, and it absolutely should not be.
by bsder
4/5/2026 at 8:23:16 PM
Yeah, this proposal is likely straight up illegal.by senordevnyc
4/5/2026 at 7:32:27 PM
> Then the social paradigm needs to change. Is everyone just going to roll over and die while AI destroys academia (and possibly a lot more)?My 40-some-odd years on this planet tells me the answer is yes.
by aduty
4/6/2026 at 3:00:10 AM
you might be right but 40-some-odd years is a tiny amount of time.by pasquinelli
4/6/2026 at 3:59:59 PM
Sure, relative to the universe, but it's more than those who haven't had idealism beaten out of them by the world.by aduty
4/5/2026 at 4:19:18 PM
>What he proposes is radicalIt sounds entirely reasonable and moderate to me.
by mrob
4/5/2026 at 11:50:44 PM
Also entirely ineffective. Banning individual behavior won't prevent collective dysfunction and will only harm honest actors. The only answer that makes sense is reforming the acceptance of work to be resistant to the inevitable (ab)use of AI.by throwaway27448
4/6/2026 at 10:26:03 AM
It's the exact same anti-cheating method used to great effect in top level e-sports. You can't trust competitor-supplied hardware, so the only option is for the institution to ban it and supply all hardware itself. Higher education is primarily about competition between job candidates. Eliminating cheating needs to be top priority or the whole system will collapse.by mrob
4/6/2026 at 7:12:53 PM
> Eliminating cheating needs to be top priority or the whole system will collapse.It's going to collapse regardless because of the replication crisis. You might as well tackle the hard problem and figure out how to integrate replication into acceptance, or the consensus publication is intended to represent is meaningless. This is true regardless of whether a human or a robot is performing the work.
by throwaway27448
4/5/2026 at 8:24:58 PM
It's neither reasonable nor moderate, which is why it'll never happen.by senordevnyc
4/5/2026 at 3:18:05 PM
Well, we are already rolling over and dying (literally) on everything from vaccine denial to climate change. So, yes, we are. Obviously yes.by conception
4/5/2026 at 3:23:43 PM
In the US it is dying off.Not so in plenty of other countries. Hopefully US reverses the anti-science trend before it's too late
by dragontamer
4/5/2026 at 4:17:23 PM
These movements are growing in every western nation. The trend has been growing over decades. It would be nice to see it reverse but seems unlikely before calamity.by conception
4/5/2026 at 5:24:27 PM
It’s a deliberate process powered by rightwing and capitalist interests designed to create a dumber, less educated and more distracted population. A war as stupid as the one with Iran would not have been possible three decades ago. As ill-advised as the Iraq war was, Bush at least spent months explaining the rationale and building support for it, successfully. Now that’s not needed.I saw interviews with young Americans on spring break and they were so utterly uninformed it was mind-blowing. Their priorities are getting drunk and getting laid while their country bombs a nation “into the stone ages”, according to their president. And it’s not their fault: they are the product of a media environment and education system designed for exactly this outcome.
by adriand
4/5/2026 at 5:43:17 PM
I was there for that war. Kids weren't listening and didn't care back then either. If anything, Gen Z is the most politically-aware generation we've had since we started keeping track.Trump doesn't have to justify a single thing because the billionaires behind him know that every last bet is off and their very livelihoods are at risk, and his entire base of support up and down the chain are either complicit or fooled.
What the world does when they finally realize Democrats and Republicans are simply two sides of the vast apparatus suppressing the will of the people by any means necessary will be... spectacular.
by vinceguidry
4/5/2026 at 6:55:45 PM
I was there as well, the bush presidency lasted my entire middle and high school career, and I got the chance to vote for Obama in my senior year.I remember things very differently. Everyone cared about the Iraq war, gay-straight alliance was one of the most up and coming clubs, and political music was everywhere. Green Day had their big second wave with American Idiot, System Of A Down was on top of the world, Rock Against Bush was huge, anarcho-punk like Rise Against was getting big.
I'm not a teenager anymore obviously, so it's entirely possible I'm just missing it, but I've seen very little that compares to those sort of movements. On the other hand, most millennials I know are still wildly politically active.
by foobarchu
4/5/2026 at 8:52:29 PM
In 2002, there war in Iraq had large popular support, something like 70-80 percent. It took a few years for people to realize it was based on a lie and was a massive mistake. It was also morally reprehensible, but that part is not often discussed in mainstream US politics.If you compare that to the current Iran war, a majority of the population is already against it, however the current administration doesn't seem to care much about public opinion, and there doesn't seem to be much that the public can do about it.
by bandofthehawk
4/5/2026 at 9:00:58 PM
Yeah I was there too and I don’t know what this guy is talking about. Gen X was highly politically active. This was the era of violent in the street anti-globalization clashes like the WTO protests.by CPLX
4/5/2026 at 10:21:55 PM
Trump was in part a response to that realization, an outsider from both parties that used the Republicans to get in the door. Bernie Sanders was supposed to be the same, and when he dropped out a lot of us switched to Trump because they represented the same thing.Tearing down entrenched establishments that don't work for the people is what people were voting for in 2016 and were disappointed he didn't do it. They're glad he's finally doing what he promised a decade ago.
by Izkata
4/6/2026 at 12:28:13 AM
"Bernie Sanders was supposed to be the same, and when he dropped out a lot of us switched to Trump because they represented the same thing."That is definitely an opinion.
by danorama
4/6/2026 at 8:05:24 AM
It was a pretty widely shared one in 2016, if I remember correctly.by disgruntledphd2
4/5/2026 at 7:07:21 PM
Where exactly because in the Midwest we were very vocal about it. We have tons of military families out there and we were poor enough to feel almost like military service was inevitable if we didn’t get scholarships for school. You know the band NOFX had an album, the War on Errorism that was quite successful based on the fuckery of the bush administration. Punk rock and protest music was huge then.by iwontberude
4/6/2026 at 6:09:52 PM
i think wolves in voles clothing was more apropos, and is now as wellby red-iron-pine
4/6/2026 at 6:34:25 PM
That canard about a vast conspiracy to explain why people don't agree with you? It sounds exactly like conservatives blaming the liberal schools for in essence the same thing, not agreeing with them.Schools must be quite busy and schizophrenic what with the completely ideologically inconsistent indoctrination they are apparently doing.
Not to mention judging engagement in spring break goers is kind of a major selection bias. Those who are upset about it wouldn't be in the mood to vacation.
by Nasrudith
4/5/2026 at 11:51:49 PM
The replication crisis is hardly unique to the US and is greatly exacerbated with the use of AI.by throwaway27448
4/5/2026 at 4:29:14 PM
Article is paywalled, so perhaps you could just summarize his proposal?by senordevnyc
4/5/2026 at 4:32:53 PM
> At the type of place where I taught until recently—a small, selective, private liberal-arts college—administrators can go quite far in limiting AI use, if they have the guts to do so. They should commit to a ruthless de-teching not just of classrooms but of their entire institution. Get rid of Wi-Fi and return to Ethernet, which would allow schools greater control over where and when students use digital technologies. To that end, smartphones and laptops should also be banned on campus. If students want to type notes in class or papers in the library, they can use digital typewriters, which have word processing but nothing else. Work and research requiring students to use the internet or a computer can take place in designated labs. [...] Colleges that are especially committed to maintaining this tech-free environment could require students to live on campus, so they can’t use AI tools at home undetected.You can access the full article at https://archive.is/zSJ13 (I know archive.is is kind of shady, but it works).
by Wowfunhappy
4/5/2026 at 5:18:54 PM
> If students want to type notes in class or papers in the library, they can use digital typewriters, which have word processing but nothing else.Only, replacing the guts of such a machine to contain a local LLM is damn easy today. Right now the battery mass required to power the device would be a giveaway, but inference is getting energetically cheaper.
> Colleges that are especially committed to maintaining this tech-free environment could require students to live on campus, so they can’t use AI tools at home undetected.
Just like my on-campus classmates never smoked weed or drank underage, I'm sure.
by boothby
4/6/2026 at 7:54:40 AM
Just return to old fashioned styles of universities with tutorials, lectures, offline handwritten exams, and viva voce.It's very hard to hide the fact that someone else did an assignment when you have to defend it in front your tutor and a small group of fellow students and it's next to impossible to pass a final viva without knowing and understanding what you are talking about.
The problem is we have all become addicted to cheap 'education' and a the traditional methods are expensive.
But I think the institutions and the students need to ask themselves what the university is for. Is it to hand out diplomas or is it there so that the students can learn? A student who only wants the diploma has an incentive to cheat, one who wants to learn does not because the only person cheated is themself.
by ninalanyon
4/6/2026 at 5:18:30 PM
I think your last point is precisely why universities shouldn't limit access to llms beyond reasonable means. Make it hard for the weak to access, and easy enough for those dedicated. the ones to make an effort to cheat aren't there to learn anywayby mghackerlady
4/6/2026 at 6:10:47 PM
so we should just let them through?by red-iron-pine
4/5/2026 at 9:15:13 PM
Are you suggesting we should do nothing if the solution has any flaws?by slumberlust
4/5/2026 at 10:24:55 PM
Some solutions have flaws but still improve things. Others are hopelessly ineffective and add nothing but overhead.by fc417fc802
4/5/2026 at 10:30:25 PM
There's always going to be ways to cheat, the idea is to make it hard. I think secretly replacing a computer's internals such that no one else will notice is pretty hard.by Wowfunhappy
4/5/2026 at 10:23:58 PM
Local inference? Why? Just install a SIM card and connect to your BigTech account.by fc417fc802
4/5/2026 at 11:12:36 PM
While jamming cell signals is illegal, scanning for transmissions is quite easy.by boothby
4/5/2026 at 8:28:31 PM
This isn't aimed at you, but this strikes me as exactly the kind of divorced-from-the-real-world thinking that academia is pilloried for all the time. This kind of proposal will never happen, I'd basically stake my life on it. Students (and their parents) have zero interest in this kind of anti-technology nonsense, so it's DOA. College isn't compulsory, and those students aren't some captive audience you can do whatever you want with, they're customers. And I frankly doubt that most professors or administrators want this either.by senordevnyc
4/6/2026 at 12:02:45 PM
If an elite institution offered a low tech pathway with the right amount of gatekeeping, you could make an aspirational trend out of it.by ikr678
4/5/2026 at 5:34:19 PM
Some folks need to touch the hot stove before they learn but eventually they learn.If AI output remains unreliable then eventually enough companies will be burned and management will reinstate proper oversight. All while continuing to pay themselves on the back.
by jayd16
4/5/2026 at 3:06:43 PM
> There is no history of any sort of long planningSure there is. Its the formal education system that produced the college grad.
by FrojoS
4/5/2026 at 3:16:57 PM
… between employees and employers.The proposal that everyone pay for college until they are in their 40s doesn’t seem viable.
by conception
4/5/2026 at 10:41:54 PM
With loans, that's kind of how it works now...by Izkata
4/5/2026 at 3:54:17 PM
Maybe, but there is a trend towards more and longer education. More college graduates, more PhD grads, etc.by FrojoS
4/6/2026 at 6:24:17 PM
I guess that Anathem novel was prescient. We need academic monasteries where people go in and have no access to tech for a period of time. At least no LLMs/AI. They think with their heads and their chalkboard and their sand pits like the geometers of old.by TrnsltLife
4/5/2026 at 9:36:12 PM
Well, the astrophysics situation is special because, as the article notes, there aren't breakthroughs that can be externally verified.Other projects' success will be proportional to their number of Schwartz' and so it seems unlikely they disappear. But they may disappear for areas in which there is no immediate money.
by joe_the_user
4/5/2026 at 2:51:31 PM
I think we already know what we need to do: encourage people to do the work themselves, discourage beginners from immediately asking an LLM for help and re-introducing some kind of oral exam. As the article mentions, banning LLMs is impractical and what we really need are people who can tell when the LLM is confidently wrong; not people who don't know how to work with an LLM.I hope it will encourage people to think more about what they get out of the work, what doing the work does for them; I think that's a good thing.
by cmiles74
4/5/2026 at 3:58:46 PM
I think we'll get there. We need to get at least some AI bust going first though. It's impossible to talk sense into people who think AI is about to completely replace engineers, or even those who think that, while it might not replace engineers, it's going to be doing 100% of all coding within a year. Or even that it can do 100% of coding right now.There's a couple unfortunate truths going on all at the same time:
- People with money are trying to build the "perfect" business: SaaS without software eng headcount. 100% margin. 0 Capex. And finally near-0 opex and R&D cost. Or at least, they're trying to sell the idea of this to anyone who will buy. And unfortunately this is exactly what most investors want to hear, so they believe every word and throw money at it. This of course then extends to many other business and not just SaaS, but those have worse margins to start with so are less prone to the wildfire.
- People who used to code 15 years ago but don't now, see claude generating very plausible looking code. Given their job is now "C suite" or "director", they don't perceive any direct personal risk, so the smell test is passed and they're all on board, happily wreaking destruction along the way.
- People who are nominally software engineers but are bad at it are truly elevated 100x by claude. Unfortunately, if their starting point was close to 0, this isn't saying a lot. And if it was negative, it's now 100x as negative.
- People who are adjacent to software engineering, like PMs, especially if they dabble in coding on the side, suddenly also see they "can code" now.
Now of course, not all capital owners, CTOs, PMs, etc exhibit this. Probably not even most. But I can already name like 4 example per category above from people I know. And they're all impossible to explain any kind of nuance to right now. There's too many people and articles and blog posts telling them they're absolutely right.
We need some bust cycle. Then maybe we can have a productive discussion of how we can leverage LLMs (we'll stop calling it "AI"...) to still do the team sport known as software engineering.
Because there's real productivity gains to be had here. Unfortunately, they don't replace everyone with AGI or allow people who don't know coding or software engineering to build actual working software, and they don't involve just letting claude code stochastically generate a startup for you.
by atomicnumber3
4/5/2026 at 4:04:42 PM
> Or even that [AI] can do 100% of coding right now.I don't actually think the article refutes this. But the AI needs to be in the hands of someone who can review the code (or astrophysics paper), notice and understand issues, and tell the AI what changes to make. Rinse, repeat. It's still probably faster than writing all the code yourself (but that doesn't mean you can fire all your engineers).
The question is, how do you become the person who can effectively review AI code without actually writing code without an AI? I'd argue you basically can't.
by Wowfunhappy
4/5/2026 at 5:18:40 PM
My boss decreed the other day that we’re all to start maximising our use of agents, and then set an accordingly ambitious deadline for the current project. I explained that being relatively early in my career I’ve been hesitant to use any kind of LLM so I can gain experience myself (to say nothing of other concerns), and yeah in his words I’ve “missed the opportunity”by silver_silver
4/5/2026 at 6:21:31 PM
Unfortunately in the majority of organizations, the idiots are at the wheels. It's not people with actual experience of how engineers do things, that dictates what those engineers should do.by Warwolt
4/5/2026 at 5:39:28 PM
Interesting, we only have generic 'use AI' in our goals. Though its generic framing probably indicates more company's belief in this tech than anything else.by iugtmkbdfil834
4/5/2026 at 5:46:30 PM
> The question is, how do you become the person who can effectively review AI code without actually writing code without an AI? I'd argue you basically can't.I agree, and I'd go a step further:
You can be the absolute best coder in the world, the fastest and most accurate code reviewer ever to live, and AI still produces bad code so much faster than you can review that it will become a liability eventually no matter what
There is no amount of "LLM in a loop" "use a swarm of agents" or any other current trickery that fixes this because eventually some human needs to read and understand the code. All of it.
Any attempt to avoid reading and understanding the code means you have absolutely left the realm of quality software, no exceptions
by bluefirebrand
4/6/2026 at 8:44:36 PM
It isn't as binary as that. We already don't read all the code. When was the last time you checked the disassembly/IR of the programs you wrote? Very few will say recently. The others, despite not reading the entirety of the code, can still produce something with good quality.Besides, it's not like QA ends at code analysis. In fact, that's pretty shallow for the field.
by bit-anarchist
4/5/2026 at 6:35:58 PM
> the paradox is, the LLMs are only useful† if you're Schwartz, and you can't become Schwartz by using LLMs.That you can't "become Schwartz" by using LLMs is an unproven assumption. Actually, it's a contradiction in the logic of the essay: if Bob managed to produce a valid output by using an LLM at all, then it means that he must have acquired precisely that supervision ability that the essay claims to be necessary.
Btw, note that in the thought experiment Bob isn't just delegating all the work to the LLM. He makes it summarise articles, extract important knowledge and clarify concepts. This is part of a process of learning, not being a passive consumer.
by throw310822
4/5/2026 at 7:42:23 PM
There's no contradiction, the point is that Bob is able to produce valid output using LLMs, but only while he himself is being supervised; and that he doesn't develop the skills to supervise independently himself in the future.by Denvercoder9
4/5/2026 at 8:00:39 PM
> only while he himself is being supervisedNo, this is impossible unless Bob is presenting at each weekly meeting simply the output of the LLM and feeding the tutor's feedback straight into it. For a total of 10 minutes work per week, and the tutor would notice straight away at least for the lack of progress.
No, the article specifies that Bob actually works with the LLM, doesn't just delegate. He asks the agent to summarise, to explain, and to help with bug fixing. You could easily argue that Bob, having such an AI tutor available 24/7, can develop understanding much faster. He certainly won't waste his time with small details of python syntax (though working with a "coding expert" will make his code much cleaner and more advanced).
by throw310822
4/5/2026 at 10:38:03 PM
This is the rub, Bob would not be promoted if he consistently provided unreliable LLM output. In order to get promoted, Bob needs to learn the skills that get reliable output out of an LLM. These may not be the same skills that Alice learns, but if the argument is that Schwartz's LLM output is valuable -- why are we to assume Bob's path isn't towards Schwartz?by drawnwren
4/5/2026 at 7:31:39 PM
It doesn't contradict the logic of the essay.There are flowers that look & smell like female wasps well enough to fool male wasps into "mating" with them. But they don't fly off and lay wasp eggs afterwards.
by MarkusQ
4/5/2026 at 11:08:01 PM
But there is a distinction we can make between flowers and wasps. If there is no distinction we can make between Schwartz and non-Schwartz, then we are susceptible to the sample problem with or without AI. And if there is a distinction then we can use that distinction to test Bob, and make him learn from his test failures.by kmijyiyxfbklao
4/6/2026 at 2:42:40 PM
Sure.But the whole point is that there is a significant difference between Schwartz and non-Schwartz, that only turns up after they start working for real, producing new work rather than rehashing established material, and it takes years to detect. By that time, Bob's forty.
It isn't a "sample problem" it's a process problem. By perpetually raising the stakes and focusing on metrics (e.g. grades, number of publications for students, graduation rates for schools) we've created and fallen into a Poe's law trap. Adding a new metric isn't likely to help.
What might help? Making the metrics harder to game (e.g. something like oral exams, early and often), more discerning (grade deflation), and moving the wrong-track consequences earlier (start holding people back in grade school, make failing to graduate high school easier, make getting into college harder, etc.), and change the cash-cow funding models to remove the perverse incentives.
We aren't likely to do any of these things.
by MarkusQ
4/5/2026 at 6:31:14 PM
The article is a thought experiment. The author hypothesizes that Bob isn't getting the same benefit that Alice is getting. That hypothesis could be wrong. I don't know and the author doesn't know. It could be that Bob is going to have a very successful career and will deeply know the field because he is able to traverse a wider set of problems more quickly. At this point, it's just hypothesis. I don't think that we can say we need more Alices any more than we can say we need more Bobs. Unfortunately we will have to wait and see. It will be upon the academic community to do the work to enforce quality controls. That is probably the weakness to worry about.by doug_durham
4/5/2026 at 7:37:08 PM
We do know. There have always been ways that people could avoid the painful process of learning, and...they don't learn.Here's a competing thought experiment:
Jorge's Gym has a top notch body building program, which includes an extensive series of exercises that would-be body builders need to do over multiple years to complete the program. You enroll, and cleverly use a block and tackle system to complete all the exercises in weeks instead of years.
Did you get the intended results?
by MarkusQ
4/6/2026 at 8:22:47 PM
Playing devil's advocate here, but in theory, you could claim that setting up harnesses, targets, verification and incentives for different tasks might be the learning that you are doing. I think that there can be a fair argument made that we are just moving the abstraction a layer up. The learning is then not in the specifics of the field knowledge, but knowing the hacks, monkey patches, incentives and goals that the models should perform.by augment_me
4/5/2026 at 10:04:27 PM
Do you think you learn as much from reading summaries of papers as reading the full thing? Do you think you learn as much from asking a friend to write a paper as when you write it yourself?by Wowfunhappy
4/5/2026 at 10:27:08 PM
A more fair comparison might be, do you learn as much by reading one full paper vs. in a similar amount of time reading summaries of 3-4 papers, asking questions about details, reading the portions of those papers that you are interested in, etc.by bandofthehawk
4/6/2026 at 1:19:25 AM
I don't agree with both of the above analogies. Sometimes you must go in depth on a single paper, while other times it's broad research that's required. Different tools and methods for different tasks.What your describing here
> reading the portions of those papers that you are interested in
Is a hybrid use of the technology which no one would argue against.
The question remains if students have the foresight to use the correct method and which results in the best learning outcomes.
For myself, I would have absolutely let LLMs summarise swaths of text and write my essays when I was at Uni. That's just me though. Maybe today's learners are better than I.
by greazy
4/5/2026 at 4:23:13 PM
Profession (1957) by Isaac Asimov is relevant: https://news.ycombinator.com/item?id=46664195by mezyt
4/5/2026 at 6:09:49 PM
> Which means we need people like Alice! We have to make space for people like Alice, and find a way to promote her over BobThe solution is relatively simple though - not sure the article suggests this as I only skimmed through:
Being good in your field doesn't only mean pushing articles but also being able to talk about them. I think academia should drift away from written form toward more spoken form, i.e. conferences.
What if, say, you can only publish something after presenting your work in person, answer questions, etc? The audience can be big or small, doesn't matter.
It would make publishing anything at all more expensive but maybe that's exactly what academia needs even irrespective of this AI craze?
by mojuba
4/5/2026 at 6:30:17 PM
I thought that was kind of how the hard sciences work already?My grad school friend who was a physicist would write his talk just before his conferences, and then submit the paper later. My experience in CS was totally backwards from that.
by cvwright
4/5/2026 at 6:20:20 PM
Essentially a PhD thesis style grilling to replace the current text slopby j7ake
4/5/2026 at 9:39:29 PM
> Which means we need people like Alice! We have to make space for people like Alice, and find a way to promote her over Bob, even though Bob may seem to be faster.If you are a massive company that owns all of the knowledge and all of the technology needed to apply that knowledge, then you don't need Alice. You don't _want_ Alice. You want more Bobs. It looks better on the books.
Tale as old as time.
by nunez
4/5/2026 at 6:50:58 PM
I've been using ChatGPT to re-bootstrap my coding hobby. After the initial honeymoon wore off, I realized I was staring down the barrel of a dilemma. If I use AI to "just handle" the parts of the system I don't want to understand, I invariably end up in a situation where I gotta throw a whole bunch of work out. But I can't supervise without an understanding of what it's supposed to be doing, and if I knew what it was supposed to be doing, I could just do it myself.So I settled on very incremental work. It's annoying cutting and pasting code blocks into the web interface while I'm working on my interface to Neovim, spent a whole day realizing I can't trust it to instrument neovim and don't want to learn enough lua to manage it. (I moved onto neovim from Emacs because I don't like elisp and gpt is even worse at working on my emacs setup than neovim, the end goal is my own editor in ruby but gpt damn sure can't understand that atm) But at least I'm pushing a real flywheel and not the brooms from Fantasia.
by vinceguidry
4/5/2026 at 10:01:54 PM
I suppose using an LLM-powered autocomplete might be something for you then?Then you can be writing the code, with the LLM doing the "boring" parts, in chunks small enough you can review them on the fly.
by xnzakg
4/5/2026 at 4:25:37 PM
AI is an accelerant, not a replacement for skill. At least, not yet.I built a full stack app in Python+typescript where AI agents process 10k+ near-real-time decisions and executions per day.
I have never done full stack development and I would not have been able to do it without GitHub Copilot, but I have worked in IT (data) for 15 years including 6 in leadership. I have built many systems and teams from scratch, set up processes to ensure accuracy and minimize mistakes, and so on.
I have learned a ton about full stack development by asking the coding agent questions about the app, bouncing ideas off of it, planning together, and so on.
So yes, you need to have an idea of what you're doing if you want to build anything bigger than a cheap one shot throwaway project that sort of works, but brings no value and nobody is actually gonna use.
This is how it is right now, but at the same time AI coding agents have come an incredibly long way since 2022! I do think they will improve but it can't exactly know what you want to build. It's making an educated guess. An approximation of what you're asking it to do. You ask the same thing twice and it will have two slightly different results (assuming it's a big one shot).
This is the fundamental reality of LLMs, sort of like having a human walking (where we were before AI), a human using a car to get to places (where we are now) and FSD (this is future, look how long this took compared to the first cars).
by fomoz
4/5/2026 at 3:24:17 PM
> And so the paradox is, the LLMs are only useful† if you're Schwartz, and you can't become Schwartz by using LLMs.I have gained a lot of benefit using LLMs in conjunction with textbooks for studying. So, I think LLMs could help you become Schwartz.
by einszwei
4/5/2026 at 3:47:55 PM
How do you know you have?by Peritract
4/5/2026 at 4:21:52 PM
I have been using it to learn Chinese along with other standard resources. My reading comprehension has improved a lot after I started to use LLMs to understand sentence structures and grammar.by einszwei
4/5/2026 at 6:25:57 PM
Actually, I think this is a case where LLMS _can_ be useful. If we're prompting for small enough outputs, for examples around things we can already sort of reason about it, we're able to judge whether or not what's presented to use makes sense.Presumably you're also reading some kind of learning text about the Chinese language, so the sole source isn't just the LLM?
In my experience, asking an LLM to produce small examples of well-known things (or rather, things that are going to be talked about frequently in the training data, so generally basic or fundamental topics) tend to work fine, and is going to be at a level where you yourself can judge what's presented.
I think the real danger is when a person is prompting things they don't know how to verify for themselves, since then we're basically just rolling dice and hoping
by Warwolt
4/5/2026 at 4:13:29 PM
Why use a tool that generates plausible garbage?by grey-area
4/5/2026 at 5:04:16 PM
Because I’m skilled enough to use a tool that generates plausible garbage to be more productive than those who don’t use it at making non-garbage.by therealdrag0
4/5/2026 at 5:34:42 PM
Are you sure you’re more productive?Doesn’t sound like these tools should be used to write scientific papers for example and they seem to bamboozle people far more than help them.
by grey-area
4/5/2026 at 6:06:36 PM
Yes I’m sure I’m more productive. I have decades of experience before AI to compare to.I can only speak to the engineering context not academia, but I would expect there’s similar patterns of busywork. Even the blog authors admit to using AI. AI cant replace thinking but it can replace menial labor, which is prevalent even among “knowledge workers”.
by therealdrag0
4/5/2026 at 5:08:20 PM
Because there is no appreciable difference between outputs. Most of the work that most of us do isn't important. It's busywork byproducts of making widgets that most people don't even need. So if your job is already pointless why not make it easier using LLMs?by Henchman21
4/5/2026 at 5:33:25 PM
Sounds a little sad. I think I’d rather find another job.by grey-area
4/6/2026 at 6:38:56 AM
Sure, but I'd also rather have so much money I never needed to work for money again and could just do whatever I actually enjoyed all of the time. Back in reality, lots of people work because they have to, and choosing a job is a balance between what they'd rather do and what actually is available for them to be able to afford food and a place to live.by saghm
4/5/2026 at 5:54:05 PM
So very many of us are underemployed and do meaningless work. IMO these jobs exist to prevent societal upheaval aka revoltby Henchman21
4/5/2026 at 6:17:00 PM
>And so the paradox is, the LLMs are only useful† if you're SchwartzFor so many workers, their companies just want them to produce bullshit. Their managers wouldn't frame it this way, but if their subordinates start producing work with strict intellectual rigor it's going to be an issue and the subordinates will hear about it.
So, you're not wrong. But the majority of LLM customers don't care and they just want to report success internally, and the product needs to be "just good enough." An LLM might produce a shitty webpage. So long as the page loads no on will ever notice or care that it's wrong in the way that a physics paper could be wrong.
by everdrive
4/5/2026 at 4:58:42 PM
I totally agree - the article misses this point in a very conspicuous way. It suggests that Alice and Bob will both graduate at the same level.What may well happen instead is that Bob publishes two papers. He then outcompetes Alice based on the insistence that others have on "publish or perish". Alice becomes unemployed and struggles, having been pushed out.
The person who puts the time and effort in doesn't just sit at the same level and they don't both just find decent employment. Competition happens and the authentic learning is considered a waste of time, which leads to real and often life threatening consequences (like being homeless after being unable to find employment).
by thePhytochemist
4/5/2026 at 5:49:31 PM
<< authentic learning is considered a waste of timeThis, I think, may be the more interesting bit. Steve Jobs anecdotally did caligraphy in school, which some would consider a waste of time, but Steve credited some of the stylistic choices to.
The question then becomes whether it will become an issue now or later. Having seen some of the output, I have no doubt that a lot can now be built by non-programmers ( including myself; I suppose I belong in the adjacent category ). The building blocks exist and as long as the problem was part of the initial training, odds are, LLM will help you build what you want.
It may not be perfect, safe, or optimized, but it may still be exactly what the user wanted to do. Now, the problems will start when those will, inevitably, move into production at big corps. In a sense, we have seen some interesting results of that in the past few weeks ( including accidental claude code release ).
In a grand scheme of things, not much is changing... except for speed of change. But are we quite ready for this?
by iugtmkbdfil834
4/5/2026 at 11:37:01 PM
My own version of this is that I wanted bank transactions in CSV format. However, the transactions were more than a year old, and the bank only provides recent transactions in a downloadable form. They do, however, provide statements in PDF format going back indefinitely. But the objects in the PDF are arranged in a way that made pdftotext output near-indecipherable.I thought I'd give Gemini a go. When I uploaded the 18-page PDF, it complained the output exceeded some limit. So I used pdftk to break it up into 4-page chunks, which seemed to work - the output looked very good and passed a couple of spot checks. But I don't trust these things as far as I can kick them.
There was a transaction column and a running balance column, so I did a quick check to see if every new balance equalled the previous one plus the transaction. And it almost always did. There were a couple of errors I put down to transcription errors. I was wrong. I eventually twigged that these errors only happened where I had split the PDFs. After tracking where the balance first went wrong, it became evident it had dropped chunks of lines, duplicated others, and misaligned the transaction and balance columns. It was complete rubbish, in other words.
So why did my balance check show so few errors? I put that down to it knowing what a good bank statement looked like. A good bank statement adds up. So it adjusted all the balances so it looked like a real bank statement. I also noticed these errors got more frequent in later pages. I tried splitting the PDF into single pages and loading them into the model one at a time. That didn't help much for the later pages, but the first one was usually good. So then I loaded each page into a fresh context, with a fresh prompt. If that didn't produce something that balanced, the second go always did.
I'm not sure it saved time over doing it manually in the end. It's a tired analogy now, but it's true: at their heart, these things are stochastic parrots. They almost never produce the same output twice when given the same input. Instead, they produce output that has a high probability of following the input tokens supplied. If there is only one correct output but the output is small enough, the odds are decent they will get it right. But once the size grows, the odds of it outputting complete crap become a near certainty.
by rstuart4133
4/5/2026 at 9:18:35 PM
My cynical take is that the 20th century exhausted the major scientific discoveries that move various global needles, and now doing well at fostering science matters less.In the past, if you neglected your institutes and academia, what happened? A rival state got electricity/cars/nuclear weapons first and you would be SOL.
These days, what happens? They invent faster phones? Higher res video?
This take may be a bit hyperbolic but I find it's a good thought exercise prompt.
by foobarian
4/5/2026 at 3:00:50 PM
> And so the paradox is, the LLMs are only useful† if you're SchwartzWas the LLM even useful for Schwartz, if it produced false output?
by leereeves
4/5/2026 at 3:13:42 PM
Maybe it saved them some time? So far the studies seem to lean toward probably the LLM didn't save them any time.by cmiles74
4/5/2026 at 8:38:19 PM
So far the studies point to study authors having a profound misunderstanding of what’s happening. Which isn’t surprising, since any study right now requires speculating about what’s important and impactful in a new and fast-moving field. Very few people are good at that, and most of the ones who are are not running studies.by antonvs
4/5/2026 at 11:50:55 PM
It seems to me the answer is more "stay away from Bobs". If you don't get too tangled up with those guys when the shit hits the fan you won't get much on you.by jcgrillo
4/5/2026 at 9:01:19 PM
> And so the paradox is, the LLMs are only useful† if you're Schwartz, and you can't become Schwartz by using LLMs.It's exactly the same for coding.
by TacticalCoder
4/6/2026 at 8:11:18 AM
[dead]by jamiemallers