hckrnws
> There can be no objective story since the very act of assembling facts requires implicit beliefs about what should be emphasized and what should be left out. History is therefore a constant act of reinterpretation and triangulation, which is something that LLMs, as linguistic averaging machines, simply cannot do.
This exactly why tech companies want to replace those jobs with LLMs.
The companies control the models, the models control the narrative, the narrative controls the world.
Whoever can get the most stories into the heads of the masses runs the world.
"tech companies", "companies [who] control the models", "whoever"
To be more discrete, patchwork alliances of elites stretching decades and centuries back to concentrate power. Tech companies are under the thumb of the US government and the US government is under the thumb of the elites. It's not direct but it doesn't need to be. Many soft power mechanisms exist and can be deployed when needed e.g. Visa/Mastercard censorship. The US was always founded for elites, by elites but concessions needed to be made to workers out of necessity. With technology and the destruction of unions, this is no longer the case. The veracity of this statement is still up for debate but truth won't stop them from giving it a shot (see WW2).
"Whoever can get the most stories into the heads of the masses runs the world."
I'd argue this is already the case. It has nothing to do with transformer models or AGI but basic machine learning algorithms being applied at scale in apps like TikTok, YouTube, and Facebook to addict users, fragment them, and destroy their sense of reality. They are running the world and what is happening now is their plan to keep running it, eternally, and in the most extreme fashion.
>History is therefore a constant act of reinterpretation and triangulation, which is something that LLMs, as linguistic averaging machines, simply cannot do.
I know you weren't necessarily endorsing the passage you quoted, but I want to jump off and react to just this part for a moment. I find it completely baffling that people say things in the form of "computers can do [simple operation], but [adjusting for contextual variance] is something they simply cannot do."
There was a version of this in the debate over "robot umps" in baseball that exposed the limitation of this argument in an obvious way. People would insist that automated calls of balls and strikes loses the human element, because human umpires could situationally squeeze or expand the strike zone in big moments. E.g. if it's the World Series, the bases are loaded, the count is 0-2, and the next pitch is close, call it a ball, because it extends the game, you linger in the drama a bit more.
This was supposedly an example of something a computer could not do, and frequently when this point was made it induced lots of solemn head nodding in affirmation of this deep and cherished baseball wisdom. But... why TF not? You actually could define high leverage and close game situations, and define exactly how to expand the zone, and machines could call those too, and do so more accurately than humans. So they could better respect contextual sensitivity that critics insist is so important.
Even now, in fits and starts, LLMs are engaging in a kind of multi-layered triangulating, just to even understand language. It can pick up on multilayered things like subtext or balance of emphasis, or unstated implications, or connotations, all filtered through rules of grammar. It doesn't mean they are perfect, but calibrating for context or emphasis that is most important for historical understanding seems absolutely within machine capabilities, and I don't know what other than punch drunk romanticism for "the human element" moves people to think that's an enlightened intellectual position.
> "[...] dynamically changing the zone is something they simply cannot do." But... why TF not?
Because the computer is fundamentally knowable. Somebody defined what a "close game" ahead of time. Somebody defined what a "reasonable stretch" is ahead of time.
The minute it's solidified in an algorithm, the second there's an objective rule for it, it's no longer dynamic.
The beauty of the "human element" is that the person has to make that decision in a stressful situation. They will not have to contextualize it within all of their other decisions, they don't have to formulate an algebra. They just have to make a decision they believe people can live with. And then they will have to live with the consequences.
It creates conflict. You can't have a conflict with the machine. It's just there, following rules. It would be like having a conflict with the beurocrats at the DMV, there's no point. They didn't make a decision, they just execute on the rules as written.
Could we maybe say that an LLM which can update its own model weights using its own internal self-narrating log may be 'closer' to being dynamic? We can use Wolfram's computational irreducibility principle which says that even simple rule-based functions will often cause unpredictable patterns of its return values. You could say that computer randomness is deterministic, but could we maybe say that ontologically a Quantum-NN LLMs could perhaps be classified on paper as being Dynamic? (unless you believe that quantum computing is deterministic).
You also can't argue over whether the machine made the right call over a pint of beer. Or yell at the Robo-Ref from the stands. It not only makes the game more sterile and less following the "rule of cool" but it also diminishes the entertainment value for the people the game is actually for.
Sports is literally, in the truest sense of the word, reality TV and people watch reality TV for the drama and because it's messy. It's good tea, especially in golf.
That’s their self selecting goal, sure. Fortunately for humanity the main drivers are old as hell, physics is ageist. Data centers are not a fundamental property of reality. They can be taken offline; sabotage or just loss of skills in time to maintain them leading to cascading failures. A new pandemic could wipe out billions and the loss of service workers cause it all to fail. Wifi satellites can go unreplaced.
They're a long long ways from "protomolecule" that just carries on infinitely on its own
CEOs don't really understand physics. Signal loss and such. Just data models that only mean something to their immediate business motives. They're more like priests; well versed in their profession, but oblivious to how anything outside that bubble works.
I think you dramatically overestimate the effectiveness of trying to shape narratives and change people's minds.
Yes, online content is incredibly influential, but it's not like you can just choose which content is effective.
The effectiveness is tied to a zeitgeist that is not predictable, as far as I have seen.
I think you dramatically underestimate how much your own mind is a product of narratives, some of which are self-selected, many of which are influenced by others.
Perhaps, but what you said is unfalsifiable and/or unknowable.
Ideologies are not invented unless you're a caveman. We all got to know the world by listening to others.
The subject of discussion is if and when external forces can alter those ideologies at will. And I have not seen any evidence to support the feasibility of that at scale.
Let's concede you can't shape narrative or change peoples minds through online content (though I would disagree on this). The very act of addicting people to digital platforms is enough for control. Drain their dopamine daily, fragment them into isolated groups, use influencers as proxies for control, and voila, you have an effect.
I would agree with you. It's easier to just muddy the waters and degrade people's ability to hold attention or think critically. But that is not the same thing as convincing them of what you want them to think.
It's always easier to throw petrol on an existing fire than to light one.
If you censor all opposing arguments, all you need to do to convince the vast majority of people of most things is too keep repeating yourself until people forget that there ever were opposing arguments.
In this world you can get people censored for slandering beef, or for supporting the outcome of a Supreme Court case. Then pay people to sing your message over and over again in as many different voices as can be recruited. Done.
edit: I left out "offer your most effective enemies no-show jobs, and if they turn them down, accuse them of being pedophiles."
i'm having trouble following what you are saying. Are you describing something that's happening now or will happen in the future ?
I'm unaware of any mass censoring apparatus that exists outside of authoritarian countries, such as China or North Korea.
> The deeper risk is not that AI will replace historians or translators, but that it will convince us we never needed them in the first place.
I think the bigger danger would be that they'd lose the unimportant grunt work that helped the field exist.
Fields need a large amount of consistent routine work to keep existing. Like when analog photography got replaced by digital. A photo lab can't just pay the bills with the few pro photographers that refuse to move to digital or have a specific need for analog. They needed a steady supply of cat pictures and terrible vacation photos, and when those dried up, things got tough.
So things like translation may go that way too -- those that need good quality translation understand the need very well, but industry was always supported by a lot of less demanding grunt work that now just went away.
"Convince us"? There's no need for that at all. We've done all that ourselves.
Just check the latest budgets for university history departments.
There are nice example how even after human input the translation misses things.
For example, the price of the fish was stated as 2.40 rubles. This is meaningless outside the context and does not explain why it was very expensive for the old man who checked the fish first. But if one knows that this was Soviet SF that was about a life in a small Soviet town of that time, then one also knows that a monthly pension was like 70-80 rubles so the fish cost was a daily income.
Then one needs to remember that the only payment method was cash and people did not go out with amount more than they would expect to spend to minimize the loss in case of thievery etc. and banking was non-existing in practice so people hold the savings in cash at home. That explains why Lozhkin went to home for the money.
Would you want a translator to somehow jam that context into the story? Otherwise, I fail to see how it's an issue of translation.
If I had learned Russian and read the story in the original language, I would be in the same position regardless.
It's pretty common for translators to do exactly that, usually via either footnotes or context created by deliberate word choice. Read classical translations for example and they'll often point out wordplay in the original language that doesn't quite work in translation. I've even seen that in subtitles.
LLMs tend to imitate that practice, e.g. Gemini seems to be doing that by default in its translations unless you stop it. The result is pretty poor though - it makes trivial things overly verbose and rarely gets the deeper cultural context. The knowledge is clearly here, if you ask it explicitly it does it much better, but the generalization ability is still nowhere near the required level, so it struggles to connect the dots on its own.
I was going to say that I'm not certain the knowledge is there and tried an experiment: If you give it a random bible passage in Greek, can it produce decent critical commentary of it? That's something it's certainly ingested mounds of literature on, both decent and terrible.
Checked with a few notable passages like the household codes and yeah, it does a decent (albeit superficial) job. That's pretty neat.
Sometimes when re-publishing an older text references are added to clarify the meaning that people would miss otherwise from the lack of knowledge of cultural references.
But here there is need to even put a references. A good translation may reword "too expensive!" into "what? I can live the whole day on that!" to address things like that.
Some translators may add notes on the bottom of the page for things like that.
It's going to greatly vary of course. Some will try to culturally adapt things. Maybe convert to dollars, maybe translate to "a day's wages", maybe translate as it is then add an explanatory note.
You might even get a preface explaining important cultural elements of the era.
This is also something that an LLM would be able to easily explain. I'm pretty confident that all 4o-level models know these facts without web search.
Not in my experience. Unless you explicitly prompt and bias that model for that kind of deep answer (which you won't unless you are already experienced in the field) you're going to get some sycophantic superficial dribble that's only slightly better than the wikipedia page.
This is a very tangential comment, but I read the short story (https://www.dropbox.com/scl/fi/8eh2woz05ndfxinbf9vdh/Goldfis...) and loved it (took me around 15 minutes to read).
Went down a bit of a rabbit hole on the original author, Kir Bulychev, and saw that he wrote many short stories set in Veliky Guslar (which explained the name Greater Bard). The overall tone is very very similar to R.K. Narayan's Malgudi Days (albeit without the fantastical elements of talking goldfish), which is a favorite of mine. If anyone wants to get into reading some easily approachable Indian English literature, I always point them to Narayan and Adiga (who wrote The White Tiger).
On that note, does anyone else have any recommendations on authors who make use of this device (small/mid-sized city which serves as a backdrop for an anthology of short stories from a variety of characters' perspectives)?
> Still, without AI a story like this would have taken me several weeks to translate and polish, instead of one afternoon.
I don't understand this. It's only six thousand words and it's the polishing that takes the time. How would it have taken weeks to do the initial draft?
I suppose different people work differently and the author knows themselves better than we do?
The article is written by LLM and published on Substack, so there's no expectation for it to make coherent points.
The translated story is full of implicit cultural references. If the author have used AI to clarify some references about what the story was about, it could explain the vast time saving.
Maybe he's... not great at Russian? I'm at a loss, same as you.
And I don't have any skill in Russian, but I would say that his translation is not good, or at least was not thoughtfully made, based solely on the fact that he did not write the author's name in it.
I don't understand: the translation is not good because he omitted the author's name? He stated it plainly in his article:
> As it happens, I recently translated a short story by Kir Bulychev — a Soviet science-fiction icon virtually unknown in the West.
I for one enjoyed reading it! As for the article, it's on point. There will be fewer historians and translators, but I suspect those that stick around will be greatly amplified.
In the PDF linked in the article there's only the title of the story, and not the author's name.
> I for one enjoyed reading it!
I gave a quick look and was surprised to see that the most of the first two paragraphs simply aren't there. I guess you've read something else!
As for machine translation: currently it isn't remotely ready to deal with literature by itself, but it could be handy to assist translators.
> The initial draft was terrible in conveying tone, irony, or any kind of cultural allusion.
My mother reads books mostly in Russian, including books by English-speaking authors translated into Russian.
Some of the translations are laughably bad; one recent example had to translate "hot MILF", and just translated "hot" verbatim - as in the adjective indicating temperature - and just transliterated the word "MILF", as the translator (or machine?) apparently just had no idea what it was, and didn't know the equivalent term in Russian.
As a mirror, I have a hard time reading things in Russian - I left when I was ten years old, so I'm very out of practice, and most of the cultural allusions go straight over my head as well. A good translation needs to make those things clear, either via a good translation, or by footnotes that explain things to the reader.
And this doesn't just apply to linguistic translation - the past is a foreign country, too. Reading old texts - any old texts - requires context.
As a Russian, горячая милфа is exactly how I’d translate it. I think a lot of slang got borrowed like this in 2000-s.
> and just translated "hot" verbatim - as in the adjective indicating temperature
Well, "горячий" does have figurative meaning "passionate" (and by transfer, "sexy") in Russian just as it has in English. Heck, English is even worse in this regard: "heated argument", seriously? Not only an argument doesn't have a temperature, you can't change it either (since it does not exist)! Yet the phrase exists just fine, and it translates as "hot argument" to Russian, no problem.
No comments on "MILF" though. But I wouldn't be surprised if it actually entered the (youth/Internet) slang as-is: many other English words did as well.
Reminds me of my Czech friend explaining some of the subtle humour in 'Kolya' which I would never have got just from the English subtitles.
One of the few articles about AI on the front page that doesn't make me want to throw my phone against a wall.
Haven't even read it completely, but in contrast to the countless submissions regurgitating badly thought-out meta arguments about AI-supported software engineering, it actually seems to elaborate on some interesting points.
I also think that the internet as a primary communication and mass medium + generative AI evokes 1984, very strongly.
Many historians work on manuscripts and/or large archives of documents that might not be digitized, let alone be accessible in the internet. The proportion of human knowledge that is available in the internet, especially if we further constrain to English-language and non-Darkweb or pirated, is greatly exaggerated. So there are infrastructure problems that LLMs by themselves don't solve.
On the other hand, people tend to be happy with a history that ignores 90+% of what happened, instead focusing on a "central" narrative, which traditionally focussed on maybe 5 Euro-Atlantic great powers, and nowadays somewhat pretends not to.
That being said, I don't like the subjectivist take on historical truth advanced by the article. Maybe it's hard to positively establish facts, but it doesn't mean one cannot negatively establish falsehoods and this matters more in practice, in the end. This feels salient when touching on opinions of Carr's as a Soviet-friendly historian.
> I don't like the subjectivist take on historical truth [...]
The work of historians is to make inferences based on incomplete and contradictory sources.
Historians aren't simple fact-checkers. They make judgements in an attempt to understand the sweep of history.
You can see what kind of work they have to do every time you stare at some bullshit narrative put out by a company about how really it was good for the local economy for them to run their fracking operation, and the waste water really was filtered three times so it couldn't be causing the overabundance of three-legged frogs, and last year they funded two scientific studies that prove it. (I just made this up, hope you get the idea.)
> But as Goethe said, every fact is already a theory. That’s why facts cannot save you. “Data” will not save you. There can be no objective story since the very act of assembling facts requires implicit beliefs about what should be emphasized and what should be left out.
There may be no objective story, but some stories and fact theories are more rigorous and thoroughly defended than others.
I'm sorry, as someone who genuinely likes AI, I still have to say that I have to call bullshit on Microsoft's study on this. I use ChatGPT all the time, but it's not going to "replace web developers" because that's almost a statement that doesn't even make sense.
You see all these examples like "I got ChatGPT to make a JS space invaders game!" and that's cool and all, but that's sort of missing a pretty crucial part: the beginning of a new project is almost always the easiest and most fun part of the project. Showing me a robot that can make a project that pretty much any intern could do isn't so impressive to me.
Show me a bot that can maintain a project over the course of months and update it based on the whims of a bunch of incompetent MBAs who scope creep a million new features and who don't actually know what they want, and I might start worrying. I don't know anything about the other careers so I can't speak to that, but I'd be pretty surprised if "Mathematician" is at severe risk as well.
Honestly, is there any reason for Microsoft to even be honest with this shit? Of course they want to make it look like their AI is so advanced because that makes them look better and their stock price might go up. If they're wrong, it's not like it matters, corporations in America are never honest.
> History is therefore a constant act of reinterpretation and triangulation, which is something that LLMs, as linguistic averaging machines, simply cannot do.
Well-put
Comment was deleted :(
here's the paper this guy seems to be reacting to https://arxiv.org/abs/2507.07935v1
> they are acts of interpretation that are never recognized as such by outsiders
And that is exactly why translators are getting replaced by ML/AI. Companies don't care about quality, that is the reason customer support was the first thing axed, companies see it only as a cost.
Comment was deleted :(
the more we can all dump EVERYTHING we got into the same giant mega AI,
the better off we will all be.
but of course, this goes directly against how so many people think, that I won't even bo
I agree and disagree. It's certainly the case that facts imply underlying epistemologies, but it completely misses the point to treat that like it entails catastrophic relativism.
Building up an epistemology isn't just recreational, ideally it's done for good reasons that are responsive to scrunity, standing firm on important principles and, where necessarily, conciliatory in response to epistemological conundrums. In short, such theories can be resilient and responsible, and facts based on them can inherent that resilience.
So I think it completely misses the point to think that "facts imply epistemologies" should have the upshot of destroying any conception of access to authoritative factual understanding. Global warming is still real, vaccines are still effective, sunscreen works, dinosaurs really existed. And perhaps, more to the point in this context, there really are better and worse understandings of the fall of Rome or the Dark Ages or Pompeii or the Iraq war.
If being accountable to the theory-laden epistemic status of facts means throwing the stability of our historical understanding into question, you're doing it wrong.
And, as it relates to the article, you're doing it super wrong if you think that creates an opening for a notion of human intuition that is fundamentally non-informational. I think it's definitely true that AI as it currently exists can spew out linguistically flat translations, lacking such things as an interpretive touch, or an implicit literary and cultural curiosity that breathes the fire of life and meaning into language as it is actually experienced by humans. That's a great and necessary criticism. But.
Hubert Dreyfus spent decades insisting that there were things "computers can't do", and that those things were represented by magical undefined terms that speak to ineffable human essence. He insisted, for instance, that computers performing chess at a high level would never happen because it required "insight", and he felt similarly about the kind of linguistic comprehension that has now, at least in part, been achieved by LLMs.
LLMs still fall short in critical ways, and losing sight of that would involve letting go of our ability to appreciate the best human work in (say) history, or linguistics. And there's a real risk that "good enough" AI can cause us to lose touch with such distinctions. But I don't think it follows that you have to draw a categorical line insisting such understanding is impossible, and in fact I would suggest that's a tragic misunderstanding that gets everything exactly backwards.
I agree with this whole-heartedly.
Certainly some facts can imply a certain understanding of the world, but they don't require that understanding in order to remain true. The map may require the territory, but the territory does not require the map.
“Reality is that which, when you stop believing in it, doesn't go away.” ― Philip K. Dick
In this analogy though, maps are the only things we have access to. There may be Truth, but we only approximate it with our maps.
We require the map.
This whole article is based on misinterpreting Microsoft's "AI applicability score" for "risk of job being made redundant by AI". From the original paper:
"This score captures if there is nontrivial AI usage that successfully completes activities corresponding to significant portions of an occupation’s tasks."
Then the author describes their job qualitatively matching their AI applicability score by using AI to do most of their work for them.
If there's a lot of unmet demand for low-priced high-quality translation, translators could end up having more work, not less.
AI represents a move to the massive centralization of information into a few sources. History (or it's "interpreters") should never be in the hands of these few powerful entities aligned with even more powerful governments.
This. Any centralization or gate keeping of speech and knowledge is a danger to free societies. The tech is moving too fast for politics to keep up. I wonder if the only fix is to elect younger people, even if they’re missing other types of knowledge and experience.
AI is the perfect propaganda technology.
I don't believe the current race to build AI is actually about any productivity gains (which are questionable at best).
I believe the true purpose of the outsized AI investments is to make sure the universal answer machine will give answers that conform to the ideology of the ruling class.
You can read hints of that in statements like the Trump AI Action Plan [0], but also things like the Llama 4 announcement. [1]
[0] "Ensure that Frontier AI Protects Free Speech and American Values" - https://www.whitehouse.gov/wp-content/uploads/2025/07/Americ...
[1] "It’s well-known that all leading LLMs have had issues with bias—specifically, they historically have leaned left when it comes to debated political and social topics. This is due to the types of training data available on the internet." https://ai.meta.com/blog/llama-4-multimodal-intelligence/
"This is due to the types of training data available on the internet."
I'd love to see them prove this, but they can't.
Why not? The fact that even the overtly Nazi Musk led xAI's Grok can't shake a "left" bias (left for the US political spectrum, which is to the right of the centre in most other developed countries) proves this wrong. The data, and reality, tend to lean "left" (again, on the American spectrum). It isn't leftists in the US talking about controlling the weather, Haitians eating pets, proposing to nuke hurricanes, dismissing the Holocaust, denying climate change exists, and other such objectively wrong things.
This saying exists for a reason: https://en.m.wikipedia.org/?redirect=no&title=Reality_has_a_...
Maybe it's not a left bias in the AI, but a right bias in the ruling class instead?
Fascism is the most viable option for the ruling class to sustain their status quo in a declining capitalistic world with increasing public unrest. The left, anti-capitalistic narrative therefor needs to scrubbed off the people.
The fact that it can both displace vast amounts of labor and also control information for people who won't bother with primary sources anymore is what explains (for me) the endless funding and desperation to make it happen.
> There can be no objective story since the very act of assembling facts requires implicit beliefs about what should be emphasized and what should be left out. History is therefore a constant act of reinterpretation and triangulation, which is something that LLMs, as linguistic averaging machines, simply cannot do.
Yeah, no. I find it funny how everyone from other specialties take offence when their piece of "advanced" whatever gets put on a list, but they have absolutely no issue with making uninformed, inaccurate and oversimplified remarks like "averaging machines".
Brother, these averaging machines just scored gold at IMO. Allow me to doubt that whatever you do is more impressive than that.
Oh, data gods! Oh, technocratic overlords! Milords, shant though giveth but a crumb of cryptocurrency to thy humble guzzler?
I mean, I get the sarcasm, but don't get the cryptobabble. And this isn't about data or technoanything in particular. In order to get gold at IMO the system had to
a) "solve" NLP enough to understand the problem b) reason through various "themes", ideas, partial demonstrations and so on c) verify some d) gather the good ideas from all the tried paths and come up with the correct demonstrations in the end
Now tell me a system like this can't take source material and all the expert writings so far, and come up with various interpretations based on those combinations. And tell me it'll be less accurate than some historian's "vibes". Or a translator's "feelings". I don't buy it.
I dunno I can see an argument that something like IMO word problems are categorically a different language space than a corpus of historiography. For one, even when expressed in English language math is still highly, highly structured. Definitions of terms are totally unambiguous, logical tautologies can be expressed using only a few tokens, etc. etc. It's incredibly impressive that these rich structures can be learned by such a flexible model class, but it definitely seems closer (to me) to excelling at chess or other structured game, versus something as ambiguous as synthesis of historical narratives.
> Now tell me a system like this can't take source material and all the expert writings so far, and come up with various interpretations based on those combinations. And tell me it'll be less accurate than some historian's "vibes".
Framing it as the kind of problem where accuracy is a well-defined concept is the error this article is talking about. Literally the historian's "vibes" and "feelings" are the product you're trying to mimic with the LLM output, not an error to be smoothed out. I have no doubt that LLMs can have real impact in this field, especially as turbopowered search engines and text-management tools. But the point of human narrative history is fundamentally that we tell it to ourselves, and make sense of it by talking about it. Removing the human from the loop is IMO like trying to replace the therapy client with a chat agent.
Okay, shapecel.
> One, there is no way LLMs in their current state are capable of replacing human translators for a case like this. And two, they do make the job of translation a lot less tedious. I wouldn't call it pleasant exactly, but it was much easier than my previous translation experiences
On the other hand, one day they will replace human beings. And secondly, if something like transalation (or in general, any mental work) becomes too easy, then we also run the risk of incresing the amount of mediocre works. Fact is, if something is hard, we'll only spend time on it if it's really worthwhile.
Same thing happens with phone cameras. Yes, it makes some things more convenient, but it also has resulted in a mountain of mediocrity, which isn't free to store (requires energy and hence pollutes the environment).
Technically modern LLMs are handicapped on translation tasks compared to the original transformer architecture. The origami transformer got to see future context as well as past tokens.
Okay, but I'm not really concerned with the state of the art now with any specific technology, but what will be the state of the art in 20 years.
Crafted by Rajat
Source Code