understand your frustration. i trust you also understand the models have some dark corners that someone could use to misrepresent the goals of our project. if you have ideas on how we could make the models more broadly accessible while avoiding that risk, please do reach out @ history-llms@econ.uzh.ch
Ok...
So as a black person should I demand that all books written before the civil rights act be destroyed?
The past is messy. But it's the only way to learn anything.
All an LLM does it's take a bunch of existing texts and rebundle them. Like it or not, the existing texts are still there.
I understand an LLM that won't tell me how to do heart surgery. But I can't fear one that might be less enlightened on race issues. So many questions to ask! Hell, it's like talking to older person in real life.
I don't expect a typical 90 year old to be the most progressive person, but they're still worth listening too.
we're on the same page.
Although...
Self preservation is the first law of nature. If you release the model someone will basically say you endorse those views and you risk your funding being cut.
You created Pandora's box and now you're afraid of opening it.
They could add a text box where users have to explicitly type the following words before it lets them interact in any way with the model: "I understand this model was created with old texts so any racial or sexual statements are a byproduct of their time an do not represent in any way the views of the researchers".
That should be more than enough to clear any chance of misunderstanding.
I would claim the public can easily handle something like this, but the media wouldn't be able to resist.
I could easily see a hit piece making its rounds on left leaning media about the AI that re-animates the problematic ideas of the past. "Just look at what it said to my child, "<insert incredibly racist quote coerced out of the LLM here>"!" Rolling stones would probably have a front page piece on it, titled "AI resurrecting racism and misogyny". There would easily be enough there to attract death threats to the developers, if it made its rounds on twitter.
"Platforming ideas" would be the issue that people would have.
i think we (whole section) are just talking past each other - we never said we'll lock it away. it was an announcement of a release, not a release. main purpose for us was getting feedback on the methodological aspects, as we clearly state. i understand you guys just wanted to talk to the thing though.
Of course, I have to assume that you have considered more outcomes than I have. Because, from my five minutes of reflection as a software geek, albeit with a passion for history, I find this the most surprising thing about the whole project.
I suspect restricting access could equally be a comment on modern LLMs in general, rather than the historical material specifically. For example, we must be constantly reminded not to give LLMs a level of credibility that their hallucinations would have us believe.
But I'm fascinated by the possibility that somehow resurrecting lost voices might give an unholy agency to minds and their supporting worldviews that are so anachronistic that hearing them speak again might stir long-banished evils. I'm being lyrical for dramatic affect!
I would make one serious point though, that do I have the credentials to express. The conversation may have died down, but there is still a huge question mark over, if not the legality, but certainly the ethics of restricting access to, and profiting from, public domain knowledge. I don't wish to suggest a side to take here, just to point out that the lack of conversation should not be taken to mean that the matter is settled.
They aren't afraid of hallucinations. Their first example is a hallucination, an imaginary biography of a Hitler who never lived.
Their concern can't be understood without a deep understanding of the far left wing mind. Leftists believe people are so infinitely malleable that merely being exposed to a few words of conservative thought could instantly "convert" someone into a mortal enemy of their ideology for life. It's therefore of paramount importance to ensure nobody is ever exposed to such words unless they are known to be extremely far left already, after intensive mental preparation, and ideally not at all.
That's why leftist spaces like universities insist on trigger warnings on Shakespeare's plays, why they're deadly places for conservatives to give speeches, why the sample answers from the LLM are hidden behind a dropdown and marked as sensitive, and why they waste lots of money training an LLM that they're terrified of letting anyone actually use. They intuit that it's a dangerous mind bomb because if anyone could hear old fashioned/conservative thought, it would change political outcomes in the real world today.
Anyone who is that terrified of historical documents really shouldn't be working in history at all, but it's academia so what do you expect? They shouldn't be allowed to waste money like this.
You know, I actually sympathize with the opinion that people should be expected and assumed to be able to resist attempts to convince them of being nazis.
The problem with it is, it already happened at least once. We know how it happened. Unchecked narratives about minorities or foreigners is a significant part of why the 20th century happened to Europe, and it’s a significant part of why colonialism and slavery happened to other places.
What solution do you propose?
Studying history better would be a good start. The Nazis came to power because they were a far left party and the population in that era thought socialism was a great idea. Hitler himself remarked many times that his movement was left wing and socialist. I expect that if you asked the LLM trained on pre-1940s text, it would have no difficulty in explaining that.
By studying history better, people wouldn't draw the wrong conclusions about what caused it. Watch out for left wing radicals promoting socialism-with-genetic-characteristics.
They said it plainly ("dark corners that someone could use to misrepresent the goals of our project"): they just don't want to see their project in headlines about "Researchers create racist LLM!".
They already represented the goals of their project clearly, and gave examples of outputs. Anyone can already misrepresent it. That isn't their true concern.
I'm not sure I do. It feels like someone might for example have compiled a full library of books, newspapers and other writing from that era, only to then limit access to that library, doing the exact censorship I imagine the project was started to alleviate.
Now were it limited in access to ask money to compensate for the time and money spent compiling the library (or training the model), sure, I'd somewhat understand. Not agree but understand.
Now it just feels like you want to prevent your model name being associated with the one guy who might use it to create a racist slur Twitter bot. There's plenty of models for that already. At least the societal balance of a model like this would also have enough weight on the positive side to be net positive.
There's no such risk so you're not going to get any sensible ideas in response to this question. The goals of the project are history, you already made that clear. There's nothing more that needs to be done.
We all get that academics now exist in some kind of dystopian horror where they can get transitively blamed for the existence of anyone to the right of Lenin, but bear in mind:
1. The people who might try to cancel you are idiots unworthy of your respect, because if they're against this project, they're against the study of history in its entirety.
2. They will scream at you anyway no matter what you do.
3. You used (Swiss) taxpayer funds to develop these models. There is no moral justification for withholding from the public what they worked to pay for.
You already slathered your README with disclaimers even though you didn't even release the model at all, just showed a few examples of what it said - none of which are in any way surprising. That is far more than enough. Just release the models and if anyone complains, politely tell them to go complain to the users.
Yet your project relies on letting an llm synthesize historical documents and presenting itself as some sort of expert from the time? You are aware of the hallucination rates surely but don't care whether the information your university presents is accurate or are you going to monitor all output from your llm?
What are the legal or other ramifications of people misrepresenting the goals of your project? What is it you're worried about exactly?
This is understandable and I think others ITT should appreciate the legal and PR ramifications involved.
A disclaimer on the site that you are not bigoted or genocidal, and that worldviews from the 1913 era were much different than today and don't necessarily reflect your project.
Movie studios have done that for years with old movies. TCM still shows Birth of a Nation and Gone with the Wind.
Edit: I saw further down that you've already done this! What more is there to do?
[flagged]