Around the time J. Robert Oppenheimer learned that Hiroshima had been struck (alongside everyone else in the world) he began to have profound regrets about his role in the creation of that bomb. At one point when meeting President Truman Oppenheimer wept and expressed that regret. Truman called him a crybaby and said he never wanted to see him again. And Christopher Nolan is hoping that when Silicon Valley audiences of his film Oppenheimer (out June 21) see his interpretation of all those events they’ll see something of themselves there too.
After a screening of Oppenheimer at the Whitby Hotel yesterday Christopher Nolan joined a panel of scientists and Kai Bird, one of the authors of the book Oppenheimer is based on to talk about the film, American Prometheus. The audience was filled mostly with scientists, who chuckled at jokes about the egos of physicists in the film, but there were a few reporters, including myself, there too.
We listened to all too brief debates on the success of nuclear deterrence and Dr. Thom Mason, the current director of Los Alamos, talked about how many current lab employees had cameos in the film because so much of it was shot nearby. But towards the end of the conversation the moderator, Chuck Todd of Meet the Press, asked Nolan what he hoped Silicon Valley might learn from the film. “I think what I would want them to take away is the concept of accountability,” he told Todd.
“Applied to AI? That’s a terrifying possibility. Terrifying.”
He then clarified, “When you innovate through technology, you have to make sure there is accountability.” He was referring to a wide variety of technological innovations that have been embraced by Silicon Valley, while those same companies have refused to acknowledge the harm they’ve repeatedly engendered. “The rise of companies over the last 15 years bandying about words like ‘algorithm,’ not knowing what they mean in any kind of meaningful, mathematical sense. They just don’t want to take responsibility for what that algorithm does.”
He continued, “And applied to AI? That’s a terrifying possibility. Terrifying. Not least because as AI systems go into the defense infrastructure, ultimately they’ll be charged with nuclear weapons and if we allow people to say that that’s a separate entity from the person’s whose wielding, programming, putting AI into use, then we’re doomed. It has to be about accountability. We have to hold people accountable for what they do with the tools that they have.”
While Nolan didn’t refer to any specific company it isn’t hard to know what he’s talking about. Companies like Google, Meta and even Netflix are heavily dependent on algorithms to acquire and maintain audiences and often there are unforeseen and frequently heinous outcomes to that reliance. Probably the most notable and truly awful being Meta’s contribution to genocide in Myanmar.
“At least is serves as a cautionary tale.”
While an apology tour is virtually guaranteed now days after a company’s algorithm does something terrible the algorithms remain. Threads even just launched with an exclusively algorithmic feed. Occasionally companies might give you a tool, as Facebook did, to turn it off, but these black box algorithms remain, with very little discussion of all the potential bad outcomes and plenty of discussion of the good ones.
“When I talk to the leading researchers in the field of AI they literally refer to this right now as their Oppenheimer moment,” Nolan said. “They’re looking to his story to say what are the responsibilities for scientists developing new technologies that may have unintended consequences.”
“Do you think Silicon Valley is thinking that right now?” Todd asked him.
“They say that they do,” Nolan replied. “And that’s,” he chuckled, “that’s helpful. That at least it’s in the conversation. And I hope that thought process will continue. I’m not saying Oppenheimer’s story offers any easy answers to these questions. But at least it serves a cautionary tale.”
We’re in the same situation. If we don’t build it, our enemies will build it first.
The difference is that Oppenheimer was ostensibly in a race against a fascist regime to get the bomb, with the fate of the free world hanging in the balance.
Zuck and Musk and Jeff just want to make more money.
Zuck and Musk and Jeff just want to make more money.
Are you sure about that? None of the
threefour listed below would ever have to get out of bed ever again, pills, powders and prostitutes included. How much money is enough for one person?Net worth as of 2023-07-17:
Mark Zuckerberg: USD$109.4B
- 1175 Trident missiles
Elon Musk: USD$250.4B
- 2689 Trident missiles
Jeff Bezos: USD$157.3B
- 1689 Trident missiles
Just-for-fun Bonus net worth as of 2023-07-17:
Mackenzie Scott (ex-wife of Jeff Bezos): USD$36.1B
- 387 Trident missiles1 Trident missile = $93,100000 (adjusted for inflation). Source.
Don’t think it’s ever just “want to make more money”. Only Uncle Scrooge wants that.
Oh, BTW, which “fascist regime” was Oppenheimer in a race with again? Careful now.
I don’t see why calling the litteral Nazis a fascist regim is somehow a problem?
The Second World War in Europe officially ended on May 8, 1945: German forces had already surrendered on most battlefronts, Mussolini had already been executed and Hitler committed suicide a week before the official surrender. Good riddance to bad rubbish—it pains me to have to clarify this. The Americans, Soviets and Europeans had already started kidnapping, coercing and collecting Nazi scientists like baseball cards. Japan soon after had sued for peace. The Manhattan Project had been completed only after the surrender of German and Italian forces.
So to address the “race” question[1]: Dr. Oppenheimer and the United States’ Manhattan Project were “ostensibly in a race against nobody.” The goalposts then moved, it was no longer Axis powers the Allies were in a race with; it was the Soviets and later the CCP, both Allies…until they weren’t. On who was there to test all this spent time, effort and money? Although advised against it by his own generals, United States President Truman said “Hiroshima.” Then after three days, Nagasaki.
While he was “working against the Nazis and the Japanese[2]” his own government had paranoically conspired against him (and his wife, friends and colleagues) spying on him since the 1930s, finally ending with his security clearance revoked in 1954.
The United States government issued loyalty tests to its workers with the following questions among the many:
- Is it proper to mix white and Negro blood plasma?
- There is a suspicion in your record that you are in sympathy with the underprivileged. Is that true?
- What were your feelings at that time concerning race equality?
- Have you ever made statements about the “downtrodden masses” and “underprivileged people”?
If I hadn’t already revealed this was the United States of America issuing these questions, to whom might you have attached such sentiment?
So, you’ll have to pardon me if I questioned exactly who this race was against or which fascist regime. I will not abide simplistic, populist phrases such as “Oppenheimer was ostensibly in a race against a fascist regime to get the bomb, with the fate of the free world hanging in the balance.” I can just picture the cover to that comic book!
HistoryReality is never cut and dried and sloganized like that.Oh, darn. We didn’t even touch on why those rocket-owning billionaires might possibly be interested in AI.
He is spot on.
Algorithms and AI aren’t even any different. AI is literally a complex system of nonlinear functions. It’s not black magic.
If I wrote a traditional nonlinear alto with computer optimized parameters it only differs from ML models in that it’s less complex. Not understanding your product is not a defense.
The problem is we have relied on self-training neural network models which are a black box to us.
The networks are numbers. Tons and tons of numbers. Weights are distributed throughout the neurons. And we don’t know what the numbers mean, why they are the way they are, or what they do.
The problem is we don’t know how they work. And until we can explain the decisions they make, we should be very cautious using them.
I am very, very, very skeptical that any modern "AI"s are intelligent at all. I don’t think they behave like intelligence. I’m more of a SALAMI believer. But people are using these LLM bots to do real work and make decisions without understanding how they are coming up with their answers, and that is dangerous. It’s not dangerous because they’ll become sentient and take over the world. It’s dangerous because we don’t know that these algorithms are ethically sound tools to use and no one can be held accountable if they aren’t.
That was hilarious. Thanks for sharing the link on SALAMI. I definitely had some bias and misunderstandings when thinking about AI.