Добавить новость
News in English





160*600

Новости сегодня на DirectAdvert

Новости сегодня от Adwile

Актуальные новости сегодня от ValueImpression.com


Опубликовать свою новость бесплатно - сейчас


The art of the squeal: What we can learn from the flood of AI resignation letters

Corporate resignations rarely make news, except at the highest levels. But in the last two years, a spate of X posts, Substack open letters, and public statements from prominent artificial intelligence researchers have created a new literary form — the AI resignation letter — with each addition becoming an event to be mined for meaning. Together, the canon of these letters — some of them apparently bound by non-disclosure agreements and other loyalties, legally compelled or not — tells us a lot about how some of the top people in AI see themselves and the trajectory of their industry. Overall, the image is bleak.

This past week brought several additions to the annals of "Why I quit this incredibly valuable company working on bleeding-edge tech" letters, including from researchers at xAI and an op-ed in The New York Times from a departing OpenAI researcher. Perhaps the most unusual was by Mrinank Sharma, who was put in charge of Anthropic's Safeguards Research Team a year ago, and who announced his departure from what is often considered the more safety-minded of the leading AI startups. He posted a 778-word letter on X that was at times romantic and brooding — he quoted the poets Rainer Maria Rilke and Mary Oliver. Opining on AI safety, his own experiences working on AI sycophancy and "AI-assisted bioterrorism," and the "poly-crisis" consuming our society, the letter had three footnotes and some ominous, if vague, warnings.

"We appear to be approaching a threshold where our wisdom must grow in equal measure to our capacity to affect the world, lest we face the consequences," Sharma wrote. "Throughout my time here, I've repeatedly seen how hard it is to truly let our values govern our actions."

Sharma noted that his final project at Anthropic was "on understanding how Al assistants could make us less human or distort our humanity" — a nod, perhaps, to the scourge of AI psychosis and other novel harms emerging from people overvaluing their relationships with chatbots. He said that he didn't know what he was going to do next, but expressed a desire to pursue "a poetry degree and devote myself to the practice of courageous speech." The researcher ended by including the full text of "The Way It Is" by the poet William Stafford.

In the annals of AI resignations, Sharma's missive might be less dramatic than the boardroom coup that ousted OpenAI CEO Sam Altman for five days in November 2023. It's less troubling than some of the other end-of-days warnings published by AI safety researchers who quit their posts believing that their employers weren't doing enough to mitigate the potential harms of artificial general intelligence, or AGI, a smarter-than-human intelligence that AI companies are racing to build. (Some AI experts question whether AGI is even achievable or what it might mean.)

But Sharma's note captures the deep attachments that top AI researchers — who are extremely well-compensated and work together in small teams — feel to their work, their colleagues, and, often, their employers. It also exposes some of the tensions that we see cropping up again and again in these resignation announcements. At top AI labs, there's an intense competition for resources between research/safety teams and people working on consumer-facing AI products. (Few, if any, public resignations seem to come from people on the product side.) There are pressures to ship without proper testing, established safeguards, or knowing what might happen when a system goes rogue. And there's a deep sense of mission and purpose that can sometimes be upended by feelings of betrayal.

Many of the people who have publicly quit AI companies work in safety and "alignment," the field tasked with making sure that AI capabilities align with human needs and welfare. Many of them seem very optimistic about AI, and even AGI, but they worry that financial pressures are eating away at safeguards. Few seem to be giving up on the field entirely — except perhaps Sharma, the aspiring poet. Either they jump ship for another seven-, eight-, or nine-figure job at a competing AI startup, or they become civic-minded AI analysts and researchers at one of a growing number of AI think tanks.

Sam Altman

All of them seem to be worried that either epic gains or epic disasters lie ahead. Announcing his departure from Anthropic to become OpenAI's Head of Preparedness earlier this month, Dylan Scandinaro wrote on LinkedIn, "AI is advancing rapidly. The potential benefits are great — and so are the risks of extreme and even irrecoverable harm." Daniel Kokotajlo, who resigned from OpenAI, said that OpenAI's systems "could be the best thing that has ever happened to humanity, but it could also be the worst if we don't proceed with care."

Recently, xAI, where co-founder Elon Musk is notorious for tinkering with the proverbial dials of the Grok chatbot, has seen a half-dozen members of its founding team leave. But the locus of the AI resignation letter, as a kind of industry artifact, is the red-hot startup OpenAI, where major figures, including top executives and safety-minded researchers, have been leaving for the last two years. Some resigned; some were fired; some were described in the press as "forced out" over internal company disputes. Seven left in a short period in the first half of 2024.

With revenue paling compared to its massive and growing infrastructure costs, OpenAI recently announced that it would begin incorporating ads into ChatGPT. That caused researcher Zoë Hitzig to quit. This week, she published a resignation letter in the Times, warning about the potential implications of ads becoming part of the substrate of chatbot conversations. "ChatGPT users have generated an archive of human candor that has no precedent, in part because people believed they were talking to something that had no ulterior agenda," she wrote. But, she warned, OpenAI seemed prepared to leverage that "archive of human candor" — much as Facebook had done — to target ads and undermine user autonomy. In the service of maximizing engagement, consumers might be manipulated — the classic sin of the modern internet.

If you think you are building a world-changing invention, you need to be able to trust your leadership. That's been a problem at OpenAI. On November 17, 2023, Altman was dramatically fired by the company's board because, it claimed, Altman was "not consistently candid in his communications with the board." Less than a week later, he performed his own boardroom coup and was reinstated, before consolidating his power. The exodus proceeded from there.

On May 14, 2024, OpenAI co-founder Ilya Sutskever announced his resignation. Sutskever was replaced as head of OpenAI's superalignment team by John Schulman, another company co-founder. A few months later, Schulman left OpenAI for Anthropic. Six months later, he announced his move to Thinking Machines Lab, an AI startup founded by former OpenAI CTO Mira Murati, who had replaced Altman as OpenAI's interim CEO during his brief firing.

The day after Sutskever left OpenAI, Jan Leike, who also helped head OpenAI's alignment work, announced on X that he had resigned. "OpenAI is shouldering an enormous responsibility on behalf of all of humanity," Leike wrote, but the company's "safety culture and processes have taken a backseat to shiny products." He thought that "OpenAI must become a safety-first AGI company." Less than two weeks later, Leike was hired by Anthropic. OpenAI and Antrhopic did not respond to requests for comment.

At OpenAI, departing researchers have said that the experts concerned with alignment and safety have often been sidelined, pushed out, or scattered among other teams, leaving researchers with the sense that AI companies are sprinting to build an invention they won't be able to control. "In short, neither OpenAI nor any other frontier lab is ready, and the world is also not ready" for AGI, wrote Miles Brundage when he resigned from OpenAI's AGI readiness team in 2024. Yet he added that "working at OpenAI is one of the most impactful things that most people could hope to do" and did not directly criticize the company. Brundage now runs AVERI, an AI research institute.

Across the AI industry, the story is much the same. In public pronouncements, top researchers gently chastise or occasionally denounce their employers for pursuing a potentially apocalyptic invention while also emphasizing the necessity of doing that research. Sometimes they offer a "cryptic warning" that leaves AI watchers scratching their heads. A few do seem genuinely alarmed at what's happening. When OpenAI safety researcher Steven Adler left the company in January 2025, he wrote that he was "pretty terrified by the pace of AI development" and wondered if it would wipe out humanity.

Yet in the many AI resignation letters, there's little discussion of how AI is being used right now. Data center construction, resource consumption, mass surveillance, ICE deportations, weapons development, automation, labor disruption, the proliferation of slop, a crisis in education — these are the areas where many people see AI affecting their lives, sometimes for the worse, and the industry's pious resignees don't have much to say about it all. Their warnings about some disaster just beyond the horizon become fodder for the tech press — and de facto cover letters for their next industry job — while failing to reach the broader public.

"Tragedies happen; people get hurt or die; and you suffer and get old," wrote William Stafford in the poem that Mrinank Sharma shared. It's a terrible thing, especially the tones of passivity and inevitability — resignation, you might call it. It can feel as if no single act of protest is enough, or, as Stafford writes in the next line: "Nothing you do can stop time's unfolding."


Jacob Silverman is a contributing writer for Business Insider. He is the author, most recently, of "Gilded Rage: Elon Musk and the Radicalization of Silicon Valley."

Read the original article on Business Insider


Загрузка...

Читайте на сайте

Другие проекты от 123ru.net








































Другие популярные новости дня сегодня


123ru.net — быстрее, чем Я..., самые свежие и актуальные новости Вашего города — каждый день, каждый час с ежеминутным обновлением! Мгновенная публикация на языке оригинала, без модерации и без купюр в разделе Пользователи сайта 123ru.net.

Как добавить свои новости в наши трансляции? Очень просто. Достаточно отправить заявку на наш электронный адрес mail@29ru.net с указанием адреса Вашей ленты новостей в формате RSS или подать заявку на включение Вашего сайта в наш каталог через форму. После модерации заявки в течении 24 часов Ваша лента новостей начнёт транслироваться в разделе Вашего города. Все новости в нашей ленте новостей отсортированы поминутно по времени публикации, которое указано напротив каждой новости справа также как и прямая ссылка на источник информации. Если у Вас есть интересные фото Вашего города или других населённых пунктов Вашего региона мы также готовы опубликовать их в разделе Вашего города в нашем каталоге региональных сайтов, который на сегодняшний день является самым большим региональным ресурсом, охватывающим все города не только России и Украины, но ещё и Белоруссии и Абхазии. Прислать фото можно здесь. Оперативно разместить свою новость в Вашем городе можно самостоятельно через форму.



Новости 24/7 Все города России




Загрузка...


Топ 10 новостей последнего часа






Персональные новости

123ru.net — ежедневник главных новостей Вашего города и Вашего региона. 123ru.net - новости в деталях, свежий, незамыленный образ событий дня, аналитика минувших событий, прогнозы на будущее и непредвзятый взгляд на настоящее, как всегда, оперативно, честно, без купюр и цензуры каждый час, семь дней в неделю, 24 часа в сутки. Ещё больше местных городских новостей Вашего города — на порталах News-Life.pro и News24.pro. Полная лента региональных новостей на этот час — здесь. Самые свежие и популярные публикации событий в России и в мире сегодня - в ТОП-100 и на сайте Russia24.pro. С 2017 года проект 123ru.net стал мультиязычным и расширил свою аудиторию в мировом пространстве. Теперь нас читает не только русскоязычная аудитория и жители бывшего СССР, но и весь современный мир. 123ru.net - мир новостей без границ и цензуры в режиме реального времени. Каждую минуту - 123 самые горячие новости из городов и регионов. С нами Вы никогда не пропустите главное. А самым главным во все века остаётся "время" - наше и Ваше (у каждого - оно своё). Время - бесценно! Берегите и цените время. Здесь и сейчас — знакомства на 123ru.net. . Разместить свою новость локально в любом городе (и даже, на любом языке мира) можно ежесекундно (совершенно бесплатно) с мгновенной публикацией (без цензуры и модерации) самостоятельно - здесь.



Загрузка...

Загрузка...

Экология в России и мире




Путин в России и мире

Лукашенко в Беларуси и мире



123ru.netмеждународная интерактивная информационная сеть (ежеминутные новости с ежедневным интелектуальным архивом). Только у нас — все главные новости дня без политической цензуры. "123 Новости" — абсолютно все точки зрения, трезвая аналитика, цивилизованные споры и обсуждения без взаимных обвинений и оскорблений. Помните, что не у всех точка зрения совпадает с Вашей. Уважайте мнение других, даже если Вы отстаиваете свой взгляд и свою позицию. Smi24.net — облегчённая версия старейшего обозревателя новостей 123ru.net.

Мы не навязываем Вам своё видение, мы даём Вам объективный срез событий дня без цензуры и без купюр. Новости, какие они есть — онлайн (с поминутным архивом по всем городам и регионам России, Украины, Белоруссии и Абхазии).

123ru.net — живые новости в прямом эфире!

В любую минуту Вы можете добавить свою новость мгновенно — здесь.






Здоровье в России и мире


Частные объявления в Вашем городе, в Вашем регионе и в России






Загрузка...

Загрузка...





Друзья 123ru.net


Информационные партнёры 123ru.net



Спонсоры 123ru.net