Sam Altman is credited for founding ChatGPT during his tenure as the CEO of OpenAI. / Photo: AFP

Hundreds of staff at OpenAI have threatened to quit the leading artificial intelligence company on Monday and join Microsoft.

They would follow OpenAI co-founder Sam Altman, who said he was starting an AI subsidiary at Microsoft following his shock sacking from the company whose ChatGPT chatbot has led the rapid rise of artificial intelligence technology.

In a letter, some of OpenAI's most senior staff members threatened to leave the company if the board did not get replaced.

"Your actions have made it obvious that you are incapable of overseeing OpenAI," said the letter, which was first released to Wired.

500 sign letter

Included in the list of names of signers was Ilya Sutskever, the company's chief scientist and one of members of the four-person board that voted to oust Altman.

It also included top executive Mira Murati, who was appointed to replace Altman as CEO when he was removed on Friday, but was herself demoted over the weekend.

"Microsoft has assured us that there are positions for all OpenAI employees at this new subsidiary should we choose to join," the letter said.

Reports said as many as 500 of OpenAI's 770 employees signed the letter.

ChatGPT

OpenAI has appointed Emmett Shear, a former chief executive of Amazon's streaming platform Twitch, as its new CEO despite pressure from Microsoft and other major investors to reinstate Altman.

The startup's board sacked Altman on Friday, with US media citing concerns that he was underestimating the dangers of its tech and leading the company away from its stated mission – claims his successor has denied.

Microsoft CEO Satya Nadella wrote on X that Altman "will be joining Microsoft to lead a new advanced AI research team," along with OpenAI co-founder Greg Brockman and other colleagues.

Altman shot to fame with the launch of ChatGPT last year, which ignited a race to advance AI research and development, as well as billions being invested in the sector.

High-profile departures

His sacking triggered several other high-profile departures from the company, as well as a reported push by investors to bring him back.

"We are going to build something new and it will be incredible. The mission continues," Brockman said, tagging former director of research Jakub Pachocki, AI risk evaluation head Aleksander Madry, and longtime researcher Szymon Sidor.

But OpenAI stood by its decision in a memo sent to employees on Sunday night, saying "Sam's behavior and lack of transparency... undermined the board's ability to effectively supervise the company," The New York Times reported.

'Badly' handled sacking

Shear confirmed his appointment as OpenAI's interim CEO in a post on X on Monday, while also denying reports that Altman had been fired over safety concerns regarding the use of AI technology.

"Today I got a call inviting me to consider a once-in-a-lifetime opportunity: to become the interim CEO of @OpenAI. After consulting with my family and reflecting on it for just a few hours, I accepted," he wrote.

"Before I took the job, I checked on the reasoning behind the change. The board did not remove Sam over any specific disagreement on safety, their reasoning was completely different from that.

"It's clear that the process and communications around Sam's removal has been handled very badly, which has seriously damaged our trust," Shear added.

More than $10 billion investment

Global tech titan Microsoft has invested more than $10 billion in OpenAI and has rolled out the AI pioneer's tech in its own products.

Microsoft's Nadella added in his post that "we look forward to getting to know Emmett Shear and OAI's new leadership team and working with them."

"We remain committed to our partnership with OpenAI and have confidence in our product roadmap," he said.

OpenAI is in fierce competition with others including Google and Meta, as well as start-ups like Anthropic and Stability AI, to develop its own AI models.

Dangers of misuse

Generative AI platforms such as ChatGPT are trained on vast amounts of data to enable them to answer questions, even complex ones, in human-like language.

They are also used to generate and manipulate imagery.

But the tech has triggered warnings about the dangers of its misuse – from blackmailing people with "deepfake" images to the manipulation of images and harmful disinformation.

AFP