Monday, November 25, 2024

Baghdad

OpenAI insiders blast lack of AI transparency

 OpenAI insiders blast lack of AI transparency

The open letter criticizing AI transparency comes amid questions about OpenAI CEO Sam Altman’s corporate leadership

San Francisco – A group of current and former employees from OpenAI on Tuesday issued an open letter warning that the world’s leading artificial intelligence companies were falling short of necessary transparency and accountability to meet the potential risks posed by the technology.

The letter raised serious concerns about AI safety risks “ranging from the further entrenchment of existing inequalities, to manipulation and misinformation, to the loss of control of autonomous AI systems potentially resulting in human extinction.”

The 16 signatories, which also included a staff member from Google DeepMind, warned that AI companies “have strong financial incentives to avoid effective oversight” and that self-regulation by the companies would not effectively change this.

“AI companies possess substantial non-public information about the capabilities and limitations of their systems, the adequacy of their protective measures, and the risk levels of different kinds of harm,” the letter said.

“However, they currently have only weak obligations to share some of this information with governments, and none with civil society. We do not think they can all be relied upon to share it voluntarily.”

That reality, the letter added, meant that employees inside the companies were the only ones who could notify the public, and the signatories called for broader whistleblower laws to protect them.

“Broad confidentiality agreements block us from voicing our concerns, except to the very companies that may be failing to address these issues,” the letter said.

The four current employees of OpenAI signed the letter anonymously because they feared retaliation from the company, The New York Times reported.

It was also signed by Yoshua Bengio, Geoffrey Hinton and Stuart Russell, who are often described as AI “godfathers” and have criticized the lack of preparation for AI’s dangers.

OpenAI in a statement pushed back at the criticism.

“We’re proud of our track record of providing the most capable and safest AI systems and believe in our scientific approach to addressing risk,” a statement said.

“We agree that rigorous debate is crucial given the significance of this technology and we’ll continue to engage with governments, civil society and other communities around the world.”

OpenAI also said it had “avenues for employees to express their concerns including an anonymous integrity hotline” and a newly formed Safety and Security Committee led by members of the board and executives, including CEO Sam Altman.

The criticism of OpenAI, which was first released to the Times, comes as questions are growing around Altman’s leadership of the company.

OpenAI has unveiled a wave of new products, though the company insists they will only get released to the public after thorough testing.

An unveiling of a human-like chatbot caused a controversy when Hollywood star Scarlett Johansson complained that it closely resembled her voice.

She had previously turned down an offer from Altman to work with the company.