OpenAI Accelerates Development of Government AI Licensing Requirements
Category Machine Learning Tuesday - July 25 2023, 05:46 UTC - 1 year ago OpenAI's internal policy memo supports the idea of requiring government licenses for advanced AI systems and suggests they are willing to make more data available to train image generators. They caution that their proposals will be different to those in the White House’s public initiative. OpenAI recognises they don't have all the answers and encourages the public to comment on how to cautiously regulate the industry.
An internal policy memo drafted by OpenAI shows the company supports the idea of requiring government licenses from anyone who wants to develop advanced artificial intelligence systems. The document also suggests the company is willing to pull back the curtain on the data it uses to train image generators. The creator of ChatGPT and DALL-E laid out a series of AI policy commitments in the internal document following a May 4 meeting between White House officials and tech executives including OpenAI Chief Executive Officer Sam Altman. "We commit to working with the U.S. government and policy makers around the world to support development of licensing requirements for future generations of the most highly capable foundation models," the San Francisco-based company said in the draft.
The idea of a government licensing system co-developed by AI heavyweights such as OpenAI sets the stage for a potential clash with startups and open-source developers who may see it as an attempt to make it more difficult for others to break into the space. It's not the first time OpenAI has raised the idea: During a U.S. Senate hearing in May, Altman backed the creation of an agency that, he said, could issue licenses for AI products and yank them should anyone violate set rules.
The policy document comes just as Microsoft Corp., Alphabet Inc.'s Google and OpenAI are expected to publicly commit Friday to safeguards for developing the technology—heeding a call from the White House. According to people familiar with the plans, the companies will pledge to responsible development and deployment of AI.
OpenAI cautioned that the ideas laid out in the internal policy document will be different from the ones that will soon be announced by the White House, alongside tech companies. Anna Makanju, the company's vice president of global affairs, said in an interview that the company isn't "pushing" for licenses as much as it believes such permitting is a "realistic" way for governments to track emerging systems.
"It's important for governments to be aware if super powerful systems that might have potential harmful impacts are coming into existence," she said, and there are "very few ways that you can ensure that governments are aware of these systems if someone is not willing to self-report the way we do." .
Makanju said OpenAI supports licensing regimes only for AI models more powerful than OpenAI's current GPT-4 one and wants to ensure smaller startups are free from too much regulatory burden. "We don't want to stifle the ecosystem," she said.
OpenAI also signaled in the internal policy document that it's willing to be more open about the data it uses to train image generators such as DALL-E, saying it was committed to "incorporating a provenance approach" by the end of the year. Data provenance—a practice used to hold developers accountable for transparency in their work and where it came from—has been raised by policy makers as critical to keeping AI tools from spreading misinformation and bias.
The commitments laid out in OpenAI's memo track closely with some of Microsoft's policy proposals announced in May. OpenAI has noted that, despite such similarities, the White House's initiative and OpenAI's proposals will be distinct from each other.
OpenAI said it's now looking to hear feedback from the public on how to cautiously regulate the industry. Makanju said the details of the proposed licensing requirements are still uncertain, but the company aims to establish them in a way that allows both responsible innovation and public accountability. "We've realized that we don't have all the answers," she added.
Share