Paul Krill
Editor at Large

OpenAI unveils specs for desired AI model behavior

news
May 9, 20242 mins

First draft of Model Spec documents how OpenAI wants its generative AI models to behave in ChatGPT and the OpenAI API.

openai
Credit: Andrew Neel

In a bid to “deepen the public conversation about how AI models should behave,” AI company OpenAI has introduced Model Spec, a document that shares the company’s approach to shaping desired model behavior.

Model Spec, now in a first draft, was introduced May 8. The document specifies OpenAI’s approach to shaping desired model behavior and how the company evaluates trade-offs when conflicts arise. The approach includes objectives, rules, and default behaviors that will guide OpenAI’s researchers and AI trainers who work on reinforcement learning from human feedback (RLHF). The company will also explore how much its models can learn directly from the Model Spec.

The Model Spec draws on documentation used at OpenAI today, the company’s experience and ongoing research in designing model behavior, and more recent work, including inputs from domain experts, OpenAI said. The company expects the Model Spec to change over time.

Objectives of the Model Spec include assisting the developer and user, benefiting humanity, and reflecting well on OpenAI. Rules include following the chain of command, complying with applicable laws, respecting creators, protecting privacy, not responding with not-safe-for-work content, and not providing information hazards. Default behaviors include encouraging fairness and kindness, using the right tool for the job, assuming best intentions from the user or developer, expressing uncertainty, and being as helpful as possible without overstepping.

OpenAI said it views its work on the Model Spec as part of an ongoing public conversation. The company seeks opportunities to engage with globally representative stakeholders, including policymakers, trusted institutions, and domain experts, to learn how they understand the approach, if they support it, and if there are additional objectives, rules, and defaults to be considered.

Paul Krill

Paul Krill is editor at large at InfoWorld. Paul has been covering computer technology as a news and feature reporter for more than 35 years, including 30 years at InfoWorld. He has specialized in coverage of software development tools and technologies since the 1990s, and he continues to lead InfoWorld’s news coverage of software development platforms including Java and .NET and programming languages including JavaScript, TypeScript, PHP, Python, Ruby, Rust, and Go. Long trusted as a reporter who prioritizes accuracy, integrity, and the best interests of readers, Paul is sought out by technology companies and industry organizations who want to reach InfoWorld’s audience of software developers and other information technology professionals. Paul has won a “Best Technology News Coverage” award from IDG.

More from this author