Recent allegations have surfaced against OpenAI, the brains behind ChatGPT, with ex-employees claiming the tech giant is venturing into risky territories with potentially harmful consequences.
To address concerns regarding AI safety, OpenAI has unveiled a new research paper outlining measures to enhance transparency in its models. The paper lays out a method to delve into the workings of the AI driving ChatGPT. This breakthrough enables the identification of potentially problematic concepts within the AI system.
While shedding light on OpenAI’s commitment to responsible AI development, the research also unveils internal strife within the company. The study was conducted by the disbanded “superalignment” team at OpenAI, focused on studying the long-term risks associated with AI technology.
Former leaders of the group, Ilya Sutskever and Jan Leike, named as coauthors, have since parted ways with OpenAI. Sutskever, a co-founder and erstwhile chief scientist, played a pivotal role in the company’s tumultuous episode last November, which culminated in the reinstatement of CEO Sam Altman.
ChatGPT, powered by GPT models that utilize artificial neural networks, represents a milestone in language models. Though these networks excel at learning from data, understanding their decision-making processes proves to be intricate due to the complexities of neural network interactions.
The researchers emphasize the challenge of comprehending neural networks, stating in a related blog post that the opacity of AI models raises concerns about potential misuse in weapon design or cyber warfare.
OpenAI’s recent paper introduces a method to uncover patterns representing specific concepts in machine learning systems. By enhancing the interpretability of these systems, the approach aims to boost efficiency and reveal the inner workings.
Demonstrating the effectiveness of this technique, OpenAI identified concept patterns within GPT-4, one of its leading AI models. The company also made the code for interpretability research accessible and introduced a visualization tool to examine word activations and concepts like profanity and sensitive content across different models. Understanding how models represent concepts could pave the way for mitigating unwanted behaviors and steering AI systems in desired directions.