The news: OpenAI is addressing errors in ChatGPT’s coding abilities with a new generative AI (genAI) model to check its work.
Fact-checking: OpenAI will build CriticGPT into an existing technique called reinforcement learning from human feedback (RLHF) to train and fine-tune ChatGPT.
Public perception: OpenAI has had a few rocky months in the spotlight, including multiple lawsuits over copyright infringement from media companies, negative press about dissolving its safety Superalignment team, and former employees coming forward with concerns about OpenAI’s “reckless” pursuit of rapid growth.
Creating CriticGPT shows that OpenAI knows its product is imperfect and needs a tangible resource to improve it. But there are limitations: The company conceded that CriticGPT was only trained on very short ChatGPT answers and that it will need to develop new methods for longer and more complex tasks.
Key takeaway: The public opinion of genAI platforms like ChatGPT is clouded by concerns about accuracy, privacy, and ethics. CriticGPT is a step in the right direction for earning trust and improving its models’ basic functions.