OpenAI has announced the development of a new AI model named CriticGPT, designed to identify and rectify mistakes in code generated by GPT-4.
The AI firm stated that the new chatbot was trained using the reinforcement learning from human feedback (RLHF) framework and was powered by one of the GPT-4 models.
The under-development chatbot was designed to improve the quality of the AI-generated code that users get from the large language models. At present, the model is not available to users or testers. OpenAI also highlighted several limitations of the model.
OpenAI shared details of the new CriticGPT model in a blog post, stating that it was based on GPT-4 and designed to identify errors in code generated by ChatGPT.
“We found that when people get help from CriticGPT to review ChatGPT code they outperform those without help 60 percent of the time,” the company claims. The model was developed using the RLHF framework and the findings have been published in a paper.
RLHF is a machine learning technique that combines machine output with humans to train AI systems. In such a system, human evaluators provide feedback to the AI’s performance. This is used to adjust and improve the model’s behavior. Humans who provide feedback to the AI are called AI trainers.
According to OpenAI, CriticGPT demonstrated a 63 percent improvement over ChatGPT in catching code errors. However, the model has certain limitations. It has primarily been trained on short code snippets and has not yet been tested on longer, more complex coding tasks.
OpenAI suggests that CriticGPT is primarily intended to enhance the company’s understanding of training techniques to produce higher-quality AI outputs.
“While there is no immediate plan to release CriticGPT to the public, it may eventually be integrated into ChatGPT to provide improved code review capabilities,” as per the reports.