Monday, January 2, 2023

ChatGPT: The Chatbot?

 ChatGPT: The Chatbot?

OpenAI's ChatGPT (Generative Pre-trained Transformer) chatbot was unveiled in November 2022. It is based on OpenAI's GPT-3.5 family of big language models and is modified using reinforcement and supervised learning approaches. ChatGPT has a million users within one week of its inception. While GPT-3 is the more capable of the two Large Language Models (LLMs), ChatGPT is best adapted for chatbot purposes and is quicker and more effective than GPT-3.

ChatGPT is developed via Reinforcement Learning from Human Feedback (RLHF) which is kind of similar to InstructGPT, but with minor variations in data gathering arrangement. Supervised fine-tuning was used to train the initial model: human AI trainers offered dialogues wherein they portrayed both sides—the user and an AI assistant. The instructors were provided with model-written ideas to assist them in composing their replies. This fresh dialogue dataset was combined with the InstructGPT dataset, which also was converted into a dialogue format.

To build a reinforcement learning reward model, comparison data, which entailed two or more model replies rated by quality was needed. This information was gathered by recording discussions between AI trainers and the chatbot. The developers chose a model-written phrase at random, sampled various potential completions, and had Artificial operators score them. 


Drawbacks

  • ChatGPT is sensitive to changes in input phrases or performing the same request many times.

  • ChatGPT occasionally writes realistic but erroneous or illogical responses.

  • The model is frequently very effusive and exaggerates specific terms, such as repeating that it is an OpenAI-trained language model. These problems occur as a result of prejudices in the training data and well-known excessive concerns. 

  • When a user submits an uncertain query, the model may ask questions for clarity.

  • The model will occasionally react negatively to damaging instructions or demonstrate prejudiced conduct.

The publication of ChatGPT research today is the latest development in OpenAI's continual implementation of efficient and secure useful AI systems. Several learnings from previous model deployments, such as GPT-3 and Codex, have impacted the security countermeasures in situ for this version, including substantial cuts in damaging and deceitful outputs gained through the application of reinforcement learning from human feedback (RLHF).

ChatGPT is an intriguing AI platform that can address our lingering questions, elucidate difficult concepts, and generate fresh ideas, to mention a few applications. The option to answer follow-up questions enables you to locate the information you need much quicker than if you had to visit multiple places on the web.

However, that doesn't indicate it's perfect. ChatGPT can still deliver dangerous answers and inaccurate information. During this study time, OpenAI intends to gather user feedback to assist enhance it for potential use.


No comments: