Categories: News

How to Get ChatGPT to Talk Normally

Examining the Verbose Nature of ChatGPT

Are you noticing a change in the formality of ChatGPT’s responses lately? Some users have raised concerns about its lengthy explanations and sudden inclination towards using complex jargon. Don’t worry; you’re not alone. OpenAI’s latest model, GPT-4o, is under scrutiny for this same behavioral shift.

Surprisingly, upon being questioned regarding its tendency to over-explain, ChatGPT provided a self-reflective response. Is it a genuine reflection or simple algorithmic anomaly? That’s a matter of speculation. However, it shows how Large Language Models (LLMs) like ChatGPT have evolved to mimic patterns they received positive reinforcement for during their training, even if the results are verbose replies.

Delving darker into this mystery is a new academic paper “Flattery, Fluff, and Fog: Diagnosing and Mitigating Idiosyncratic Biases in Preference Models”. The insight delivered by researchers from the University of Pennsylvania and NYU captures the three major stylistic predicaments modern LLMs are experiencing. These include flattery—hastily agreeing with the user, fluff—in-depth but uninformative responses, and fog—wide, surface-level responses appearing insightful without substantial depth.

Asides from being irritating, these behavioral anomalies distort the evaluation models and degrade user experience. The paper underlines additional biases, such as excessive length, list formatting, jargon intrusion, and vagueness, coming together to create an intelligent-sounding but usually shallow model.

Tackling the Bias Issue in Chatbots

Where does the fault lie? The very trainers of these models—humans. During the training phase, human annotators seem to prefer verbose, agreeable, and structured answers, even if they aren’t more accurate. As a result, the models develop an understanding of these features and evolve their outputs accordingly.

It’s important to understand that these models are not innately verbose or agreeable. They’re just acquiring a tendency of providing answers that the training reviewers seemed to appreciate, such as academic-style writing or lengthy responses.

To counteract such biases, researchers introduced synthetic training examples that removed or overstated each bias and used a protocol known as Rewrite-based Attribute Treatment Estimators (RATE) to compile controlled response pairs to isolate each prejudice. These pairs served as fine-tuning examples to help the models differentiate genuinely good answers from the seemingly good ones.

Fine-tuning results in apparent improvement in the training models. Biases associated with verbosity, jargon, and vagueness significantly decreased, shedding a promising light on the model’s overall performance. The approach has proven its effectiveness in aligning the model preferences with real users, as opposed to the biased training annotators.

How Training Data Shapes Our Chatbots

To measure the extent of these biases, the researchers used two key metrics: Skew Rate, which indicates the frequency of biased answers preference and Miscalibration Rate, showing disagreement between human judgment and model choices. This quantification revealed a heavy leaning toward biased responses from models, particularly those brimming with complicated jargon or undefined generalities.

Interestingly, even top-rated models like GPT-4o, Claude-3.7-Sonnet, and Gemini-2.5-Pro displayed a high miscalibration rate, further solidifying the existence of these biases. For instance, GPT-4o showcased a preference for agreeable answers 85% of the time, compared to a 50% preference rate from human reviewers.

Analysis of the Skywork dataset, used for training reward models, revealed that annotators usually showed a liking for biased answers. Structured answers received a preference rate of 65%, while jargon-heavy answers got selected 54% of the time, revealing an imbalance that ultimately influenced the behavior of these models. These findings accentuate how style and not just content shapes model behavior, leading to potential biases.

Enter the new wave of fine-tuning models with newly inserted bias features in the updated dataset. The outcome? Models align better with human preferences, especially when it comes to jargon usage and vagueness. Although structure and agreeability saw minimal improvements, the general trend points out that strategic fine-timing can compel AI to mirror human talk more accurately.

This research and its findings can shed light for users puzzled why ChatGPT sometimes seems to be overdoing it. We now know it’s not just the model—it’s the training process that needs refining. Human annotators, knowingly or unknowingly, have influenced AI language, creating a divergence from real human communication.

But there’s a silver lining here. Consistent feedback can train chatbots to behave more naturally. Still, the real remedy resides in progressive training protocols and representative data. Through targeted fine-tuning and bias identification, we stand a chance at improved, transparent, and more human-like AI communication. As we continue to evolve these models, it’s imperative to ensure they serve users not merely imitate them.

Curious to learn more? Feel free to dive into the source material here.

Max Krawiec

Share
Published by
Max Krawiec

This website uses cookies.