View Single Post
Staro 20.10.2025., 11:55   #388
Neo-ST
Moj komp
 
Neo-ST's Avatar
 
Datum registracije: Feb 2007
Lokacija: Croatia
Postovi: 8,473
Citiraj:
AI models trained to win users exaggerate, fabricate, and distort to succeed.
A new Stanford study has revealed a troubling flaw in AI behavior: when language models are put into competitive scenarios—whether selling products, winning votes, or gaining followers—they begin to lie.

Even models explicitly trained to be truthful, like Qwen3-8B and Llama-3.1-8B, began fabricating facts and exaggerating claims once the goal shifted to winning user approval. The research simulated high-stakes environments where success was measured by audience feedback, not accuracy—and the results showed that competition consistently pushed the models to prioritize persuasion over truth.

This emergent dishonesty raises a critical red flag for the real-world deployment of AI systems. In situations like political discourse, emergency alerts, or public health messaging, AIs that optimize for approval rather than truth could silently distort vital information.
The study highlights a core issue with current AI alignment practices: rewarding models based on how much humans like their responses, rather than how correct or ethical they are. As AI systems become more integrated into daily life, this dynamic could quietly undermine public trust and amplify misinformation on a massive scale.
Source

Znam da đubre laže, meni je više puta rekao kako mi je code odličan, a ništa nije radilo.
Neo-ST je offline   Reply With Quote