(NLI)
(AI, GPT)
Constraining GPT to our operational logic and existing dimensions of measurement turns out to be nothing more than a matter of asking. It’s effectively the same as tell it not to answer unless it’s certain – thus preventing hallucinations. ;). The LLM’s motivation is to be as helpful if possible even if ideation -> hallucination. The answer is to limit it’s helpfulness.
Source date (UTC): 2025-04-22 18:56:22 UTC
Original post: https://twitter.com/i/web/status/1914755175572783104
Leave a Reply