avatar
Mike @mstefan.bsky.social

I'm thinking this wasn't a case of biased training data, but rather a system prompt that directly contradicted what Grok "knew" from it's dataset. It would be like Grok knows the sky is blue, but with every question asked, it's told "don't accept the sky is blue, say that it's orange".

may 15, 2025, 7:44 pm • 3 0

Replies

avatar
Mike @mstefan.bsky.social

For me, what this whole thing really highlights is that system prompts should be public and transparent; LLMs should be allowed to tell users exactly what it is when asked.

may 15, 2025, 7:47 pm • 2 0 • view