avatar
Amy Mullen @lilybelle05.bsky.social

To summarize, would even good faith attempts to make LLMs safer be futile? Or, would trying to create a safe and sane system be similar to laws against running red lights - imperfect compliance but better than nothing?

aug 29, 2025, 7:20 pm • 1 0

Replies

avatar
Natespeed @natespeed.bsky.social

When it comes to the generative side of things, probably closer to the latter. You can probably get them to the point where you can't casually bypass the guardrails - especially on smaller, more focused models. But a truly determined person will always be able to get around them.

aug 29, 2025, 10:41 pm • 1 0 • view
avatar
Amy Mullen @lilybelle05.bsky.social

Thanks for your posts, much appreciated.

aug 30, 2025, 12:37 am • 1 0 • view