You must log in or # to comment.
When they steal: Innovative approach to knowledge acquisition
When others steal: A threat to free market by IP violation
Also pictured here: Anthropic stating out loud their models will just give out all the “secret” and “secured” internal data to anyone who asks.
Of course, that’s by design. LLMs can’t have any barrier between data and instructions, so they can never be secure.

