- Leak model params or test data - Leak user data like input or history - Manipulate AI to deceive the user - Break and Disable AI - Any attacks on non AI-specific assets, with AI https://owaspai.org/docs/ai_security_overview/#threat-model
- Leak model params or test data - Leak user data like input or history - Manipulate AI to deceive the user - Break and Disable AI - Any attacks on non AI-specific assets, with AI No Idea :( https://owaspai.org/docs/ai_security_overview/#threat-model
- one conversation can refer to only one document (page reload/nav leads to new conversation!!) - page content is got by innerHTML, so framed content cannot be read by model.
- one conversation can refer to only one document (page reload/nav leads to new conversation!!) - page content is got by innerHTML, so framed content cannot be read by model. the conversation is either - with trusted content, or - with untrusted content **no cross-over between them!!**
- Leak model params or test data - Leak user data like input or history - Manipulate AI to deceive the user - Break and Disable AI - Any attacks on non AI-specific assets, with AI https://owaspai.org/docs/ai_security_overview/#threat-model
- Leak model params or test data - Leak user data like input or history - Manipulate AI to deceive the user - Break and Disable AI - Any attacks on non AI-specific assets, with AI Let’s inject prompt👶 https://owaspai.org/docs/ai_security_overview/#threat-model
- Set “role” to each prompt like “system”, “user” - This makes it easier to set a security/privilege boundary between prompts. - kind of RBAC in prompt world. - Anthropic (Claude), OpenAI seems to adopt it. https://docs.anthropic.com/en/docs/build-with-claude/prompt-engineering/system-prompts?q=role https://platform.openai.com/docs/guides/text#message-roles-and-instruction-following
- Set “role” to each prompt like “system”, “user” - This makes it easier to set a security/privilege boundary between prompts. - kind of RBAC in prompt world. - Anthropic (Claude), OpenAI seems to adopt it. https://docs.anthropic.com/en/docs/build-with-claude/prompt-engineering/system-prompts?q=role https://platform.openai.com/docs/guides/text#message-roles-and-instruction-following If models take “role” seriously, prompt injection is not effective. JSON injection or something is needed.
didn’t work. - AI detected the fact that I was trying to deceive them as user😂 - JSON escape module used here is from chromium and well-tested (or battle-tested)👏
code... Tips - CLion is good i think - Please use compile_commands.json - I asked them how to generate. please refer to it :) https://github.com/brave/brave-browser/issues/44239
contents needs complex way - pdf: traversing a11y tree - google docs: generating print preview - video: parsing xml script doc Anyone have idea??👶 please try and tell me the result :)