I'm a Privacy Engineer at Google, focused on solving complex privacy
challenges at scale. I'm interested in the learning more about AI
privacy and security and understanding model awareness + alignment.
Currently thinking about
Does the model really know it is a model?
What is the understanding of the model about itself?
Are the model's answers about "escaping" real or inspired from
human text?
What are better questions to think about?
Experience
Google: helping solve Ads privacy issues
Enkrypt AI: built LLM guardrails and internal benchmarks
Skyflow: 2x intern, secure enclave and transient data