Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

We need something like Perl's tainted strings to hinder sandbox escapes.


Wouldn’t help. The problem isn’t unsafe interpolation, the problem is unsafe interpretation. Models make decisions based on strings; that’s what they’re for. Problem is, once external data is “appended to the string” (updates the context), the model makes decisions based on the whole composite string, and existentially has no way to delineate trusted from untrusted data.




Consider applying for YC's Summer 2026 batch! Applications are open till May 4

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: