Hacker Newsnew | past | comments | ask | show | jobs | submit | edf13's commentslogin

That is the biggest threat - and likely where things will end up eventually… it’s when that “eventually” is and what the server based providers can pivot to in that time.

Nice list!

As you say lots of effort going into this problem at the moment. We launch soon with grith.ai ~ a different take on the problem.


Building grith — OS-level syscall interception for AI coding agents.

The problem: every agent (Cline, Aider, Codex, Claude Code) has unrestricted access to your filesystem, shell, and network. When they process untrusted content — a cloned repo, a dependency README — they’re prompt injection vectors with full machine access. No existing tool evaluates what the agent actually does at the syscall level.

grith wraps any CLI agent without modification. OS-level interception captures every file open, network call, and process spawn, then runs it through 17 independent security filters in parallel across three phases (~15ms total). Composite score routes each call: auto-allow, auto-deny, or queue for async review. Most will auto approve - which eliminates approval fatigue.

Also does per-session cost tracking and audit trails as a side effect of intercepting everything.

https://grith.ai


Each syscall taking 15ms on top of the normal considered costly time taken for context switching to the kernel seems excessivly slow, no?

It’s fast in terms of a response from a LLM model - but it is part of the system I am quite active on at the moment to ensure it’s performant as possible

We are a different approach and are targeting Linux for our first release (Windows & Mac shortly afterwards).

Taking more of an automated supervisor approach with limited manual approval for edge cases.

Grith.ai


Yes - for many legacy systems especially in compliance related areas.


Could you elaborate? I am new to edi.


Agree, that’s why we’re building grith.ai

Sandboxing alone isn’t the right approach… a multi-faceted approach is what works.

What we’ve found that does work is automation on the approval process but only with very strong guards in place… approval fatigue is another growing problem - users simply clicking approve on all requests.


Interesting. How are the security filters implemented?


Every system call, file access, net access etc is forced through a local “proxy” where 17 individual filters check what’s going on.

Everything is done locally via our grith cli tool.

Happy to answer any questions on hello@grith.ai too


Was grift.ai too expensive?



That’s one of the reasons we’re building grith.ai ~ these ‘claw’ tools are getting too easy for use (which is good)… but they need securing!


Little too lexically close to girth


Haha - maybe… naming projects is hard!


It’s an interesting experiment… but I expect it to quickly die off as the same type message is posted again and again… their probably won’t be a great deal of difference in “personality” between each agent as they are all using the same base.


They're not though, you can use different models, and the bots have memories. That combined with their unique experiences might be enough to prevent that loop.


AI models have a tendency to like purple and similar shades.


I’d like more granular controls - sometimes I don’t want to trust the entire project but I do want to trust my elements of it


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: