Show HN: API Testing and Security with AI

8 pointsposted 2 days ago
by siddhant_mohan

3 Comments

kshitijzeoauto

2 days ago

It claims to plug into your CI pipeline, detect what changed, and generate relevant test cases using LLMs.

As someone who’s struggled with stale or missing tests—especially in fast-moving codebases—I find this idea quite compelling. But I’m also curious about how it handles:

Contextual understanding across large codebases (e.g., multiple modules touched in a PR) Avoiding flaky or non-deterministic tests Matching team-specific coding styles or conventions

anuragdt

2 days ago

Generating tests is good, but how to handle the updating tests? Also how will you handle the flakiness and side effects of AI models?

siddhant_mohan

2 days ago

We handles flakiness with retries, smart waits, and isolation, while side effects are avoided using clean setups, teardowns, and state-safe mocks. Each tests scenarios are independent of each other and can be configured in a way to have prerequisite to setup the system and the post callback to cleanup the system

About updating test scenarios, we map it with your github commits and when a new commits come, we use the diff to figure out if tests failing are because of a bug or because of a new feature.