AI Security CI
Visit WebsiteGitHub RepoAI SecurityIdea / Pre-seed (early open-source project; minimal traction signals such as 1 star)Unknown (not indicated in repository metadata provided)
Description
Automated AI prompt security testing for CI. Detect jailbreaks, prompt leakage, and unsafe behavior before production.
Founders
Arpit Bhasin (inferred from GitHub owner: arpitbhasin1)
Discovered
December 16, 2025
Added to Database
January 26, 2026
Notes
Targets an emerging pain point: shifting LLM prompt-injection and jailbreak testing left into CI/CD via automated checks. Positioned as a developer-native security tool (GitHub Action) that could become a standard layer in AI app deployment pipelines.