🚀 Generate .cursorrules and AGENTS.md with automatic, format-based skill discovery across mixed .agent sources
Explore the docs »
View Demo
·
Report Bug
·
Request Feature
Table of Contents
This generator supports Dual-Mode Execution:
- Mode A (Enhanced): If you clone the full repo, the AI uses Python scripts (
indexer.py,wizard.py) for maximum efficiency. - Mode B (Autonomous): If you only use the workflow file, the AI falls back to its native file-reading capabilities to emulate the scripts—no Python required.
A structured 6-stage workflow (Stage 0–5) for creating professional project rules (.cursorrules) and AI agent guidelines (AGENTS.md). This workflow automatically discovers and integrates relevant AI skills from any compatible skill source — users never have to manually browse or select skills.
- ✅ Automatic Skill Selection — AI analyzes your project and finds relevant skills for you
- ✅ Format-Based Discovery — Works with CATALOG, FOLDER, SEARCH_ENGINE, README, and WORKFLOW sources
- ✅ Multi-Platform — Generates files for 9+ AI tools (Cursor, Claude, Gemini, Copilot, etc.)
- ✅ Custom Specification Support — Prioritizes high-fidelity architectural intent via Smart Intent Discovery (auto-detecting specs, requirement docs, and architectural guides)
- ✅ Quality Scoring — Built-in verification with a default 38/50 pass threshold and config-aware overrides
- ✅ Time-Saving — Complete in 30-60 minutes
- ✅ Verified on Mixed
.agent/Layouts — Current release is checked against the same combined skill setup documented below - ✅ Works with Specialized Skill Packs — Domain-specific sources such as
claude-scientific-skillsfit the same discovery pipeline without custom code paths - ✅ Never Outdated — Dynamic skill discovery, no hardcoded skill names
Release details are tracked in CHANGELOG.md.
| Feature | Description |
|---|---|
| 6-Stage Workflow | Structured process: Preferences → Analyze → Discover → .cursorrules → AGENTS.md → Verify |
| Format-Based Auto-Detect | Automatically detects skill sources by format (CATALOG/FOLDER/SEARCH_ENGINE/README/WORKFLOW) |
| Native MCP Support | 🌟 New in v1.9.0: Auto-discovery of MCP servers for seamless tool execution context |
| Memory & Audit Logging | 🌟 New in v1.9.2: Deeply integrated project-local audit trails (scripts/audit.py) and automated system prompt injection via memory summarization (scripts/memory_manager.py). |
| Multi-Platform Output | Generates correct files for Cursor, Claude, Antigravity, Gemini, Copilot, and more |
| 17 Keyword Categories | Comprehensive mapping from project types to skill search terms |
| Relevance-Ranked Search | Scores matches across catalogs, folders, READMEs, and search tools; supports --limit for large roots |
| Quality & Confidence | 🌟 New in v1.9.0: Heuristic verification (38/50 default pass) and confidence gating to ensure strict relevance |
| Interactive Wizard | CLI-based configuration generator (scripts/wizard.py) with integrated audit logging |
| Flexible Agentic Engine | 🌟 New in v1.9.2: 100% functional via native AI context (Zero-Install) with scripted Mode A fallback |
| Interactive Mode | Optional preferences system with config file (.rulesrc.yaml) support |
| Multi-Language | Output generation in 9 languages (en, th, ja, zh, ko, es, fr, de, pt) |
| Preview Mode | Review planned output structure before generation begins |
| Incremental Update | Diff-based update mode for existing files — preserves user customizations |
| Generation Statistics | Dashboard showing skills scanned, lines generated, quality scores |
| Shared Skill Roots | Supports project-local .agent/ plus configured external/shared skill directories |
| 10 Template Gallery | Pre-made examples for React, Express, Next.js, Go, Unity, CLI, LangChain, and more |
| Validation Scripts | Config-aware PowerShell/Bash validators with heuristic scorecards and threshold output |
- 🌐 Web Frontend — React, Vue, Angular, Svelte, Next.js
- ⚙️ Backend API — Node.js, Python, Go, Rust, Express, FastAPI
- 🧩 Browser Extensions — Chrome/Firefox Manifest V3
- 📱 Mobile Apps — React Native, Flutter, SwiftUI
- 💻 CLI Tools — Bash, PowerShell, Commander
- 🤖 AI/ML Projects — LLM, RAG, Agents, LangChain
- 🎮 Game Development — Unity, Godot, Unreal, Bevy
- 🔒 Security — Auth, OAuth, JWT, OWASP
- 📄 Documentation — API docs, Architecture docs
- 🧪 Testing — Jest, Pytest, Playwright, TDD
- ☁️ DevOps — Docker, Kubernetes, CI/CD, Terraform
- 🏗️ Architecture — Design patterns, Clean code, SOLID
- 🏢 Monorepo — Turborepo, Nx, Lerna, Workspaces
- 🔄 Microservices — Event-driven, CQRS, gRPC, Saga
- ☁️ Serverless — Lambda, Cloud Functions, Edge
- 🗄️ Database — PostgreSQL, MongoDB, Redis, ORM
- 📦 Package/Library — npm, PyPI, SDK publishing
| AI Tool | Icon | Output File(s) |
|---|---|---|
| Cursor | 🟠 | .cursorrules or .cursor/rules/*.mdc |
| Claude Code | 🟤 | CLAUDE.md + .claude/skills/ |
| Antigravity IDE | 🔴 | .agent/skills/*/SKILL.md + .agent/workflows/ |
| Gemini CLI | 🔵 | GEMINI.md |
| Codex CLI | 🟢 | AGENTS.md |
| Kiro IDE/CLI | 🟠 | AGENTS.md + .kiro/steering/ |
| GitHub Copilot | 🩵 | .github/copilot-instructions.md + .github/prompts/ |
| OpenCode | ⚪ | AGENTS.md |
| AdaL CLI | 🌸 | AGENTS.md |
- An AI assistant that supports workflows or skills (Cursor, Claude Code, Antigravity, etc.)
- At least one AI skill source installed in your project-local
.agent/directory or a configured shared skill root - Your project source code
Don't have skills yet? Clone one of these recommended collections into your
.agent/directory:
- antigravity-awesome-skills — large
CATALOG.md-driven index- awesome-claude-skills — community skill packs
- anthropic-skills — official Anthropic skills
- techleads-agent-skills — curated registry layout
- jeffallan-claude-skills — broad full-stack developer set
- ui-ux-pro-max-skill — UI/UX design intelligence
- othman-planning-with-files — planning and persistence workflows
- claude-scientific-skills — scientific and research-focused workflows
- andrej-karpathy-skills — Andrej Karpathy's skills and workflows
What Gets Installed?
When you use the setup scripts below, only the workflow file (.agent/workflows/create-project-rules.md) and the selected skill repositories are downloaded to your project. The Python helper scripts are not installed, as the AI will perform all analysis and discovery using its native capabilities.
Local project setup:
Linux/macOS:
curl -sL https://raw.githubusercontent.com/naravid19/ai-project-rules-generator/main/setup.sh | bashWindows (PowerShell):
irm https://raw.githubusercontent.com/naravid19/ai-project-rules-generator/main/setup.ps1 | iexShared skill root setup:
Linux/macOS:
curl -sL https://raw.githubusercontent.com/naravid19/ai-project-rules-generator/main/setup.sh -o setup.sh
bash setup.sh --skill-source all --skill-root /shared/.agentWindows (PowerShell):
iwr https://raw.githubusercontent.com/naravid19/ai-project-rules-generator/main/setup.ps1 -OutFile setup.ps1
.\setup.ps1 -SkillSource all -SkillRoot "C:\shared\team-skills\.agent"The workflow file still installs locally at
.agent/workflows/create-project-rules.md.SkillRootonly changes where optional skill repositories are cloned. Theworkflows/folder is reserved for installed workflow files and is intentionally ignored byscripts/discover-skills.pywhen enumerating external skill sources.
Single scientific-source setup (scientific installer key):
Linux/macOS:
bash setup.sh --skill-source scientific --skill-root /shared/.agentWindows (PowerShell):
.\setup.ps1 -SkillSource scientific -SkillRoot "C:\shared\team-skills\.agent"Use the
scientificsource when you want the shared root to focus on research-heavy work such as literature review, bioinformatics, cheminformatics, medical/clinical analysis, time-series science, or scientific writing workflows.
- Download the workflow file to your project:
mkdir -p .agent/workflows curl -o .agent/workflows/create-project-rules.md https://raw.githubusercontent.com/naravid19/ai-project-rules-generator/main/workflows/create-project-rules.md
For power users who want automated Python scripts for discovering skills, building catalogs, and validating outputs:
git clone https://github.com/naravid19/ai-project-rules-generator.git
cd ai-project-rules-generator(The AI will automatically use the Python scripts from scripts/ if it detects them in your project.)
If you used the Quick Start script (setup.sh or setup.ps1) to install the tool into your own project, the workflow is entirely AI-self-sufficient. No manual script execution is required.
Simply ask your AI assistant (e.g., Cursor, Claude Code, Antigravity IDE, Gemini CLI):
/create-project-rules
Or just say:
"Create professional project rules for this project"
The AI will read the workflow file and handle everything natively: discovering skills, prompting for preferences, generating rules, and verifying accuracy.
(Note: The interactive wizard script is only available if you cloned the full repository. If you used the Quick Start scripts in your own project, the AI will ask you questions directly or you can manually create a .rulesrc.yaml file. See Configuration below).
- If you cloned this repository, you can generate a
.rulesrc.yamlfile interactively:python scripts/wizard.py
- Open
.agent/workflows/create-project-rules.md - Follow the 6 stages step by step (Stage 0 is optional)
- Get
.cursorrulesandAGENTS.mdtailored to your project
Use .rulesrc.yaml when you want deterministic output instead of interactive prompts.
target_platforms:
- Cursor
- Codex
- Antigravity IDE
severity_level: balanced
output_language: en
template_style: progressive
quality_threshold: 38
preview_mode: false
existing_files: asktarget_platforms:
- codex
severity_level: strict
template_style: minimal
quality_threshold: 42
skill_sources:
- path: "C:/shared/team-skills/.agent"
- path: .agent
custom_keywords:
- planning
- memory
- workflow| Field | Behavior |
|---|---|
target_platforms |
Overrides AI-tool auto-detection when set |
severity_level |
Changes the density of critical vs advisory rules |
sections.include / sections.exclude |
Filters optional sections before generation |
skill_sources |
Discovery roots scanned in order; duplicate source names prefer the first root |
custom_keywords |
Merged with auto-detected keywords using case-insensitive dedupe |
template_style |
progressive, flat, or minimal formatting density |
quality_threshold |
Default validator threshold unless a CLI threshold overrides it |
preview_mode |
Enables the preview step before writing files |
existing_files |
Chooses ask / overwrite / merge / skip behavior |
- Relative
skill_sourcespaths resolve from the project root. - Absolute paths are allowed.
- Roots are searched in the order listed, so shared roots should appear before
.agentwhen they should win. - Companion docs next to a skill entry (
AGENTS.md,CLAUDE.md,README.md) are indexed with that skill instead of treated as separate matches. - Capability extraction prefers
SKILL.md, thenAGENTS.md, thenCLAUDE.md, and reports adjacentreferences/andrules/directories when present. - If a
formatvalue is supplied for a root, the workflow trusts it and skips format inference for that root. .agent/skillsremains the default Antigravity CATALOG directory for backward compatibility.
🌟 What's New in v1.9.2 (2026-04-26): Implemented Flexible Agentic Execution (Dual-Mode Strategy). The generator now automatically adapts to its environment: utilizing Python scripts if available (Mode A), or emulating them using native IDE tools if not (Mode B). Added strict Anti-Overload Rules for autonomous skill discovery to protect context window limits.
| Stage / Step | Time | Description |
|---|---|---|
| 0. User Preferences | 2-5 min | Config file or interactive: platforms, severity, language, preview |
| 1. Project Analysis | 10-15 min | Autonomous scan: structure, tech stack, patterns, AI tool detection |
| 2. Skill Discovery | 5-10 min | Auto-detect sources by format, search all by 17 keyword categories |
| 3. Create .cursorrules | 10-20 min | Coding standards with severity levels and progressive disclosure |
| Preview (optional) | 2-3 min | Review planned structure before writing files |
| 4. Create AGENTS.md | 10-15 min | AI guidelines with multi-platform output and dynamic skill section |
| 5. Verification | 5-10 min | Quality scoring (default 38/50), smell detection, generation statistics |
| Total Time | 30-60m |
The key innovation: instead of hardcoding specific skill repositories, the workflow resolves skill roots in this order and classifies each discovered source by format. The current release has been rechecked against the same mixed .agent/ layout installed by the setup scripts.
skill_sourcesfrom.rulesrc.yaml, scanned in the order listed- Project-local
.agent/when no explicit roots are configured - A user-provided shared root can be listed before
.agentto become the preferred baseline
.agent/skillsremains the default Antigravity CATALOG path for backward compatibility. Add shared roots through configuration or setup flags instead of renaming the legacy directory.
Examples:
- Local root:
.agent/ - Shared root:
C:/shared/team-skills/.agent - Reserved local output:
.agent/workflows/(workflow storage, not a skill source)
Resolve configured roots
│
├── Found CATALOG.md?
│ → Format: CATALOG (keyword search in table)
│
├── Found visible skill directories with SKILL.md / AGENTS.md / CLAUDE.md?
│ → Format: FOLDER (one skill entity per directory; companion docs stay attached)
│
├── Found search.py or search engine?
│ → Format: SEARCH_ENGINE (run search with keywords)
│
├── Found README.md with skill listing?
│ → Format: README (browse categorized list)
│
└── Found root CLAUDE.md / AGENTS.md plus hidden integrations such as .claude-plugin?
→ Format: WORKFLOW (read the workflow entrypoint and follow its instructions)
| Format | Detection Signal | Search Method |
|---|---|---|
| CATALOG | Contains CATALOG.md with skill table |
Search table rows by keywords |
| FOLDER | Contains visible skill directories with SKILL.md, AGENTS.md, or CLAUDE.md |
Build one skill entity per directory, then read the primary entry plus companion docs |
| SEARCH_ENGINE | Contains search.py or similar tool |
Run search tool with keywords |
| README | Contains README.md with categorized list |
Browse category headings |
| WORKFLOW | Root CLAUDE.md/AGENTS.md plus hidden integrations or .shared/, with no visible skill tree |
Read workflow, follow its instructions |
Clone these into your project-local .agent/ directory or a shared skill root referenced from .rulesrc.yaml:
| Source | Format | Notes | Link |
|---|---|---|---|
| antigravity-awesome-skills | CATALOG | Large catalog-style index | GitHub |
| awesome-claude-skills | FOLDER | Community skill packs | GitHub |
| anthropic-skills | FOLDER | Official skill collection | GitHub |
| techleads-agent-skills | FOLDER | Curated registry layout | GitHub |
| jeffallan-claude-skills | FOLDER | Broad full-stack set | GitHub |
| claude-scientific-skills | FOLDER | Scientific/research workflows | GitHub |
| ui-ux-pro-max-skill | WORKFLOW | Workflow-first package | GitHub |
| othman-planning-with-files | FOLDER | Planning/persistence set | GitHub |
You can use any skill source — as long as it matches one of the supported formats, the workflow will auto-detect and search it.
When to choose
claude-scientific-skills: Pick it for projects with scientific or research-heavy tasks such as literature review, genomics, chemistry, clinical analysis, forecasting, or scientific communication. Keep the broader general-purpose sources above for mixed engineering work.
| Check | Result | Notes |
|---|---|---|
| Native MCP Discovery | Supported | Auto-discovers and registers MCP intents via templates/mcp_registry.yaml for enhanced agent capabilities. |
| Traceability Metadata | Enforced | Validators strictly require Skill_Source_Path and Confirmed_Skill_Source: true tags. |
Repo-local .agent/ sample |
Supported | The committed sample root still resolves the mixed-format layout used in tests and documentation. |
| Shared-root baseline | Supported | The live shared root at C:/shared/team-skills/.agent was rechecked as the compatibility baseline for the current implementation. |
| Hybrid skill folders | Supported | Skill directories can now keep SKILL.md as the primary entry while surfacing companion AGENTS.md, CLAUDE.md, and README.md files. |
| Root precedence | Strict | Multi-root discovery strictly prefers the explicitly confirmed root for predictability. |
| Format scan benchmark | ~280 ms per root | Measured with python scripts/discover-skills.py (full repo clone only). |
| Keyword search benchmark | ~1.3-1.9 s | Measured with python scripts/discover-skills.py (full repo clone only). |
These commands are only available if you cloned the full repository. Quick Start users can skip this — the AI performs equivalent discovery natively.
python scripts/discover-skills.py --agent-dir "C:/shared/team-skills/.agent" --agent-dir .agent --format
python scripts/discover-skills.py --agent-dir "C:/shared/team-skills/.agent" --agent-dir .agent --keywords planning workflow python testing --limit 25
python scripts/extract-capabilities.py .agent/othman-planning-with-files/skills/planning-with-filesUse --limit whenever broad keywords can hit the large CATALOG source. The discovery script still searches all sources first, but the trimmed output is easier for agents to rank and summarize correctly.
python scripts/discover-skills.py --agent-dir "C:/shared/team-skills/.agent" --agent-dir .agent --format
python scripts/discover-skills.py --agent-dir "C:/shared/team-skills/.agent" --agent-dir .agent --keywords planning workflow python testing --limit 25
python scripts/extract-capabilities.py .agent/skills/skills/dbos-python
python -m unittest discover -s tests -v# Project Rules: TaskFlow API
> RESTful API for task management with FastAPI.
## Project Identity
- **Type**: Backend API
- **Language**: Python 3.11+
- **Framework**: FastAPI + SQLAlchemy + Alembic
## Critical Rules (🔴)
1. ❌ Never commit `.env` — use `app/core/config.py` with `pydantic-settings`
2. ✅ Always validate input with Pydantic schemas
## Code Smells
| ❌ Smell | ✅ Instead Do |
| ------------------------ | -------------------------------------- |
| `db = SessionLocal()` | `Depends(get_db)` injection |
| `import os; os.getenv()` | `from app.core.config import settings` |# AI Agent Guidelines — TaskFlow API
## 🎯 Available Skills
> [!IMPORTANT]
> Scan configured skill roots before starting work!
### How to Find Skills
| If You Find | Search Method |
| ------------------- | ---------------------------- |
| `CATALOG.md` | Search table by keywords |
| `SKILL.md` folders | Browse folder names |
| Workflow `.md` file | Read and follow instructions |
### Helpful Keywords
- Backend: `api`, `fastapi`, `rest`, `backend`
- Testing: `testing`, `pytest`, `unit`, `tdd`Pre-made .cursorrules examples for common project types are available in templates/:
| Template | Directory |
|---|---|
| React + TypeScript | templates/react-typescript/ |
| Python FastAPI | templates/python-fastapi/ |
| Flutter Mobile | templates/flutter-mobile/ |
| Node.js Express | templates/nodejs-express/ |
| Chrome Extension | templates/chrome-extension/ |
| Next.js Full-Stack | templates/nextjs-fullstack/ |
| Go Microservice | templates/go-microservice/ |
| Unity Game | templates/unity-game/ |
| CLI Tool | templates/cli-tool/ |
| LangChain RAG | templates/langchain-rag/ |
These are references only — the workflow generates custom versions tailored to your specific project. Use
templates/rulesrc-template.yamlto configure workflow behavior for your project.
Quick Start users: The AI performs equivalent validation natively using the inline checklist from Stage 5.5 of the workflow. You don't need these scripts.
Validation is heuristic scoring, not semantic proof. The scripts check structure, placeholders, formatting quality, repo-local path references, and cross-file consistency signals. They do not guarantee that generated rules are perfect for every downstream project.
powershell -ExecutionPolicy Bypass -File .\scripts\validate-output.ps1
powershell -ExecutionPolicy Bypass -File .\scripts\validate-output.ps1 -Threshold 42./scripts/validate-output.sh
./scripts/validate-output.sh --threshold 42The validators use this order:
- CLI threshold (
-Threshold/--threshold) .rulesrc.yaml→quality_threshold- Default
38/50
Each file is scored across five 0-10 dimensions:
- Completeness
- Accuracy
- Specificity
- Scannability
- Consistency
The scripts print both raw checks and final scorecards so you can see why a file passed or failed.
- Core workflow with 6 stages (Stage 0–5)
- Time estimates for each stage
- Example outputs
- Quick reference card
- Format-Based Auto-Detect skill discovery
- Multi-platform AI tool support (9 tools)
- Quality scoring verification (≥38/50)
- Content smell detection
- Severity levels (🔴/🟠/🟡)
- 17 keyword categories (expanded from 12)
- WORKFLOW format support
- Quick Start Scripts (
setup.sh/setup.ps1) - Template Gallery (10 templates)
- Python FastAPI example (replacing Chrome Extension)
- Interactive mode with config file (
.rulesrc.yaml) - Multi-language output support (9 languages)
- Preview mode (dry run before writing)
- Incremental update / diff mode
- Generation statistics dashboard
- Validation scripts (
scripts/validate-output.ps1/.sh) - Setup script improvements (
--non-interactive,--uninstall,--verbose) - Plugin system for custom skill formats
- Web UI for workflow configuration
See the open issues for a full list of proposed features (and known issues).
Contributions are what make the open source community such an amazing place to learn, inspire, and create. Any contributions you make are greatly appreciated.
If you have a suggestion that would make this better, please fork the repo and create a pull request. You can also simply open an issue with the tag "enhancement". Don't forget to give the project a star! Thanks again!
- Fork the Project
- Create your Feature Branch (
git checkout -b feature/AmazingFeature) - Commit your Changes (
git commit -m 'Add some AmazingFeature') - Push to the Branch (
git push origin feature/AmazingFeature) - Open a Pull Request
Distributed under the MIT License. See LICENSE for more information.
naravid19 - GitHub Profile
Project Link: https://github.com/naravid19/ai-project-rules-generator