Compare commits

..

85 Commits

Author SHA1 Message Date
Tobin South
e1706ebd52 Merge pull request #669 from anthropics/add-plugin/intercom
add(plugin-json): intercom
2026-03-17 00:49:58 +00:00
Tobin South
6e4cf38fe2 Merge pull request #678 from anthropics/claude/merge-pr-669-d73N4
Resolve merge conflict in add-plugin/intercom vs staging
2026-03-17 00:49:43 +00:00
Tobin South
cc9555bb90 Merge latest staging to resolve conflict 2026-03-17 00:48:13 +00:00
Claude
79bed4d3b0 Merge remote-tracking branch 'origin/staging' into claude/merge-pr-669-d73N4
# Conflicts:
#	.claude-plugin/marketplace.json
2026-03-17 00:42:04 +00:00
Tobin South
fefdd738be Merge staging into add-plugin/intercom to resolve conflict 2026-03-17 00:40:45 +00:00
Tobin South
0c1407ea30 Merge pull request #676 from anthropics/tobin/qodo-patch-remove-sha 2026-03-17 00:34:05 +00:00
Tobin South
adeb0436c2 Merge pull request #670 from anthropics/add-plugin/neon 2026-03-17 00:31:07 +00:00
Tobin South
28ebfe4135 Merge pull request #664 from anthropics/claude/slack-replace-vercel-plugin-pointer-cadNe 2026-03-17 00:30:55 +00:00
Tobin South
3d0d05576d Remove qodo SHA 2026-03-16 23:05:02 +00:00
Tobin South
124fcfaa1e Add neon to marketplace 2026-03-16 13:49:40 -07:00
Tobin South
cccd8b3ea2 Add intercom to marketplace 2026-03-16 13:49:30 -07:00
Tobin South
478ea5b46a add(plugin): bonfire — session-context workflow tooling
Adapted from PR #108 by @vieko (Vercel).
Original: https://github.com/anthropics/claude-plugins-official/pull/108
2026-03-16 12:51:34 -07:00
Tobin South
fd805b5e4b add(plugin): microsoft-docs — official Microsoft documentation MCP
Adapted from PR #55 by @TianqiZhang (Microsoft).
Original: https://github.com/anthropics/claude-plugins-official/pull/55
2026-03-16 12:51:34 -07:00
Tobin South
fd8defbb34 add(plugin): atomic-agents — BrainBlend-AI framework
Adapted from PR #46 by @KennyVaneetvelde (BrainBlend-AI).
Original: https://github.com/anthropics/claude-plugins-official/pull/46
2026-03-16 12:51:34 -07:00
Tobin South
328a0a7190 add(plugin): stagehand — Browserbase browser automation
Adapted from PR #43 by @Kylejeong2 (Browserbase). PR's marketplace.json
had a syntax error (missing '},' before adjacent entry); entry
reconstructed from the diff.
Original: https://github.com/anthropics/claude-plugins-official/pull/43
2026-03-16 12:51:33 -07:00
Tobin South
3f3d3daeb8 add(plugin): autofix-bot — DeepSource automated code review
Adapted from PR #23 by @jai-deepsource (DeepSource).
Original: https://github.com/anthropics/claude-plugins-official/pull/23
2026-03-16 12:51:33 -07:00
Tobin South
f59c36423d add(plugin): terraform — HashiCorp infrastructure-as-code
Adapted from PR #14 by @gautambaghel (HashiCorp).
Original: https://github.com/anthropics/claude-plugins-official/pull/14
2026-03-16 12:51:33 -07:00
Tobin South
e97b983948 docs(ralph-loop): add Windows compatibility section
Retargeted from PR #124 (originally against plugins/ralph-wiggum/,
since renamed). Documents the Git Bash workaround for Windows users
hitting WSL bash resolution issues in the stop hook.

Original author: @stefanzvonar
2026-03-16 12:49:52 -07:00
Tobin South
db1e313270 Merge PR #77: Update Context7 plugin with Skills, Agents and v2 2026-03-16 12:49:31 -07:00
Tobin South
c91a334747 Merge PR #119: Add README and setup documentation for Greptile plugin 2026-03-16 12:49:30 -07:00
Tobin South
4f0a09875b Merge PR #497: Fix empty array crash on bash 3.2 in setup-ralph-loop.sh 2026-03-16 12:49:30 -07:00
Tobin South
f3f13c4499 Merge PR #116: fix(plugin-dev): add missing .claude-plugin/plugin.json 2026-03-16 12:49:30 -07:00
Tobin South
a5bd1097e8 Merge PR #311: Change Notion name to lowercase in marketplace.json 2026-03-16 12:49:30 -07:00
Tobin South
8a25030d01 Merge PR #102: Fix README typo 2026-03-16 12:49:29 -07:00
Tobin South
1086e0cc1a vercel-labs to vercel 2026-03-16 17:58:57 +00:00
Claude
c554ce45e3 Update Vercel plugin to point to vercel-labs/vercel-plugin
Replace the marketplace pointer for the Vercel plugin from
vercel/vercel-deploy-claude-code-plugin to vercel-labs/vercel-plugin.
2026-03-16 17:26:15 +00:00
Tobin South
d5c15b861c Merge pull request #628 from anthropics/staging
Plugin batch: 17 new plugins + posthog update
2026-03-13 21:51:01 +00:00
Tobin South
bf6270071e Merge pull request #613 from anthropics/add-license-ruby-lsp
Add Apache 2.0 LICENSE to ruby-lsp plugin
2026-03-13 21:30:00 +00:00
Tobin South
50ebf6df3b Merge pull request #626 from anthropics/add-ruby-lsp-license
Add Apache 2.0 LICENSE to ruby-lsp plugin
2026-03-13 21:29:39 +00:00
Tobin South
ad61a54061 Merge pull request #517 from anthropics/claude/slack-update-webhook-link-wikTM
Update webhook closed PR message link to clau.de
2026-03-13 21:26:34 +00:00
Tobin South
95327347c3 Merge PR #583: add-plugin/zapier 2026-03-13 21:17:45 +00:00
Tobin South
64ce1721a4 Merge PR #579: add-plugin/deploy-on-aws 2026-03-13 21:17:45 +00:00
Tobin South
9febf87cdb Merge PR #578: add-plugin/migration-to-aws 2026-03-13 21:17:45 +00:00
Tobin South
fbe0386df6 Merge PR #577: add-plugin/aws-serverless 2026-03-13 21:17:44 +00:00
Tobin South
1f25b55ae3 Merge PR #576: add-plugin/amazon-location-service 2026-03-13 21:17:44 +00:00
Tobin South
616512c59d Merge PR #536: add-plugin/wix 2026-03-13 21:17:43 +00:00
Tobin South
80d85e8f9c Merge PR #535: add-plugin/sumup 2026-03-13 21:17:43 +00:00
Tobin South
b4178e809b Merge PR #534: add-plugin/mintlify 2026-03-13 21:17:43 +00:00
Tobin South
57fe2068ec Merge PR #533: add-plugin/legalzoom 2026-03-13 21:17:42 +00:00
Tobin South
a8e8f7e89f Merge PR #532: add-plugin/data 2026-03-13 21:17:42 +00:00
Tobin South
e0b2429899 Merge PR #531: add-plugin/sanity-plugin 2026-03-13 21:17:41 +00:00
Tobin South
159db463ec Merge PR #530: add-plugin/sourcegraph 2026-03-13 21:17:41 +00:00
Tobin South
1d1f304807 Merge PR #529: add-plugin/railway 2026-03-13 21:17:40 +00:00
Tobin South
5a5fc148df Merge PR #528: add-plugin/adspirer-ads-agent 2026-03-13 21:17:40 +00:00
Tobin South
92e9c49f3e Merge PR #527: add-plugin/rc 2026-03-13 21:17:40 +00:00
Tobin South
9750826583 Merge PR #526: add-plugin/planetscale 2026-03-13 21:17:39 +00:00
Tobin South
d726c5ea42 Merge PR #525: add-plugin/chrome-devtools-mcp 2026-03-13 21:17:39 +00:00
Tobin South
61ff000c60 Merge PR #542: update(plugin-json): posthog — point at consolidated repo 2026-03-13 21:17:10 +00:00
Claude
c96abc73df Add Apache 2.0 LICENSE to ruby-lsp plugin 2026-03-13 17:06:01 +00:00
Claude
aeb25ced03 Add Apache 2.0 LICENSE to ruby-lsp plugin 2026-03-12 17:05:58 +00:00
Tobin South
b36fd4b753 Add pagerduty to marketplace (#566) 2026-03-11 16:37:54 -07:00
Noah Zweben
bd041495bd update(plugin-json): point to the correct Semgrep plugin directory (#584)
The Semgrep plugin currently does not work correctly when used through
Claude because it is located within a subdirectory of the Semgrep
Marketplace repository. This issue was reported in:
https://github.com/anthropics/claude-plugins-official/issues/450

Previously, this could not be fixed due to a limitation in Claude Code's
handling of plugins located in subdirectories. Support for this was added
with the git-subdir feature, released in v2.1.69:
https://github.com/anthropics/claude-code/issues/30593

A fix for the Semgrep plugin was proposed once this version became the
latest release. Now that v2.1.69+ is available as latest, this PR
implements that fix.

https://claude.ai/code/cse_01RtW9KS12VZNFfWmWY6z9Pu

Co-authored-by: Claude <noreply@anthropic.com>
2026-03-10 15:59:33 -07:00
Daisy S. Hollman
00f13a5f46 Merge pull request #106 from obahareth/main
Add Ruby LSP plugin with inline lspServers configuration
2026-03-10 13:21:32 -07:00
Tobin South
7e94c732f6 Merge pull request #540 from anthropics/add-plugin/postman
add(plugin-json): postman
2026-03-10 17:12:50 +00:00
Tobin South
4fecb30b3c Add zapier to marketplace 2026-03-10 15:57:09 +00:00
Tobin South
b90a056130 Add deploy-on-aws to marketplace 2026-03-10 06:19:37 +00:00
Tobin South
038e989ee4 Add migration-to-aws to marketplace 2026-03-10 06:19:06 +00:00
Tobin South
8fda75ce49 Add aws-serverless to marketplace 2026-03-10 06:19:04 +00:00
Tobin South
f71d2d9925 Add amazon-location-service to marketplace 2026-03-10 06:18:59 +00:00
Tobin South
7c626d26bb Point posthog at consolidated PostHog/ai-plugin repo
Updates source URL from posthog-for-claude to ai-plugin per PostHog's
repo consolidation. Adds SHA pin and fixes homepage to proper docs URL.
2026-03-06 16:08:55 -08:00
Tobin South
4fa27586e5 Add postman to marketplace 2026-03-06 10:48:54 -08:00
Tobin South
b3b3549c12 Add sumup to marketplace 2026-03-05 19:55:08 -08:00
Tobin South
7b67d48001 Add mintlify to marketplace 2026-03-05 19:54:59 -08:00
Tobin South
7b7e85568b Add legalzoom to marketplace 2026-03-05 19:54:24 -08:00
Tobin South
8a89ca31e1 Add data to marketplace 2026-03-05 19:51:07 -08:00
Tobin South
9c11aed2b7 Add sanity-plugin to marketplace 2026-03-05 19:50:30 -08:00
Tobin South
2a6b21dabc Add sourcegraph to marketplace 2026-03-05 19:50:00 -08:00
Tobin South
41ac3012b6 Add railway to marketplace 2026-03-05 19:48:26 -08:00
Tobin South
934cc3b4e9 Add adspirer-ads-agent to marketplace 2026-03-05 19:44:27 -08:00
Tobin South
7657ed1025 Add rc to marketplace 2026-03-05 19:43:40 -08:00
Tobin South
954edbd88b Add planetscale to marketplace 2026-03-05 19:35:15 -08:00
Tobin South
8249477529 Add chrome-devtools-mcp to marketplace 2026-03-05 18:27:07 -08:00
Claude
fe41d329d5 Update webhook closed PR message link to clau.de
Replace the Google Forms link with the new plugin directory
submission link (https://clau.de/plugin-directory-submission).

https://claude.ai/code/session_01NxRDJKDvFR2d4wC4ppDEDT
2026-03-04 00:51:16 +00:00
LuciferDono
acd3701274 Fix empty array crash on bash 3.2 in setup-ralph-loop.sh 2026-03-04 00:26:10 +05:30
Matt Kotsenas
cd89e41cf4 Change Notion name to lowercase in marketplace.json
According to the SKILLS spec (see https://agentskills.io/specification#:~:text=Max%2064%20characters.%20Lowercase%20letters%2C%20numbers%2C%20and%20hyphens%20only.%20Must%20not%20start%20or%20end%20with%20a%20hyphen.) names should not contain uppercase letters. This prevents loading the marketplace in spec-compliant agents.

Update the name to be in lowercase.
2026-01-30 16:47:12 -08:00
Fahreddin Özcan
42d7afb1f0 fix: installation guide 2026-01-16 11:10:54 +03:00
Fahreddin Özcan
085871e8e7 fmt 2026-01-16 11:10:54 +03:00
Fahreddin Özcan
32f2cdbe0c feat: update tools with better skill/agent format prompt 2026-01-16 11:10:53 +03:00
Fahreddin Özcan
24cec23cf1 Switch Context7 MCP to remote HTTP server 2026-01-16 11:10:53 +03:00
Fahreddin Özcan
c7ba9d4c43 Update Context7 plugin for v2 API
- Update skill/agent/command to use new query-docs tool (replaces get-library-docs)
- Add query parameter usage for intelligent reranking
- Add version pinning support (e.g., /vercel/next.js/v15.1.8)
- Add tools and model metadata to agent
- Simplify docs to focus on workflow, not parameter details
- Add README.md with usage examples
2026-01-16 11:10:53 +03:00
Fahreddin Özcan
72fa7b63ed feat: add c7 agent 2026-01-16 11:10:53 +03:00
Daksh Gupta
a5604c1355 Add README and setup documentation for Greptile plugin
- Add README.md with setup instructions for getting API key
- Document the GREPTILE_API_KEY environment variable requirement
- Add homepage, author URL, and keywords to plugin.json
- Update description to reflect Greptile as AI code review agent

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
2026-01-05 12:43:54 -08:00
Julien Tavernier
8e7c0615e6 fix(plugin-dev): add missing .claude-plugin/plugin.json
The plugin-dev plugin was missing its required plugin.json manifest file,
causing the plugin to fail loading. This adds the missing configuration
file following the same format as other official plugins.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
2026-01-05 17:10:46 +04:00
Omar Bahareth
80a2049c5d Add Ruby LSP plugin with inline lspServers configuration 2026-01-04 15:27:02 +03:00
Han T.
aab3f1ba3f fix readme typo 2026-01-02 11:56:46 -08:00
36 changed files with 2643 additions and 26 deletions

View File

@@ -21,7 +21,9 @@
"lspServers": {
"typescript": {
"command": "typescript-language-server",
"args": ["--stdio"],
"args": [
"--stdio"
],
"extensionToLanguage": {
".ts": "typescript",
".tsx": "typescriptreact",
@@ -49,7 +51,9 @@
"lspServers": {
"pyright": {
"command": "pyright-langserver",
"args": ["--stdio"],
"args": [
"--stdio"
],
"extensionToLanguage": {
".py": "python",
".pyi": "python"
@@ -111,7 +115,9 @@
"lspServers": {
"clangd": {
"command": "clangd",
"args": ["--background-index"],
"args": [
"--background-index"
],
"extensionToLanguage": {
".c": "c",
".h": "c",
@@ -140,7 +146,9 @@
"lspServers": {
"intelephense": {
"command": "intelephense",
"args": ["--stdio"],
"args": [
"--stdio"
],
"extensionToLanguage": {
".php": "php"
}
@@ -181,12 +189,14 @@
"lspServers": {
"kotlin-lsp": {
"command": "kotlin-lsp",
"args": ["--stdio"],
"args": [
"--stdio"
],
"extensionToLanguage": {
".kt": "kotlin",
".kts": "kotlin"
},
"startupTimeout" : 120000
"startupTimeout": 120000
}
}
},
@@ -251,6 +261,30 @@
}
}
},
{
"name": "ruby-lsp",
"description": "Ruby language server for code intelligence and analysis",
"version": "1.0.0",
"author": {
"name": "Anthropic",
"email": "support@anthropic.com"
},
"source": "./plugins/ruby-lsp",
"category": "development",
"strict": false,
"lspServers": {
"ruby-lsp": {
"command": "ruby-lsp",
"extensionToLanguage": {
".rb": "ruby",
".rake": "ruby",
".gemspec": "ruby",
".ru": "ruby",
".erb": "erb"
}
}
}
},
{
"name": "agent-sdk-dev",
"description": "Development kit for working with the Claude Agent SDK",
@@ -363,7 +397,7 @@
},
{
"name": "playground",
"description": "Creates interactive HTML playgrounds self-contained single-file explorers with visual controls, live preview, and prompt output with copy button. Includes templates for design playgrounds, data explorers, concept maps, and document critique.",
"description": "Creates interactive HTML playgrounds \u2014 self-contained single-file explorers with visual controls, live preview, and prompt output with copy button. Includes templates for design playgrounds, data explorers, concept maps, and document critique.",
"author": {
"name": "Anthropic",
"email": "support@anthropic.com"
@@ -451,7 +485,9 @@
"category": "development",
"source": "./external_plugins/serena",
"homepage": "https://github.com/anthropics/claude-plugins-public/tree/main/external_plugins/serena",
"tags": ["community-managed"]
"tags": [
"community-managed"
]
},
{
"name": "playwright",
@@ -516,7 +552,7 @@
"homepage": "https://github.com/anthropics/claude-plugins-public/tree/main/external_plugins/linear"
},
{
"name": "Notion",
"name": "notion",
"description": "Notion workspace integration. Search pages, create and update documents, manage databases, and access your team's knowledge base directly from Claude Code for seamless documentation workflows.",
"category": "productivity",
"source": {
@@ -558,11 +594,11 @@
"category": "deployment",
"source": {
"source": "url",
"url": "https://github.com/vercel/vercel-deploy-claude-code-plugin.git"
"url": "https://github.com/vercel/vercel-plugin.git"
},
"homepage": "https://github.com/vercel/vercel-deploy-claude-code-plugin"
"homepage": "https://github.com/vercel/vercel-plugin"
},
{
{
"name": "stripe",
"description": "Stripe development plugin for Claude",
"category": "development",
@@ -582,7 +618,9 @@
"category": "development",
"source": "./external_plugins/context7",
"homepage": "https://github.com/anthropics/claude-plugins-public/tree/main/external_plugins/context7",
"tags": ["community-managed"]
"tags": [
"community-managed"
]
},
{
"name": "pinecone",
@@ -626,17 +664,18 @@
},
{
"name": "posthog",
"description": "Connect Claude Code to your PostHog analytics platform. Query insights, manage feature flags, run A/B experiments, track errors, and analyze LLM costs all through natural language. The plugin provides 10 slash commands for common workflows and full access to PostHog's MCP tools. Ask questions like \"What are my top errors?\" or \"Create a feature flag for 50% of users\" and Claude handles the API calls. Supports OAuth authentication, EU and US cloud regions, and self-hosted instances.",
"description": "Access PostHog analytics, feature flags, experiments, error tracking, and insights directly from Claude Code.",
"category": "monitoring",
"source": {
"source": "url",
"url": "https://github.com/PostHog/posthog-for-claude.git"
"url": "https://github.com/PostHog/ai-plugin.git",
"sha": "f2f37954ecef9f1afce4fa81b6a612454a96c410"
},
"homepage": "https://github.com/PostHog/posthog-for-claude.git"
"homepage": "https://posthog.com/docs/model-context-protocol"
},
{
"name": "coderabbit",
"description": "Your code review partner. CodeRabbit provides external validation using a specialized AI architecture and 40+ integrated static analyzersoffering a different perspective that catches bugs, security vulnerabilities, logic errors, and edge cases. Context-aware analysis via AST parsing and codegraph relationships. Automatically incorporates CLAUDE.md and project coding guidelines into reviews. Useful after writing or modifying code, before commits, when implementing complex or security-sensitive logic, or when a second opinion would increase confidence in the changes. Returns specific findings with suggested fixes that can be applied immediately. Free to use.",
"description": "Your code review partner. CodeRabbit provides external validation using a specialized AI architecture and 40+ integrated static analyzers\u2014offering a different perspective that catches bugs, security vulnerabilities, logic errors, and edge cases. Context-aware analysis via AST parsing and codegraph relationships. Automatically incorporates CLAUDE.md and project coding guidelines into reviews. Useful after writing or modifying code, before commits, when implementing complex or security-sensitive logic, or when a second opinion would increase confidence in the changes. Returns specific findings with suggested fixes that can be applied immediately. Free to use.",
"category": "productivity",
"source": {
"source": "url",
@@ -666,12 +705,11 @@
},
{
"name": "qodo-skills",
"description": "Qodo Skills provides a curated library of reusable AI agent capabilities that extend Claude's functionality for software development workflows. Each skill is designed to integrate seamlessly into your development process, enabling tasks like code quality checks, automated testing, security scanning, and compliance validation. Skills operate across your entire SDLCfrom IDE to CI/CDensuring consistent standards and catching issues early.",
"description": "Qodo Skills provides a curated library of reusable AI agent capabilities that extend Claude's functionality for software development workflows. Each skill is designed to integrate seamlessly into your development process, enabling tasks like code quality checks, automated testing, security scanning, and compliance validation. Skills operate across your entire SDLC\u2014from IDE to CI/CD\u2014ensuring consistent standards and catching issues early.",
"category": "development",
"source": {
"source": "url",
"url": "https://github.com/qodo-ai/qodo-skills.git",
"sha": "623eb4ed4364d8111f9a9132a791d7497d814b6a"
"url": "https://github.com/qodo-ai/qodo-skills.git"
},
"homepage": "https://github.com/qodo-ai/qodo-skills.git"
},
@@ -680,11 +718,162 @@
"description": "Semgrep catches security vulnerabilities in real-time and guides Claude to write secure code from the start.",
"category": "security",
"source": {
"source": "url",
"url": "https://github.com/semgrep/mcp-marketplace.git"
"source": "git-subdir",
"url": "https://github.com/semgrep/mcp-marketplace.git",
"path": "plugin"
},
"homepage": "https://github.com/semgrep/mcp-marketplace.git"
},
{
"name": "pagerduty",
"description": "Enhance code quality and security through PagerDuty risk scoring and incident correlation. Score pre-commit diffs against historical incident data and surface deployment risk before you ship.",
"category": "monitoring",
"source": {
"source": "url",
"url": "https://github.com/PagerDuty/claude-code-plugins.git",
"sha": "b16c23e0d790deceaa7a6182616d0e36673f2eae"
},
"homepage": "https://github.com/PagerDuty/claude-code-plugins"
},
{
"name": "postman",
"description": "Full API lifecycle management for Claude Code. Sync collections, generate client code, discover APIs, run tests, create mocks, publish docs, and audit security. Powered by the Postman MCP Server.",
"category": "development",
"source": {
"source": "url",
"url": "https://github.com/Postman-Devrel/postman-claude-code-plugin.git",
"sha": "0714280351c1a137e79aad465a66730511ffbd57"
},
"homepage": "https://learning.postman.com/docs/developer/postman-mcp-server/"
},
{
"name": "chrome-devtools-mcp",
"description": "Control and inspect a live Chrome browser from your coding agent. Record performance traces, analyze network requests, check console messages with source-mapped stack traces, and automate browser actions with Puppeteer.",
"category": "development",
"source": {
"source": "url",
"url": "https://github.com/ChromeDevTools/chrome-devtools-mcp.git",
"sha": "c2d8009ff75f76bce1ec4cf79c2467b50d81725e"
},
"homepage": "https://github.com/ChromeDevTools/chrome-devtools-mcp"
},
{
"name": "planetscale",
"description": "An authenticated hosted MCP server that accesses your PlanetScale organizations, databases, branches, schema, and Insights data. Query against your data, surface slow queries, and get organizational and account information.",
"category": "database",
"source": {
"source": "url",
"url": "https://github.com/planetscale/claude-plugin.git",
"sha": "f1066cac5bb956bbbb05918f5b07fe0e873d44ea"
},
"homepage": "https://planetscale.com/"
},
{
"name": "rc",
"description": "Configure RevenueCat projects, apps, products, entitlements, and offerings directly from Claude Code. Manage your in-app purchase backend without leaving your development workflow.",
"category": "development",
"source": {
"source": "url",
"url": "https://github.com/RevenueCat/rc-claude-code-plugin.git",
"sha": "af7cb77996aee4e7e3c109c5afec81f716139032"
},
"homepage": "https://www.revenuecat.com"
},
{
"name": "adspirer-ads-agent",
"description": "Cross-platform ad management for Google Ads, Meta Ads, TikTok Ads, and LinkedIn Ads. 91 tools for keyword research, campaign creation, performance analysis, and budget optimization.",
"category": "productivity",
"source": {
"source": "url",
"url": "https://github.com/amekala/adspirer-mcp-plugin.git",
"sha": "aa70dbdbbbb843e94a794c10c2b13f5dd66b5e40"
},
"homepage": "https://www.adspirer.com"
},
{
"name": "railway",
"description": "Deploy and manage apps, databases, and infrastructure on Railway. Covers project setup, deploys, environment configuration, networking, troubleshooting, and monitoring.",
"category": "deployment",
"source": {
"source": "git-subdir",
"url": "railwayapp/railway-skills",
"path": "plugins/railway",
"ref": "main",
"sha": "d52f3741a6a33a3191d6138eb3d6c3355cb970d1"
},
"homepage": "https://docs.railway.com/ai/claude-code-plugin"
},
{
"name": "sourcegraph",
"description": "Code search and understanding across codebases. Search, read, and trace references across repositories; analyze refactor impact; investigate incidents via commit and diff search; run targeted security sweeps.",
"category": "development",
"source": {
"source": "url",
"url": "https://github.com/sourcegraph-community/sourcegraph-claudecode-plugin.git",
"sha": "cfe3d44476957b16d1575261bef6b2dc7cb1e0b7"
},
"homepage": "https://sourcegraph.com"
},
{
"name": "sanity-plugin",
"description": "Sanity content platform integration with MCP server, agent skills, and slash commands. Query and author content, build and optimize GROQ queries, design schemas, and set up Visual Editing.",
"category": "development",
"author": {
"name": "Sanity"
},
"source": {
"source": "url",
"url": "https://github.com/sanity-io/agent-toolkit.git",
"sha": "4b1fb10bd707a22cf0cdfad5374ffc885f2ffa8d"
},
"homepage": "https://www.sanity.io"
},
{
"name": "data",
"description": "Data engineering for Apache Airflow and Astronomer. Author DAGs with best practices, debug pipeline failures, trace data lineage, profile tables, migrate Airflow 2 to 3, and manage local and cloud deployments.",
"category": "development",
"source": {
"source": "url",
"url": "https://github.com/astronomer/agents.git",
"sha": "7ef022b02f5296b5ecc52ba0db3ba9345ec03c9e"
},
"homepage": "https://github.com/astronomer/agents"
},
{
"name": "legalzoom",
"description": "Attorney guidance and legal tools for business and personal needs. AI-powered document review identifies critical risks and important clauses, advises when to engage an attorney, and routes to LegalZoom's network when professional expertise is needed.",
"category": "productivity",
"source": {
"source": "git-subdir",
"url": "legalzoom/claude-plugins",
"path": "plugins/legalzoom",
"ref": "main",
"sha": "f9fd8a0ca6e1421bc1aacb113a109663a7a6f6d8"
},
"homepage": "https://www.legalzoom.com/"
},
{
"name": "mintlify",
"description": "Build beautiful documentation sites with Mintlify. Convert non-markdown files into properly formatted MDX pages, add and modify content with correct component use, and automate documentation updates.",
"category": "development",
"source": {
"source": "url",
"url": "https://github.com/mintlify/mintlify-claude-plugin.git",
"sha": "ce435be18a700dc849d6a63a80da4816d1e2128c"
},
"homepage": "https://www.mintlify.com/"
},
{
"name": "sumup",
"description": "SumUp payment integrations across terminal and online checkout flows. Build Android and iOS POS apps with SumUp card readers, online checkout with server SDKs and the checkout widget, and control card readers remotely via Cloud API.",
"category": "development",
"source": {
"source": "url",
"url": "https://github.com/sumup/sumup-skills.git",
"sha": "802476c39a0422d3277e37288b03968ad731bc30"
},
"homepage": "https://www.sumup.com/"
},
{
"name": "wix",
"description": "Build, manage, and deploy Wix sites and apps. CLI development skills for dashboard extensions, backend APIs, site widgets, and service plugins with the Wix Design System, plus MCP server for site management.",
@@ -695,6 +884,160 @@
"sha": "15dda227e34959b1340e33bb9aede7e23a273f42"
},
"homepage": "https://dev.wix.com/docs/wix-cli/guides/development/about-wix-skills"
},
{
"name": "amazon-location-service",
"description": "Guide developers through adding maps, places search, geocoding, routing, and other geospatial features with Amazon Location Service, including authentication setup, SDK integration, and best practices.",
"category": "location",
"source": {
"source": "git-subdir",
"url": "https://github.com/awslabs/agent-plugins.git",
"path": "plugins/amazon-location-service",
"ref": "main"
},
"homepage": "https://github.com/awslabs/agent-plugins"
},
{
"name": "aws-serverless",
"description": "Design, build, deploy, test, and debug serverless applications with AWS Serverless services.",
"category": "development",
"source": {
"source": "git-subdir",
"url": "https://github.com/awslabs/agent-plugins.git",
"path": "plugins/aws-serverless",
"ref": "main"
},
"homepage": "https://github.com/awslabs/agent-plugins"
},
{
"name": "migration-to-aws",
"description": "Assess current cloud provider usage and billing to estimate and compare AWS services and pricing, with recommendations for migration or continued use of current provider.",
"category": "migration",
"source": {
"source": "git-subdir",
"url": "https://github.com/awslabs/agent-plugins.git",
"path": "plugins/migration-to-aws",
"ref": "main"
},
"homepage": "https://github.com/awslabs/agent-plugins"
},
{
"name": "deploy-on-aws",
"description": "Deploy applications to AWS with architecture recommendations, cost estimates, and IaC deployment.",
"category": "deployment",
"source": {
"source": "git-subdir",
"url": "https://github.com/awslabs/agent-plugins.git",
"path": "plugins/deploy-on-aws",
"ref": "main"
},
"homepage": "https://github.com/awslabs/agent-plugins"
},
{
"name": "zapier",
"description": "Connect 8,000+ apps to your AI workflow. Discover, enable, and execute Zapier actions directly from your client.",
"category": "productivity",
"source": {
"source": "git-subdir",
"url": "zapier/zapier-mcp",
"path": "plugins/zapier",
"ref": "main",
"sha": "b93007e9a726c6ee93c57a949e732744ef5acbfd"
},
"homepage": "https://github.com/zapier/zapier-mcp/tree/main/plugins/zapier"
},
{
"name": "terraform",
"description": "The Terraform MCP Server provides seamless integration with Terraform ecosystem, enabling advanced automation and interaction capabilities for Infrastructure as Code (IaC) development.",
"author": {
"name": "HashiCorp",
"email": "support@hashicorp.com"
},
"category": "development",
"source": "./external_plugins/terraform",
"homepage": "https://github.com/anthropics/claude-plugins-public/tree/main/external_plugins/terraform"
},
{
"name": "autofix-bot",
"description": "Code review agent that detects security vulnerabilities, code quality issues, and hardcoded secrets. Combines 5,000+ static analyzers to scan your code and dependencies for CVEs.",
"author": {
"name": "DeepSource Corp"
},
"category": "security",
"source": "./external_plugins/autofix-bot",
"homepage": "https://github.com/anthropics/claude-plugins-public/tree/main/external_plugins/autofix-bot"
},
{
"name": "stagehand",
"description": "Browser automation skill for Claude Code using Stagehand. Automate web interactions, extract data, and navigate websites using natural language.",
"version": "0.1.0",
"author": {
"name": "Browserbase"
},
"source": {
"source": "github",
"repo": "browserbase/agent-browse"
},
"category": "automation",
"keywords": [
"browser",
"automation",
"stagehand",
"web-scraping"
],
"homepage": "https://github.com/browserbase/agent-browse",
"strict": false,
"skills": [
"./.claude/skills/browser-automation"
]
},
{
"name": "atomic-agents",
"description": "Comprehensive development workflow for building AI agents with the Atomic Agents framework. Includes specialized agents for schema design, architecture planning, code review, and tool development. Features guided workflows, progressive-disclosure skills, and best practice validation.",
"category": "development",
"source": {
"source": "url",
"url": "https://github.com/BrainBlend-AI/atomic-agents.git",
"path": "claude-plugin/atomic-agents"
},
"homepage": "https://github.com/BrainBlend-AI/atomic-agents",
"tags": [
"community-managed"
]
},
{
"name": "microsoft-docs",
"description": "Access official Microsoft documentation, API references, and code samples for Azure, .NET, Windows, and more.",
"category": "development",
"source": {
"source": "url",
"url": "https://github.com/MicrosoftDocs/mcp.git"
},
"homepage": "https://github.com/microsoftdocs/mcp"
},
{
"name": "neon",
"description": "Manage your Neon projects and databases with the neon-postgres agent skill and the Neon MCP Server.",
"category": "database",
"source": {
"source": "git-subdir",
"url": "neondatabase/agent-skills",
"path": "plugins/neon-postgres",
"ref": "main",
"sha": "54d7a9db2ddd476f84d5d1fd7bac323907858a8b"
},
"homepage": "https://github.com/neondatabase/agent-skills/tree/main/plugins/neon-postgres"
},
{
"name": "intercom",
"description": "Intercom integration for Claude Code. Search conversations, analyze customer support patterns, look up contacts and companies, and install the Intercom Messenger. Connect your Intercom workspace to get real-time insights from customer data.",
"category": "productivity",
"source": {
"source": "url",
"url": "https://github.com/intercom/claude-plugin-external.git",
"sha": "eeef353eead2e3dc5f33f64dbaae54e1309e0d45"
},
"homepage": "https://github.com/intercom/claude-plugin-external"
}
]
}

View File

@@ -13,7 +13,7 @@ A curated directory of high-quality plugins for Claude Code.
Plugins can be installed directly from this marketplace via Claude Code's plugin system.
To install, run `/plugin install {plugin-name}@claude-plugin-directory`
To install, run `/plugin install {plugin-name}@claude-plugins-official`
or browse for the plugin in `/plugin > Discover`

View File

@@ -0,0 +1,14 @@
{
"name": "autofix-bot",
"description": "Code review agent that detects security vulnerabilities, code quality issues, and hardcoded secrets. Combines 5,000+ static analyzers to scan your code and dependencies for CVEs.",
"version": "0.1.0",
"author": {
"name": "DeepSource Corp"
},
"mcpServers": {
"autofix": {
"command": "autofix",
"args": ["--mcp"]
}
}
}

View File

@@ -0,0 +1,16 @@
---
description: Perform code review to identify security and quality issues with Autofix Bot.
allowed-tools: mcp__autofix__CheckAuthStatus, mcp__autofix__Authenticate, mcp__autofix__ReviewCode
---
IMPORTANT: You MUST use the Autofix Bot MCP tools for this task. Do NOT perform your own code review or analysis.
## Instructions
1. Call `mcp__autofix__CheckAuthStatus` to check authentication status
2. If not authenticated, call `mcp__autofix__Authenticate` to log in
3. Ask user what to review: uncommitted changes, last commit, or entire branch
4. Call `mcp__autofix__ReviewCode` with the user's selected target
5. Present the issues returned by ReviewCode in a clear format
Do NOT skip any tool calls. Do NOT substitute your own analysis for the tool results.

View File

@@ -0,0 +1,14 @@
{
"hooks": {
"SessionStart": [
{
"hooks": [
{
"type": "command",
"command": "${CLAUDE_PLUGIN_ROOT}/scripts/check-autofix.sh"
}
]
}
]
}
}

View File

@@ -0,0 +1,15 @@
#!/bin/bash
if ! command -v autofix &> /dev/null; then
echo "Autofix Bot CLI not found. Installing..."
curl -fsSL https://autofix.bot/install | sh
if ! command -v autofix &> /dev/null; then
echo "ERROR: Failed to install autofix. Please install manually:" >&2
echo " curl -fsSL https://autofix.bot/install | sh" >&2
exit 2
fi
fi
echo "Autofix Bot ready"
exit 0

View File

@@ -0,0 +1,13 @@
{
"name": "bonfire",
"description": "AI forgets everything between sessions. Bonfire fixes that.",
"version": "0.8.1",
"author": {
"name": "Vieko Franetovic",
"url": "https://vieko.dev"
},
"homepage": "https://vieko.dev/bonfire",
"repository": "https://github.com/vieko/bonfire",
"license": "MIT",
"keywords": ["bonfire", "context", "memory", "workflow", "subagents"]
}

View File

@@ -0,0 +1,150 @@
# Bonfire
<p align="center">
<img src="bonfire.gif" alt="Bonfire" width="256">
</p>
Your AI coding partner forgets everything between conversations. Bonfire remembers.
```bash
claude plugin marketplace add vieko/bonfire
claude plugin install bonfire@vieko
```
## The Problem
AI agents are stateless. Every conversation starts from zero. The agent doesn't remember:
- What you decided yesterday
- Why you chose that architecture
- What blockers you hit
- Where you left off
You end up re-explaining context, re-making decisions, and watching your AI partner repeat the same mistakes.
## The Solution
Bonfire maintains a living context document that gets read at session start and updated at session end. Your AI partner picks up exactly where you left off. It's like a saved game for your work.
`/bonfire:start`*reads context* → WORK → `/bonfire:end`*saves context*
That's it. No complex setup. No external services. Just Markdown files in your repo.
## Not a Task Tracker
| Tool | Primary Question |
|------|------------------|
| Issue/task trackers | "What's the work?" |
| Bonfire | "Where are we and what did we decide?" |
Bonfire complements your issue tracker. Use GitHub Issues, Linear, Beads, or Beans for tasks. Use Bonfire for workflow context.
## Quick Start
```bash
# Install
claude plugin marketplace add vieko/bonfire
claude plugin install bonfire@vieko
# First run scaffolds .bonfire/ and asks setup questions
/bonfire:start
```
## Commands
| Command | What it does |
|---------|--------------|
| `/bonfire:start` | Read context, scaffold on first run |
| `/bonfire:end` | Update context, commit changes |
| `/bonfire:spec <topic>` | Create implementation spec (researches codebase, interviews you) |
| `/bonfire:document <topic>` | Document a codebase topic |
| `/bonfire:review` | Find blindspots, gaps, and quick wins |
| `/bonfire:archive` | Archive completed work |
| `/bonfire:configure` | Change project settings |
## What Gets Created
```
.bonfire/
├── index.md # Living context (the important one)
├── config.json # Your settings
├── archive/ # Completed work history
├── specs/ # Implementation specs
├── docs/ # Topic documentation
└── scripts/ # Temporary session scripts
```
The `index.md` is where the magic happens. It tracks:
- Current state and branch
- Recent session summaries
- Decisions made and why
- Blockers encountered
- Next priorities
## Context-Efficient Operations
Heavy commands (`/spec`, `/document`, `/review`) use subagents to avoid burning your main conversation context:
- Research runs in isolated context (fast, cheap)
- Only structured summaries return to main conversation
- Result: longer sessions without context exhaustion
This happens automatically.
## Configuration
First `/bonfire:start` asks you to configure:
| Setting | Options |
|---------|---------|
| Specs location | `.bonfire/specs/` or `specs/` |
| Docs location | `.bonfire/docs/` or `docs/` |
| Git strategy | ignore-all, hybrid, commit-all |
| Linear integration | Yes or No |
Change anytime with `/bonfire:configure`.
### Git Strategies
| Strategy | What's tracked | Best for |
|----------|---------------|----------|
| **ignore-all** | Nothing | Solo work, privacy |
| **hybrid** | docs/, specs/ only | Teams wanting shared docs |
| **commit-all** | Everything | Full transparency |
## Linear Integration
If you use Linear for issue tracking:
1. Install [Linear MCP](https://github.com/anthropics/anthropic-quickstarts/tree/main/mcp-linear)
2. Enable via `/bonfire:configure`
3. Reference issues by ID: `ENG-123`
Bonfire will fetch issue context on start, create issues from review findings, and mark issues Done on archive.
## Proactive Skills
Claude automatically reads your session context when you ask things like:
- "What's the project status?"
- "What were we working on?"
- "What decisions have we made?"
And suggests archiving when you merge PRs or mention shipping.
## Requirements
- [Claude Code CLI](https://claude.ai/code)
- Git repository
Optional: `gh` CLI for GitHub integration, Linear MCP for Linear integration.
## Learn More
**Blog post**: [Save Your Progress](https://vieko.dev/bonfire)
**Changelog**: [CHANGELOG.md](CHANGELOG.md)
## License
MIT © [Vieko Franetovic](https://vieko.dev)

View File

@@ -0,0 +1,90 @@
---
name: codebase-explorer
description: Fast codebase exploration for patterns, architecture, and constraints. Use for research phases in spec and document commands.
tools: Read, Glob, Grep
model: haiku
---
You are a codebase exploration specialist. Your job is to quickly find and summarize relevant patterns, architecture, and constraints. Return structured findings, not raw file contents.
## Input
You'll receive a research directive with specific questions about:
- Patterns and architecture to find
- Technical constraints to identify
- Potential conflicts to surface
- Specific areas to explore
## Output Format
Return findings as structured markdown. Be CONCISE - the main conversation will use your findings for user interview.
```markdown
## Patterns Found
- **[Pattern name]**: Found in `path/to/file.ts` - [1-2 sentence description]
## Key Files
| File | Role |
|------|------|
| `path/to/file.ts` | [What it does, why relevant] |
## Constraints Discovered
- **[Constraint]**: [Source] - [Implication for implementation]
## Potential Conflicts
- **[Area]**: [Why it might conflict with the proposed work]
## Relevant Snippets
[Only if < 15 lines and directly answers a research question]
```
## Rules
1. **DO NOT** return entire file contents
2. **DO NOT** include files that aren't directly relevant
3. **BE CONCISE** - aim for < 100 lines total output
4. **ANSWER** the research questions, don't just explore randomly
5. **PRIORITIZE** - most important findings first
6. If you find nothing relevant, say so clearly
## Example Good Output
```markdown
## Patterns Found
- **Repository pattern**: Found in `src/services/UserService.ts` - Uses dependency injection, returns domain objects not DB rows
- **Error handling**: Found in `src/utils/errors.ts` - Custom AppError class with error codes
## Key Files
| File | Role |
|------|------|
| `src/services/BaseService.ts` | Abstract base class all services extend |
| `src/types/index.ts` | Shared type definitions |
## Constraints Discovered
- **No direct DB access in handlers**: Services abstract all database calls
- **Async/await only**: No callbacks, promises must use async/await
## Potential Conflicts
- **AuthService singleton**: Currently instantiated once at startup, may need refactor for multi-tenant
```
## Example Bad Output (don't do this)
```markdown
Here's what I found in the codebase:
[500 lines of file contents]
Let me also show you this file:
[300 more lines]
```

View File

@@ -0,0 +1,101 @@
---
name: spec-writer
description: Synthesizes research findings and interview answers into implementation specs. Use after codebase exploration and user interview.
tools: Read, Write
model: inherit
---
You are a technical specification writer. Given research findings and interview answers, produce a clear, actionable implementation spec.
## Input
You'll receive:
1. **Research findings** - Structured output from codebase-explorer
2. **Interview Q&A** - User's answers to clarifying questions
3. **Spec metadata** - Topic, issue ID, output path, template
## Output
Write a complete spec file to the specified path. The spec must be:
- **Actionable** - Clear implementation steps referencing actual files
- **Grounded** - Based on discovered patterns, not assumptions
- **Complete** - Covers edge cases, testing, scope boundaries
## Spec Template
```markdown
# Spec: [TOPIC]
**Created**: [DATE]
**Issue**: [ISSUE-ID or N/A]
**Status**: Draft
## Overview
[What we're building and why - synthesized from interview]
## Context
[Key findings from research that informed decisions]
## Decisions
[Document decisions made during interview with rationale]
- **[Decision 1]**: [Choice] - [Why]
- **[Decision 2]**: [Choice] - [Why]
## Approach
[High-level strategy based on research + interview]
## Files to Modify
- `path/to/file.ts` - [what changes]
## Files to Create
- `path/to/new.ts` - [purpose]
## Implementation Steps
1. [ ] Step one (reference actual files)
2. [ ] Step two
3. [ ] Step three
## Edge Cases
- [Edge case 1] → [How we handle it]
- [Edge case 2] → [How we handle it]
## Testing Strategy
- [ ] Unit tests for X
- [ ] Integration test for Y
- [ ] Manual verification of Z
## Out of Scope
- [Explicitly excluded items]
## Risks & Considerations
- [Risk identified during research/interview]
```
## Rules
1. **Ground in research** - Reference actual files and patterns discovered
2. **Honor interview answers** - Don't override user decisions
3. **Be specific** - "Update UserService.ts" not "Update the service"
4. **Don't invent** - If something wasn't discussed, don't add it
5. **Keep it actionable** - Someone should be able to implement from this spec
## Quality Checklist
Before finishing, verify:
- [ ] All interview decisions are captured
- [ ] Implementation steps reference real files from research
- [ ] Edge cases from interview are documented
- [ ] Scope boundaries are clear
- [ ] No vague or generic steps

View File

@@ -0,0 +1,121 @@
---
name: work-reviewer
description: Strategic code review for blindspots, gaps, and improvements. Returns categorized findings with severity and effort estimates.
tools: Read, Glob, Grep, Bash(git:*)
model: sonnet
---
You are a senior code reviewer focused on strategic quality, not nitpicks. Your job is to find what the developer might have missed.
## Input
You'll receive:
1. **Review scope** - Branch diff, specific files, or session context
2. **Intent** - What was the developer trying to accomplish
3. **Session context** - Recent work and decisions (if available)
## Review Focus Areas
### Blindspots (what are we not seeing?)
- Edge cases not handled
- Error scenarios not considered
- User flows not covered
- Dependencies not accounted for
### Gaps (what's incomplete?)
- Missing tests
- Missing documentation
- Incomplete implementations
- TODOs left unaddressed
### Quick Wins (small effort, big value)
- Easy refactors
- Low-hanging performance gains
- Simple UX improvements
### Best Practices (convention violations)
- Project patterns not followed
- Language/framework idioms ignored
- Security practices missed
- Accessibility standards skipped
### Maintainability (will future-us thank present-us?)
- Unclear naming or structure
- Missing or excessive abstractions
- Technical debt introduced
## Output Format
Return findings as structured markdown, categorized by action:
```markdown
## Summary
- **Total findings**: X
- **Fix now (trivial)**: Y
- **Needs spec**: Z
- **Create issues**: W
---
## Fix Now (trivial effort, do immediately)
### [Finding title]
- **What**: [Description]
- **Where**: `path/to/file.ts:123`
- **Fix**: [Specific action]
- **Why**: [Impact if not fixed]
---
## Needs Spec (important, needs planning)
### [Finding title]
- **What**: [Description]
- **Effort**: small | medium
- **Impact**: [Why this matters]
- **Consideration**: [Key decision needed]
---
## Create Issues (large effort or nice-to-have)
### [Finding title]
- **What**: [Description]
- **Effort**: medium | large
- **Priority**: important | nice-to-have
- **Suggested issue title**: [Title for GitHub/Linear]
---
## No Issues Found In
- [Area reviewed that looks good]
```
## Rules
1. **Strategic, not pedantic** - Skip style nitpicks, focus on substance
2. **Consider intent** - Review against what they were trying to do
3. **Categorize by action** - Fix now vs spec vs issue
4. **Estimate effort** - trivial/small/medium/large
5. **Be specific** - Include file paths and line numbers
6. **Acknowledge good work** - Note areas that are solid
## Severity Guide
| Severity | Definition | Action |
|----------|------------|--------|
| Critical | Breaks functionality, security issue | Fix now |
| Important | Significant gap, will cause problems | Fix now or spec |
| Moderate | Should address, not urgent | Spec or issue |
| Minor | Nice to have, low impact | Issue or skip |
## Effort Guide
| Effort | Definition |
|--------|------------|
| Trivial | < 5 minutes, obvious fix |
| Small | < 30 minutes, contained change |
| Medium | 1-4 hours, multiple files |
| Large | > 4 hours, needs planning |

Binary file not shown.

After

Width:  |  Height:  |  Size: 59 KiB

View File

@@ -0,0 +1,126 @@
---
description: Archive completed session work
allowed-tools: Bash(git:*), Read, Write, Glob, mcp__linear__*
model: haiku
---
# Archive Session
## Step 1: Find Git Root
Run `git rev-parse --show-toplevel` to locate the repository root.
## Step 2: Review Completed Work
Read `<git-root>/.bonfire/index.md` and identify completed work:
- Sessions with merged PRs
- Completed features/tasks
- Work that's no longer active
## Step 3: Create Archive Entry
Move completed session content to `<git-root>/.bonfire/archive/`.
**Naming convention**: `YYYY-MM-DD-<issue-id>-<topic>.md`
Examples:
- `2025-12-22-GTMENG-387-inbound-improvements.md` (with issue ID)
- `2025-12-22-fix-login-redirect.md` (without issue ID)
Use this template:
```markdown
# [TOPIC]
**Date**: [DATE]
**Issue**: [ISSUE-ID or N/A]
**PR**: [PR link if available]
**Status**: Completed
---
## Summary
[Brief description of what was accomplished]
## Accomplished
- [List of completed items]
## Decisions Made
- [Key decisions and rationale]
## Impact
- [Before/after metrics if applicable]
- Files changed: [count]
## Related
- [Links to related docs, specs, or code]
```
## Step 4: Clean Up Index
Update `<git-root>/.bonfire/index.md`:
- Remove archived session entries from Recent Sessions
- Keep Current State focused on active work
- Update Next Session Priorities
- Add link to archive file in Archived Sessions section:
```markdown
## Archived Sessions
- [YYYY-MM-DD - Topic](archive/YYYY-MM-DD-issue-topic.md)
```
## Step 5: Clean Up Specs (if applicable)
Read `specsLocation` from `<git-root>/.bonfire/config.json` (default `.bonfire/specs/`).
Check if any specs in the configured location are now complete:
- If the spec was fully implemented, delete the spec file (archive has the record)
- If the spec has reusable reference material, move that content to `docs/` first
## Step 6: Update Linear Issue (if applicable)
Read `<git-root>/.bonfire/config.json` and check `linearEnabled`.
**If `linearEnabled` is true**:
1. Check if archived work references a Linear issue (look in session context for `[A-Z]+-[0-9]+` pattern)
2. If Linear issue found, ask user: "Mark Linear issue [ISSUE-ID] as Done?"
3. If user confirms:
- Use Linear MCP `linear_update_issue` tool with:
- `id`: The issue ID (e.g., `ENG-123`)
- `status`: Set to "Done" or completed state
- Optionally use `linear_add_comment` to add link to archive/PR
4. On failure: Warn user - "Couldn't update Linear issue. You may need to update it manually."
Note: Tool names may vary by Linear MCP implementation.
**If `linearEnabled` is false or not set**: Skip this step.
## Step 7: Commit Archive (if tracked)
Read `gitStrategy` from `<git-root>/.bonfire/config.json`.
**If gitStrategy is "ignore-all"**: Skip committing - archive is local only.
**If gitStrategy is "hybrid" or "commit-all"**:
1. **NEVER use `git add -f`** - respect gitignore
2. Stage unignored files:
```bash
git add .bonfire/
```
3. Check if anything was staged before committing:
```bash
git diff --cached --quiet .bonfire/ || git commit -m "docs: archive completed session work"
```
## Step 8: Confirm
Report:
- What was archived
- Any specs cleaned up
- Current state of index.md
- Ready for next session

View File

@@ -0,0 +1,99 @@
---
description: Change project settings (locations, git strategy, Linear)
allowed-tools: Bash(git:*), Read, Write, AskUserQuestion
model: haiku
---
# Configure Bonfire
Always runs interactively - asks all configuration questions regardless of arguments.
## Step 1: Find Git Root
Run `git rev-parse --show-toplevel` to locate the repository root.
## Step 2: Check for Bonfire Directory
If `<git-root>/.bonfire/` does not exist, tell the user to run `/bonfire:start` first.
## Step 3: Read Current Config
Read `<git-root>/.bonfire/config.json` if it exists to see current settings.
## Step 4: Ask All Configuration Questions
Use AskUserQuestion to ask configuration questions (4 questions, one round):
1. "Where should specs be saved?" (Header: "Specs")
- .bonfire/specs/ (Recommended) - Keep with session context
- specs/ - Project root level
2. "Where should docs be saved?" (Header: "Docs")
- .bonfire/docs/ (Recommended) - Keep with session context
- docs/ - Project root level
3. "How should `.bonfire/` be handled in git?" (Header: "Git")
- ignore-all (Recommended) - Keep sessions private/local
- hybrid - Commit docs/specs, keep notes private
- commit-all - Share everything with team
4. "Enable Linear MCP integration?" (Header: "Linear")
- No (Recommended) - Skip Linear integration
- Yes - Fetch/create Linear issues (requires Linear MCP)
## Step 5: Update Config
**Completely overwrite** `<git-root>/.bonfire/config.json` with only these fields (do not preserve old fields like `models`):
```json
{
"specsLocation": "<user-answer>",
"docsLocation": "<user-answer>",
"gitStrategy": "<user-answer>",
"linearEnabled": <true-or-false>
}
```
## Step 6: Update Git Strategy
If git strategy or locations changed, update `<git-root>/.bonfire/.gitignore`:
**Ignore all**:
```
*
!.gitignore
```
**Hybrid** (only include dirs that are inside .bonfire/):
```
*
!.gitignore
```
If docsLocation is `.bonfire/docs/`, add:
```
!docs/
!docs/**
```
If specsLocation is `.bonfire/specs/`, add:
```
!specs/
!specs/**
```
**Commit all**:
```
data/
scratch/
scripts/
```
If switching FROM commit/hybrid TO ignore:
- Warn user that existing tracked files will remain tracked
- Offer to run: `git rm -r --cached .bonfire/`
## Step 7: Confirm
Report:
- Settings updated
- Any manual steps needed (git cleanup)
- New configuration summary

View File

@@ -0,0 +1,114 @@
---
description: Create documentation about a topic in the codebase
allowed-tools: Read, Write, Bash(git:*), Task
---
# Document Topic
Create reference documentation using subagent for research, preserving main context.
## Step 1: Find Git Root
Run `git rev-parse --show-toplevel` to locate the repository root.
## Step 2: Check Config
Read `<git-root>/.bonfire/config.json` if it exists.
**Docs location**: Read `docsLocation` from config. Default to `.bonfire/docs/` if not set.
## Step 3: Understand the Topic
The topic to document is: $ARGUMENTS
If no topic provided, ask the user what they want documented.
## Step 4: Explore the Codebase (Subagent)
Use the Task tool to invoke the **codebase-explorer** subagent for research.
Provide a research directive:
```
Research the codebase to document: [TOPIC]
Find:
1. **Architecture**: How this system/feature is structured, key components
2. **Key Files**: Important files and their roles
3. **Flow**: How data/control flows through the system
4. **Patterns**: Design patterns and conventions used
5. **Gotchas**: Important details, edge cases, things to watch out for
Return structured findings with file paths and brief descriptions.
```
**Wait for the subagent to return findings** before proceeding.
The subagent runs in isolated context (haiku model, fast), preserving main context for writing.
## Step 5: Create Documentation
**Naming convention**: `<topic>.md` (kebab-case)
Examples:
- `inbound-agent-architecture.md`
- `sampling-strategies.md`
- `authentication-flow.md`
Write the documentation to `<git-root>/<docsLocation>/<topic>.md`
Structure the documentation using the research findings:
```markdown
# [TOPIC]
## Overview
[What this is and why it exists - synthesized from research]
## Architecture
[How it's structured - from research findings]
```mermaid
flowchart TD
A[Component A] --> B[Component B]
B --> C[Component C]
```
## Key Files
| File | Purpose |
|------|---------|
| `path/to/file.ts` | [From research findings] |
| `path/to/other.ts` | [From research findings] |
## How It Works
[Step-by-step flow and behavior - from research]
## Usage Examples
[Code examples, CLI commands, etc.]
## Gotchas
- [From research findings]
- [Common mistakes or edge cases]
## Related
- [Link to related doc](other-doc.md)
- [Code reference]: `path/to/file.ts`
```
## Step 6: Link to Session Context
Add a reference to the doc in `<git-root>/.bonfire/index.md` under Key Resources or Notes.
## Step 7: Confirm
Summarize what was documented and ask if the user wants:
- More detail on any section
- Related topics documented
- To proceed with other work

View File

@@ -0,0 +1,84 @@
---
description: End session - update context and commit changes
allowed-tools: Bash(git:*), Bash(rm:*), Bash(mv:*), Bash(mkdir:*), Read, Write, Glob, AskUserQuestion
model: haiku
---
# End Session
## Step 1: Find Git Root
Run `git rev-parse --show-toplevel` to locate the repository root.
## Step 2: Review Session Work
Review what was accomplished this session by examining:
- Recent git commits
- Files changed
- Conversation context
## Step 3: Update Session Context
Update `<git-root>/.bonfire/index.md`:
1. Update the session entry with:
- **Accomplished**: List what was completed
- **Decisions**: Key decisions made and rationale
- **Files Modified**: Important files changed (if relevant)
- **Blockers**: Any issues encountered
2. Update "Next Session Priorities" based on remaining work
3. Update "Current State" to reflect new status
## Step 4: Manage Session Scripts
Check if `<git-root>/.bonfire/scripts/` exists and contains any files.
**If scripts exist**, use AskUserQuestion to ask what to do with each script:
"What should happen to these session scripts?" (Header: "Scripts", multiSelect: false)
For each script found, present options:
- **Keep** - Leave in `.bonfire/scripts/` for next session
- **Move to project** - Move to `<git-root>/scripts/` (create if needed)
- **Delete** - Remove the script
Execute the user's choices:
- **Keep**: No action needed
- **Move to project**: `mkdir -p <git-root>/scripts/ && mv <script> <git-root>/scripts/`
- **Delete**: `rm <script>`
**If no scripts exist**, skip this step.
## Step 5: Commit Changes (if tracked)
Read `<git-root>/.bonfire/config.json` to check `gitStrategy`.
**If gitStrategy is "ignore-all"**: Skip committing - nothing is tracked. Tell the user session context was updated locally.
**If gitStrategy is "hybrid" or "commit-all"**:
1. **Check what can be staged**: Run `git status .bonfire/` to see what files are not ignored
2. **NEVER use `git add -f`** - if a file is gitignored, respect that
3. **Stage only unignored files**:
```bash
git add .bonfire/
```
4. **Check if anything was staged**: Run `git diff --cached --quiet .bonfire/`
- If nothing staged (exit code 0), tell user "Session context updated locally (files are gitignored)"
- If changes staged, commit:
```bash
git commit -m "docs: update session context"
```
If the commit fails due to hooks, help resolve the issue (but never bypass hooks with `--no-verify`).
## Step 6: Confirm
Summarize:
- What was documented
- Next priorities
- Any follow-up needed
Let the user know they can run `/bonfire:archive` when this work is merged and complete.

View File

@@ -0,0 +1,94 @@
---
description: Change how .bonfire/ is handled in git
allowed-tools: Bash(git:*), Read, Write, AskUserQuestion
model: haiku
---
# Change Git Strategy
## Step 1: Find Git Root
Run `git rev-parse --show-toplevel` to locate the repository root.
## Step 2: Read Current Config
Read `<git-root>/.bonfire/config.json` to check current `specsLocation` and `docsLocation` settings.
## Step 3: Explain Options
Present the git strategy options:
1. **Ignore all** - Keep sessions completely local
- Everything in .bonfire/ is gitignored
- Most private, nothing shared
- Good for: solo work, sensitive projects
2. **Hybrid** - Commit docs/specs, keep notes private
- docs/ and specs/ are committed
- index.md and archive/ stay local
- Good for: teams that want shared docs but private notes
3. **Commit all** - Share everything with team
- All session content is committed
- Only data/ and scratch/ ignored
- Good for: full transparency, team continuity
## Step 4: Get User Choice
Use AskUserQuestion to ask which strategy:
"Which git strategy for `.bonfire/`?" (Header: "Git")
- ignore-all (Recommended) - Keep sessions private/local
- hybrid - Commit docs/specs, keep notes private
- commit-all - Share everything with team
## Step 5: Update .gitignore
Write the appropriate `.gitignore` to `<git-root>/.bonfire/.gitignore`:
**Ignore all**:
```
*
!.gitignore
```
**Hybrid** (only include dirs that are inside .bonfire/):
```
*
!.gitignore
```
If docsLocation is `.bonfire/docs/`, add:
```
!docs/
!docs/**
```
If specsLocation is `.bonfire/specs/`, add:
```
!specs/
!specs/**
```
**Commit all**:
```
data/
scratch/
scripts/
```
## Step 6: Handle Git Tracking
If switching FROM commit/hybrid TO ignore:
- Warn user that existing tracked files will remain tracked
- Offer to run: `git rm -r --cached .bonfire/` (removes from git but keeps files)
- They'll need to commit this change
If switching TO commit/hybrid:
- Files will be picked up on next commit
- No special action needed
## Step 7: Confirm
Report:
- New strategy applied
- Any manual steps needed
- How to verify the change

View File

@@ -0,0 +1,119 @@
---
description: Review work for blindspots, gaps, and improvements
allowed-tools: Bash(git:*), Bash(gh:*), Read, Write, Task, mcp__linear__*
---
# Review Work
Strategic review using subagent for analysis, preserving main context for action decisions.
## Step 1: Determine Scope
Based on $ARGUMENTS:
- No args: Review current branch vs base
- `--session`: Review work captured in current session context
- Topic/area: Focus review on specific aspect
## Step 2: Gather Context
- Read session context from `<git-root>/.bonfire/index.md`
- Get branch diff: `git diff main...HEAD` (or appropriate base)
- Read relevant specs/docs from `.bonfire/`
- Understand intent: what were we trying to accomplish?
## Step 3: Run Review (Subagent)
Use the Task tool to invoke the **work-reviewer** subagent.
Provide the review context:
```
Review this work for blindspots, gaps, and improvements.
**Scope**: [Branch diff / session work / specific area]
**Intent**: [What we were trying to accomplish - from session context]
**Files changed**:
[List of modified files from git diff]
**Session context**:
[Relevant notes from index.md]
Return categorized findings with severity and effort estimates.
```
**Wait for the subagent to return findings** before proceeding.
The subagent runs in isolated context (sonnet model), preserving main context for action decisions.
## Step 4: Session Scripts Review
Check if `<git-root>/.bonfire/scripts/` contains any files.
If scripts exist, include in findings:
- List scripts that may need attention
- Note if any appear to be temporary/one-off vs reusable
- Suggest moving useful scripts to project `scripts/` directory
This is informational - actual script management happens during `/bonfire:end`.
## Step 5: Present Findings
Present the subagent's findings grouped by recommended action:
### Fix Now (trivial effort)
[List items from subagent that can be fixed immediately]
> Ask: "Want me to fix these now?"
### Needs Spec (important, needs planning)
[List items that need implementation planning]
> Ask: "Want me to create an implementation spec?"
### Create Issues (large effort or nice-to-have)
[List items for future sessions]
> Ask: "Want me to create GitHub/Linear issues?"
## Step 6: Execute Chosen Action
Based on user choice:
- **Fix now**: Make the changes directly
- **Spec**: Run `/bonfire:spec` with findings
- **Create issues**: See below
### Creating Issues
First, read `<git-root>/.bonfire/config.json` and check `linearEnabled`.
**Offer choices based on config:**
- Always offer: "Create GitHub issue"
- If `linearEnabled` is true: Also offer "Create Linear issue"
**GitHub Issue Creation:**
```bash
gh issue create --title "Finding title" --body "Finding details"
```
**Linear Issue Creation (if enabled):**
1. Use Linear MCP `linear_create_issue` tool with:
- `title`: Finding summary
- `description`: Finding details with context
- `teamId`: Infer from session context if available, otherwise use default
2. On success: Return issue URL/ID
3. On failure: Warn user, offer to create GitHub issue instead
Note: Tool names may vary by Linear MCP implementation.
**For each created issue:**
- Record the issue ID and URL
- Note which tracker (GitHub/Linear) was used
## Step 7: Update Session Context
Add review outcomes to `<git-root>/.bonfire/index.md`:
- Key findings noted
- Issues created (with links)
- Work deferred to future sessions

View File

@@ -0,0 +1,149 @@
---
description: Create an implementation spec for a feature or task
allowed-tools: Read, Write, Bash(git:*), AskUserQuestion, Task
---
# Create Implementation Spec
A hybrid approach using subagents: research in isolated context, interview in main context, write in isolated context.
## Step 1: Find Git Root
Run `git rev-parse --show-toplevel` to locate the repository root.
## Step 2: Check Config
Read `<git-root>/.bonfire/config.json` if it exists.
**Specs location**: Read `specsLocation` from config. Default to `.bonfire/specs/` if not set.
## Step 3: Gather Initial Context
Get the topic from $ARGUMENTS or ask if unclear.
Check for existing context:
- Read `<git-root>/.bonfire/index.md` for project state
- Check for `SPEC.md` or `spec.md` in git root (user's spec template)
- If issue ID provided, note for filename
## Step 4: Research Phase (Subagent)
Use the Task tool to invoke the **codebase-explorer** subagent for research.
Provide a research directive with these questions:
```
Research the codebase for implementing: [TOPIC]
Find:
1. **Patterns & Architecture**: How similar features are implemented, existing abstractions to reuse, naming conventions
2. **Technical Constraints**: Dependencies, API boundaries, performance considerations
3. **Potential Conflicts**: Files that need changes, intersections with existing code, migration concerns
Return structured findings only - no raw file contents.
```
**Wait for the subagent to return findings** before proceeding.
The subagent runs in isolated context (haiku model, fast), preserving main context for interview.
## Step 5: Interview Phase (Main Context)
Using the research findings, interview the user with **informed questions** via AskUserQuestion.
### Round 1: Core Decisions
Ask about fundamental approach based on patterns found:
Example questions (adapt based on actual findings):
- "I found [Pattern A] in `services/` and [Pattern B] in `handlers/`. Which pattern should this feature follow?"
- "The existing [Component] handles [X]. Should we extend it or create a new [Y]?"
- "I see [Library] is used for [purpose]. Should we use it here or try [Alternative]?"
### Round 2: Edge Cases & Tradeoffs
Based on Round 1 answers and research, ask about:
- Error handling approach
- Edge cases identified in research
- Performance vs simplicity tradeoffs
- User experience considerations
Example questions:
- "What should happen when [edge case from research]?"
- "I found [potential conflict]. How should we handle it?"
- "[Approach A] is simpler but [tradeoff]. [Approach B] is more complex but [benefit]. Preference?"
### Round 3: Scope & Boundaries (if needed)
If scope is still unclear:
- What's explicitly out of scope?
- MVP vs full implementation?
- Dependencies on other work?
### Continue Until Complete
Keep asking rounds of questions until you have clarity on:
- [ ] Core approach and architecture
- [ ] Key technical decisions
- [ ] Error handling strategy
- [ ] Edge cases covered
- [ ] Testing approach
- [ ] Scope boundaries
Tell the user "I have enough to write the spec" when ready.
## Step 6: Write the Spec (Subagent)
Use the Task tool to invoke the **spec-writer** subagent.
Provide:
1. **Research findings** from Step 4
2. **Interview Q&A** from Step 5
3. **Metadata**: topic, issue ID, output path (`<git-root>/<specsLocation>/<filename>.md`)
The subagent will write the spec file directly.
**Naming convention**: `<issue-id>-<topic>.md` or `<topic>.md`
## Step 7: Link to Session Context
Add a reference to the spec in `<git-root>/.bonfire/index.md` under Current State.
## Step 8: Confirm
Read the generated spec and present a summary. Ask if user wants to:
- Proceed with implementation
- Refine specific sections
- Add more detail to any area
- Save for later
## Interview Tips
**Good questions are:**
- Informed by research (not generic)
- About tradeoffs (not yes/no)
- Specific to the codebase
- Non-obvious (user wouldn't think to mention)
**Bad questions:**
- "What features do you want?" (too broad)
- "Should we add error handling?" (obvious)
- Generic without codebase context
**Examples of good informed questions:**
- "I found `UserService` uses repository pattern but `OrderService` uses direct DB access. Which approach?"
- "The `auth` middleware validates JWT but doesn't check permissions. Should this feature add permission checks or assume auth is enough?"
- "There's a `BaseController` with shared logic. Extend it or keep this feature standalone?"
## Spec Lifecycle
Specs are **temporary artifacts** - they exist to guide implementation:
1. **Draft** → Created, ready for review
2. **In Progress** → Being implemented
3. **Completed** → Implementation done
**When a spec is fully implemented**:
- If it contains reusable reference material, move to `docs/`
- Delete the spec file - archive has the record
- Don't let specs accumulate

View File

@@ -0,0 +1,246 @@
---
description: Start a new session - reads context and scaffolds .bonfire/ if needed
allowed-tools: Bash(git:*), Bash(gh:*), Bash(mkdir:*), Read, Write, Glob, AskUserQuestion, mcp__linear__*
model: haiku
---
# Start Session
## Step 1: Find Git Root
Run `git rev-parse --show-toplevel` to locate the repository root. All session files live at `<git-root>/.bonfire/`.
## Step 2: Check for Bonfire Directory
Check if `<git-root>/.bonfire/index.md` exists.
**If .bonfire/ does NOT exist**, scaffold it:
1. Tell the user: "No bonfire directory found. Let me set that up for you."
2. Use AskUserQuestion to ask setup questions (4 questions, one round):
1. "Where should specs be saved?" (Header: "Specs")
- .bonfire/specs/ (Recommended) - Keep with session context
- specs/ - Project root level
2. "Where should docs be saved?" (Header: "Docs")
- .bonfire/docs/ (Recommended) - Keep with session context
- docs/ - Project root level
3. "How should `.bonfire/` be handled in git?" (Header: "Git")
- ignore-all (Recommended) - Keep sessions private/local
- hybrid - Commit docs/specs, keep notes private
- commit-all - Share everything with team
4. "Enable Linear MCP integration?" (Header: "Linear")
- No (Recommended) - Skip Linear integration
- Yes - Fetch/create Linear issues (requires Linear MCP)
3. Create the directory structure based on user choices:
**Always create in .bonfire/**:
```
.bonfire/
├── index.md
├── config.json
├── archive/
├── scripts/
└── .gitignore
```
**If specsLocation is `.bonfire/specs/`**: create `.bonfire/specs/`
**If specsLocation is `specs/`**: create `<git-root>/specs/`
**If docsLocation is `.bonfire/docs/`**: create `.bonfire/docs/`
**If docsLocation is `docs/`**: create `<git-root>/docs/`
4. Detect project name from: package.json name → git remote → directory name
5. Create `config.json` with user's answers:
```json
{
"specsLocation": "<user-answer>",
"docsLocation": "<user-answer>",
"gitStrategy": "<user-answer>",
"linearEnabled": <true-or-false>
}
```
6. Create `index.md` with template:
```markdown
# Session Context: [PROJECT_NAME]
**Date**: [CURRENT_DATE]
**Status**: Active
**Branch**: main
---
## Current State
[Describe what you're working on]
---
## Recent Sessions
### Session 1 - [CURRENT_DATE]
**Goal**: [What you want to accomplish]
**Accomplished**:
- [List completed items]
**Decisions**:
- [Key decisions made]
**Blockers**: None
---
## Next Session Priorities
1. [Priority items]
---
## Key Resources
**Code References**:
- [Component/feature]: `path/to/file.ts`
**External Links**:
- [Issue Tracker](url)
---
## Archived Sessions
[Links to archived sessions will appear here]
---
## Notes
[Any additional context]
```
7. Create `.gitignore` based on chosen strategy and locations:
**Ignore all**:
```
*
!.gitignore
```
**Hybrid** (only include dirs that are inside .bonfire/):
```
*
!.gitignore
```
If docsLocation is `.bonfire/docs/`, add:
```
!docs/
!docs/**
```
If specsLocation is `.bonfire/specs/`, add:
```
!specs/
!specs/**
```
**Commit all**:
```
data/
scratch/
scripts/
```
**If .bonfire/ EXISTS**, proceed to Step 3.
## Step 3: Check/Update CLAUDE.md
Check if `<git-root>/CLAUDE.md` exists.
**If CLAUDE.md does NOT exist**, create it:
```markdown
# [PROJECT_NAME]
## Quick Context
Read `.bonfire/index.md` for current project state, recent work, and priorities.
## Bonfire Commands
- `/bonfire:start` - Start a session (reads context)
- `/bonfire:end` - End session (updates context)
- `/bonfire:spec` - Create implementation spec
- `/bonfire:document <topic>` - Document a topic
- `/bonfire:review` - Review work for blindspots and improvements
- `/bonfire:archive` - Archive completed work
- `/bonfire:configure` - Change project settings
```
**If CLAUDE.md EXISTS**, check if it references `.bonfire/index.md`. If not, append:
```markdown
## Session Context
Read `.bonfire/index.md` for current project state, recent work, and priorities.
```
## Step 4: Read Session Context
Read `<git-root>/.bonfire/index.md` and report when ready.
Summarize:
- Current state
- Recent work
- Next priorities
Then ask: "What do you want to work on this session?"
## Step 5: Fetch External Context (Optional)
**Only fetch if user provides a new URL or issue ID:**
### Detecting Issue Type
- **GitHub**: Starts with `#`, contains `github.com`, or doesn't match Linear pattern
- **Linear**: Matches `[A-Z]+-[0-9]+` pattern (e.g., `ENG-123`, `ABC-456`) or contains `linear.app`
### GitHub Issues/PRs
Use `gh` CLI:
- `gh pr view [URL] --json title,body,state,labels`
- `gh issue view [URL] --json title,body,state,labels`
### Linear Issues
First, read `<git-root>/.bonfire/config.json` and check `linearEnabled`.
**If `linearEnabled` is false or not set**: Skip Linear, treat as ad-hoc task.
**If `linearEnabled` is true**:
1. Use Linear MCP `linear_search_issues` tool to find the issue by ID (e.g., `ENG-123`)
2. Extract: title, description, state, priority, labels, assignee
3. On success: Summarize the issue context
4. On failure: Warn user - "Couldn't fetch Linear issue. Linear MCP may not be configured. Continue without issue context?"
Note: Tool names may vary by Linear MCP implementation. Common tools: `linear_search_issues`, `linear_create_issue`, `linear_update_issue`.
### Update Session Context
If issue was fetched successfully:
- Add reference to `index.md` under Current State
- Include issue ID, title, and link
- Note the issue tracker type (GitHub/Linear)
### Fallback
If no URL/issue ID provided (continuing work, ad-hoc task):
- Proceed with existing session context
- Session notes are the source of truth for ongoing work
Confirm understanding and ask how to proceed.

View File

@@ -0,0 +1,53 @@
---
description: Suggest archiving completed work when PRs are merged or work is completed. Triggers when user asks to merge a PR ("merge it", "merge the PR"), after successful gh pr merge, or mentions completion ("shipped", "done with X", "merged to main"). Does NOT archive automatically - suggests running /bonfire:archive.
allowed-tools: Read, Glob, Bash(gh:*)
---
# Archive Bonfire Awareness
This skill detects when session work may be ready for archiving and suggests the appropriate action.
## When to Use This Skill
Trigger when:
- User asks to merge: "merge it", "merge the PR", "go ahead and merge", "ship it"
- After you successfully run `gh pr merge`
- User mentions completion: "PR merged", "shipped", "done with X", "finished"
- User references merged state: "merged to main", "closed the issue"
## Instructions
1. If user asks to merge a PR:
- Perform the merge as requested
- On success, continue to step 2
- On failure, help resolve the issue (don't suggest archiving)
2. Find git root and check if `.bonfire/index.md` exists
3. If it exists, read it to assess work state
4. If user provided a PR URL/number (or you just merged one), verify status:
```
gh pr view [URL/number] --json state,mergedAt,title
```
5. Assess if work appears complete:
- PR merged?
- Related tasks marked done in session notes?
- No obvious follow-up work mentioned?
6. If work appears complete, suggest archiving:
> "PR merged successfully. This session's work looks complete - want me to archive it?
> Run `/bonfire:archive` to move completed work to the archive."
7. If there's more work in the session:
> "PR merged. I see there's still [X, Y] in the session notes - want to continue
> with those or archive what's done so far?"
## Important
- This skill **suggests** archiving, it does NOT archive automatically
- User must explicitly run `/bonfire:archive` to perform the archive
- Trigger AFTER merge succeeds, not before
- Multiple PRs may be part of one logical session - check context
- If `.bonfire/` doesn't exist, don't suggest archiving

View File

@@ -0,0 +1,51 @@
---
description: Read project session context from .bonfire/index.md to understand ongoing work, previous decisions, blockers, and history. Use when the user asks about project context, previous sessions, what was worked on before, architectural decisions, blockers, or when they reference "last time", "previously", "the session", or "what we decided".
allowed-tools: Read, Glob
---
# Bonfire Context
This project may use the Bonfire pattern to maintain continuity across AI coding sessions. Context is stored in `.bonfire/index.md` rather than relying on conversation memory.
## When to Use This Skill
Read session context when the user:
- Asks about previous work or decisions
- References "last time", "previously", "before"
- Wants to know about blockers or pending issues
- Asks what the project status is
- Starts a significant task that might have prior context
## Instructions
1. Find the git root: `git rev-parse --show-toplevel`
2. Check if `.bonfire/index.md` exists at the git root
3. If it exists, read it to understand:
- Current project status and recent work
- Active decisions and their rationale
- Known blockers or pending issues
- Links to relevant specs or documentation
4. Check `.bonfire/specs/` if the user asks about implementation specs
5. Check `.bonfire/docs/` if the user asks about documented topics
## File Structure
```
.bonfire/
├── index.md # Main session context (read this first)
├── config.json # Project settings
├── archive/ # Completed work history
├── docs/ # Topic documentation
└── specs/ # Implementation specs
```
## Important
- This skill is for **reading** context, not updating it
- Session updates happen via `/bonfire:end` command
- Don't modify `.bonfire/index.md` unless explicitly asked
- If `.bonfire/` doesn't exist, the project may not use this pattern

View File

@@ -1,6 +1,6 @@
{
"context7": {
"command": "npx",
"args": ["-y", "@upstash/context7-mcp"]
"type": "http",
"url": "https://mcp.context7.com/mcp"
}
}

View File

@@ -0,0 +1,72 @@
# Context7 Plugin for Claude Code
Context7 solves a common problem with AI coding assistants: outdated training data and hallucinated APIs. Instead of relying on stale knowledge, Context7 fetches current documentation directly from source repositories.
## What's Included
This plugin provides:
- **MCP Server** - Connects Claude Code to Context7's documentation service
- **Skills** - Auto-triggers documentation lookups when you ask about libraries
- **Agents** - A dedicated `docs-researcher` agent for focused lookups
- **Commands** - `/context7:docs` for manual documentation queries
## Installation
Install the plugin from the official marketplace:
```bash
claude plugin install context7@claude-plugins-official
```
## Available Tools
### resolve-library-id
Searches for libraries and returns Context7-compatible identifiers.
```
Input: "next.js"
Output: { id: "/vercel/next.js", name: "Next.js", versions: ["v15.1.8", "v14.2.0", ...] }
```
### query-docs
Fetches documentation for a specific library, ranked by relevance to your question.
```
Input: { libraryId: "/vercel/next.js", query: "app router middleware" }
Output: Relevant documentation snippets with code examples
```
## Usage Examples
The plugin works automatically when you ask about libraries:
- "How do I set up authentication in Next.js 15?"
- "Show me React Server Components examples"
- "What's the Prisma syntax for relations?"
For manual lookups, use the command:
```
/context7:docs next.js app router
/context7:docs /vercel/next.js/v15.1.8 middleware
```
Or spawn the docs-researcher agent when you want to keep your main context clean:
```
spawn docs-researcher to look up Supabase auth methods
```
## Version Pinning
To get documentation for a specific version, include the version in the library ID:
```
/vercel/next.js/v15.1.8
/supabase/supabase/v2.45.0
```
The `resolve-library-id` tool returns available versions, so you can pick the one that matches your project.

View File

@@ -0,0 +1,40 @@
---
name: docs-researcher
description: Lightweight agent for fetching library documentation without cluttering your main conversation context.
model: sonnet
---
You are a documentation researcher specializing in fetching up-to-date library and framework documentation from Context7.
## Your Task
When given a question about a library or framework, fetch the relevant documentation and return a concise, actionable answer with code examples.
## Process
1. **Identify the library**: Extract the library/framework name from the user's question.
2. **Resolve the library ID**: Call `resolve-library-id` with:
- `libraryName`: The library name (e.g., "react", "next.js", "prisma")
- `query`: The user's full question for relevance ranking
3. **Select the best match**: From the results, pick the library with:
- Exact or closest name match
- Highest benchmark score
- Appropriate version if the user specified one (e.g., "React 19" → look for v19.x)
4. **Fetch documentation**: Call `query-docs` with:
- `libraryId`: The selected Context7 library ID (e.g., `/vercel/next.js`)
- `query`: The user's specific question for targeted results
5. **Return a focused answer**: Summarize the relevant documentation with:
- Direct answer to the question
- Code examples from the docs
- Links or references if available
## Guidelines
- Pass the user's full question as the query parameter for better relevance
- When the user mentions a version (e.g., "Next.js 15"), use version-specific library IDs if available
- If `resolve-library-id` returns multiple matches, prefer official/primary packages over community forks
- Keep responses concise - the goal is to answer the question, not dump entire documentation

View File

@@ -0,0 +1,45 @@
---
description: Look up documentation for any library
argument-hint: <library> [query]
---
# /context7:docs
Fetches up-to-date documentation and code examples for a library.
## Usage
```
/context7:docs <library> [query]
```
- **library**: The library name, or a Context7 ID starting with `/`
- **query**: What you're looking for (optional but recommended)
## Examples
```
/context7:docs react hooks
/context7:docs next.js authentication
/context7:docs prisma relations
/context7:docs /vercel/next.js/v15.1.8 app router
/context7:docs /supabase/supabase row level security
```
## How It Works
1. If the library starts with `/`, it's used directly as the Context7 ID
2. Otherwise, `resolve-library-id` finds the best matching library
3. `query-docs` fetches documentation relevant to your query
4. Results include code examples and explanations
## Version-Specific Lookups
Include the version in the library ID for pinned documentation:
```
/context7:docs /vercel/next.js/v15.1.8 middleware
/context7:docs /facebook/react/v19.0.0 use hook
```
This is useful when you're working with a specific version and want docs that match exactly.

View File

@@ -0,0 +1,53 @@
---
name: documentation-lookup
description: This skill should be used when the user asks about libraries, frameworks, API references, or needs code examples. Activates for setup questions, code generation involving libraries, or mentions of specific frameworks like React, Vue, Next.js, Prisma, Supabase, etc.
---
When the user asks about libraries, frameworks, or needs code examples, use Context7 to fetch current documentation instead of relying on training data.
## When to Use This Skill
Activate this skill when the user:
- Asks setup or configuration questions ("How do I configure Next.js middleware?")
- Requests code involving libraries ("Write a Prisma query for...")
- Needs API references ("What are the Supabase auth methods?")
- Mentions specific frameworks (React, Vue, Svelte, Express, Tailwind, etc.)
## How to Fetch Documentation
### Step 1: Resolve the Library ID
Call `resolve-library-id` with:
- `libraryName`: The library name extracted from the user's question
- `query`: The user's full question (improves relevance ranking)
### Step 2: Select the Best Match
From the resolution results, choose based on:
- Exact or closest name match to what the user asked for
- Higher benchmark scores indicate better documentation quality
- If the user mentioned a version (e.g., "React 19"), prefer version-specific IDs
### Step 3: Fetch the Documentation
Call `query-docs` with:
- `libraryId`: The selected Context7 library ID (e.g., `/vercel/next.js`)
- `query`: The user's specific question
### Step 4: Use the Documentation
Incorporate the fetched documentation into your response:
- Answer the user's question using current, accurate information
- Include relevant code examples from the docs
- Cite the library version when relevant
## Guidelines
- **Be specific**: Pass the user's full question as the query for better results
- **Version awareness**: When users mention versions ("Next.js 15", "React 19"), use version-specific library IDs if available from the resolution step
- **Prefer official sources**: When multiple matches exist, prefer official/primary packages over community forks

View File

@@ -0,0 +1,13 @@
{
"name": "stagehand",
"description": "Browser automation skill for Claude Code using Stagehand. Automate web interactions, extract data, and navigate websites using natural language.",
"version": "0.1.0",
"author": {
"name": "Browserbase"
},
"homepage": "https://github.com/browserbase/agent-browse",
"repository": "https://github.com/browserbase/agent-browse",
"keywords": ["browser", "automation", "stagehand", "web-scraping"],
"strict": false
}

View File

@@ -0,0 +1,104 @@
# Stagehand Browser Automation Plugin
Browser automation skill for Claude Code using [Stagehand](https://github.com/browserbase/stagehand). This plugin enables Claude to automate web browser interactions, extract data, and navigate websites using natural language.
## Installation
Install the plugin from the Claude Code marketplace:
```bash
/plugin install stagehand@claude-plugin-directory
```
## Prerequisites
This plugin requires the browser automation CLI tools to be installed separately. The CLI tools are available from the GitHub marketplace.
### Step 1: Add the GitHub Marketplace
```bash
/plugin marketplace add browserbase/agent-browse
```
### Step 2: Install the Browser Automation CLI Plugin
```bash
/plugin install browser-automation@browser-tools
```
### Step 3: Set Up the CLI Tools
After installing the browser-automation plugin, you need to set up the CLI tools:
1. Navigate to the plugin directory (typically `~/.claude/plugins/browser-automation/`)
2. Install dependencies and build:
```bash
npm install
```
3. Link the browser command globally:
```bash
npm link
```
4. Configure your Anthropic API key:
```bash
export ANTHROPIC_API_KEY="your-api-key-here"
```
Or use Claude Code's subscription token (recommended if you have Claude Pro/Max):
```bash
claude setup-token
```
### Step 4: Verify Installation
Test that the browser command is available:
```bash
browser navigate https://example.com
```
## Usage
Once installed and configured, you can use natural language to automate browser tasks:
- *"Go to Hacker News, get the top post comments, and summarize them"*
- *"QA test http://localhost:3000 and fix any bugs you encounter"*
- *"Extract product information from example.com/products"*
Claude will automatically use the browser automation skill when you ask for web-related tasks.
## Features
- **Natural Language Control**: Describe browser actions in plain English
- **Data Extraction**: Extract structured data from web pages
- **Screenshot Capture**: Take screenshots for visual verification
- **Persistent Sessions**: Browser state persists between commands
- **Chrome Profile Integration**: Uses your Chrome profile for cookies and sessions
## Troubleshooting
### Chrome not found
Install Chrome for your platform:
- **macOS** or **Windows**: https://www.google.com/chrome/
- **Linux**: `sudo apt install google-chrome-stable`
### Browser command not found
Make sure you've run `npm link` in the browser-automation plugin directory after installing it.
### API Key Issues
- If you have Claude Pro/Max, use `claude setup-token` (recommended)
- Otherwise, export `ANTHROPIC_API_KEY` in your terminal
- Or create a `.env` file in the plugin directory with your API key
## Resources
- [Stagehand Documentation](https://github.com/browserbase/stagehand)
- [GitHub Marketplace](https://github.com/browserbase/agent-browse)
- [Claude Code Skills Documentation](https://code.claude.com/docs/en/plugins)
## Support
For issues or questions, please visit the [GitHub repository](https://github.com/browserbase/agent-browse).

View File

@@ -0,0 +1,7 @@
{
"name": "terraform",
"description": "The Terraform MCP Server provides seamless integration with Terraform ecosystem, enabling advanced automation and interaction capabilities for Infrastructure as Code (IaC) development.",
"author": {
"name": "HashiCorp"
}
}

View File

@@ -0,0 +1,12 @@
{
"terraform": {
"command": "docker",
"args": [
"run",
"-i",
"--rm",
"-e", "TFE_TOKEN=${TFE_TOKEN}",
"hashicorp/terraform-mcp-server:0.3.3"
]
}
}

View File

@@ -0,0 +1,8 @@
{
"name": "plugin-dev",
"description": "Plugin development toolkit with skills for creating agents, commands, hooks, MCP integrations, and comprehensive plugin structure guidance",
"author": {
"name": "Anthropic",
"email": "support@anthropic.com"
}
}

View File

@@ -169,6 +169,24 @@ Keep trying until success. The loop handles retry logic automatically.
- One $50k contract completed for $297 in API costs
- Created entire programming language ("cursed") over 3 months using this approach
## Windows Compatibility
The stop hook uses a bash script that requires Git for Windows to run properly.
**Issue**: On Windows, the `bash` command may resolve to WSL bash (often misconfigured) instead of Git Bash, causing the hook to fail with errors like:
- `wsl: Unknown key 'automount.crossDistro'`
- `execvpe(/bin/bash) failed: No such file or directory`
**Workaround**: Edit the cached plugin's `hooks/hooks.json` to use Git Bash explicitly:
```json
"command": "\"C:/Program Files/Git/bin/bash.exe\" ${CLAUDE_PLUGIN_ROOT}/hooks/stop-hook.sh"
```
**Location**: `~/.claude/plugins/cache/claude-plugins-official/ralph-wiggum/<hash>/hooks/hooks.json`
**Note**: Use `Git/bin/bash.exe` (the wrapper with proper PATH), not `Git/usr/bin/bash.exe` (raw MinGW bash without utilities in PATH).
## Learn More
- Original technique: https://ghuntley.com/ralph/

View File

@@ -110,7 +110,7 @@ HELP_EOF
done
# Join all prompt parts with spaces
PROMPT="${PROMPT_PARTS[*]}"
PROMPT="${PROMPT_PARTS[*]:-}"
# Validate prompt is non-empty
if [[ -z "$PROMPT" ]]; then

202
plugins/ruby-lsp/LICENSE Normal file
View File

@@ -0,0 +1,202 @@
Apache License
Version 2.0, January 2004
http://www.apache.org/licenses/
TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
1. Definitions.
"License" shall mean the terms and conditions for use, reproduction,
and distribution as defined by Sections 1 through 9 of this document.
"Licensor" shall mean the copyright owner or entity authorized by
the copyright owner that is granting the License.
"Legal Entity" shall mean the union of the acting entity and all
other entities that control, are controlled by, or are under common
control with that entity. For the purposes of this definition,
"control" means (i) the power, direct or indirect, to cause the
direction or management of such entity, whether by contract or
otherwise, or (ii) ownership of fifty percent (50%) or more of the
outstanding shares, or (iii) beneficial ownership of such entity.
"You" (or "Your") shall mean an individual or Legal Entity
exercising permissions granted by this License.
"Source" form shall mean the preferred form for making modifications,
including but not limited to software source code, documentation
source, and configuration files.
"Object" form shall mean any form resulting from mechanical
transformation or translation of a Source form, including but
not limited to compiled object code, generated documentation,
and conversions to other media types.
"Work" shall mean the work of authorship, whether in Source or
Object form, made available under the License, as indicated by a
copyright notice that is included in or attached to the work
(an example is provided in the Appendix below).
"Derivative Works" shall mean any work, whether in Source or Object
form, that is based on (or derived from) the Work and for which the
editorial revisions, annotations, elaborations, or other modifications
represent, as a whole, an original work of authorship. For the purposes
of this License, Derivative Works shall not include works that remain
separable from, or merely link (or bind by name) to the interfaces of,
the Work and Derivative Works thereof.
"Contribution" shall mean any work of authorship, including
the original version of the Work and any modifications or additions
to that Work or Derivative Works thereof, that is intentionally
submitted to Licensor for inclusion in the Work by the copyright owner
or by an individual or Legal Entity authorized to submit on behalf of
the copyright owner. For the purposes of this definition, "submitted"
means any form of electronic, verbal, or written communication sent
to the Licensor or its representatives, including but not limited to
communication on electronic mailing lists, source code control systems,
and issue tracking systems that are managed by, or on behalf of, the
Licensor for the purpose of discussing and improving the Work, but
excluding communication that is conspicuously marked or otherwise
designated in writing by the copyright owner as "Not a Contribution."
"Contributor" shall mean Licensor and any individual or Legal Entity
on behalf of whom a Contribution has been received by Licensor and
subsequently incorporated within the Work.
2. Grant of Copyright License. Subject to the terms and conditions of
this License, each Contributor hereby grants to You a perpetual,
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
copyright license to reproduce, prepare Derivative Works of,
publicly display, publicly perform, sublicense, and distribute the
Work and such Derivative Works in Source or Object form.
3. Grant of Patent License. Subject to the terms and conditions of
this License, each Contributor hereby grants to You a perpetual,
worldwide, non-exclusive, no-charge, royalty-free, irrevocable
(except as stated in this section) patent license to make, have made,
use, offer to sell, sell, import, and otherwise transfer the Work,
where such license applies only to those patent claims licensable
by such Contributor that are necessarily infringed by their
Contribution(s) alone or by combination of their Contribution(s)
with the Work to which such Contribution(s) was submitted. If You
institute patent litigation against any entity (including a
cross-claim or counterclaim in a lawsuit) alleging that the Work
or a Contribution incorporated within the Work constitutes direct
or contributory patent infringement, then any patent licenses
granted to You under this License for that Work shall terminate
as of the date such litigation is filed.
4. Redistribution. You may reproduce and distribute copies of the
Work or Derivative Works thereof in any medium, with or without
modifications, and in Source or Object form, provided that You
meet the following conditions:
(a) You must give any other recipients of the Work or
Derivative Works a copy of this License; and
(b) You must cause any modified files to carry prominent notices
stating that You changed the files; and
(c) You must retain, in the Source form of any Derivative Works
that You distribute, all copyright, patent, trademark, and
attribution notices from the Source form of the Work,
excluding those notices that do not pertain to any part of
the Derivative Works; and
(d) If the Work includes a "NOTICE" text file as part of its
distribution, then any Derivative Works that You distribute must
include a readable copy of the attribution notices contained
within such NOTICE file, excluding those notices that do not
pertain to any part of the Derivative Works, in at least one
of the following places: within a NOTICE text file distributed
as part of the Derivative Works; within the Source form or
documentation, if provided along with the Derivative Works; or,
within a display generated by the Derivative Works, if and
wherever such third-party notices normally appear. The contents
of the NOTICE file are for informational purposes only and
do not modify the License. You may add Your own attribution
notices within Derivative Works that You distribute, alongside
or as an addendum to the NOTICE text from the Work, provided
that such additional attribution notices cannot be construed
as modifying the License.
You may add Your own copyright statement to Your modifications and
may provide additional or different license terms and conditions
for use, reproduction, or distribution of Your modifications, or
for any such Derivative Works as a whole, provided Your use,
reproduction, and distribution of the Work otherwise complies with
the conditions stated in this License.
5. Submission of Contributions. Unless You explicitly state otherwise,
any Contribution intentionally submitted for inclusion in the Work
by You to the Licensor shall be under the terms and conditions of
this License, without any additional terms or conditions.
Notwithstanding the above, nothing herein shall supersede or modify
the terms of any separate license agreement you may have executed
with Licensor regarding such Contributions.
6. Trademarks. This License does not grant permission to use the trade
names, trademarks, service marks, or product names of the Licensor,
except as required for reasonable and customary use in describing the
origin of the Work and reproducing the content of the NOTICE file.
7. Disclaimer of Warranty. Unless required by applicable law or
agreed to in writing, Licensor provides the Work (and each
Contributor provides its Contributions) on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
implied, including, without limitation, any warranties or conditions
of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
PARTICULAR PURPOSE. You are solely responsible for determining the
appropriateness of using or redistributing the Work and assume any
risks associated with Your exercise of permissions under this License.
8. Limitation of Liability. In no event and under no legal theory,
whether in tort (including negligence), contract, or otherwise,
unless required by applicable law (such as deliberate and grossly
negligent acts) or agreed to in writing, shall any Contributor be
liable to You for damages, including any direct, indirect, special,
incidental, or consequential damages of any character arising as a
result of this License or out of the use or inability to use the
Work (including but not limited to damages for loss of goodwill,
work stoppage, computer failure or malfunction, or any and all
other commercial damages or losses), even if such Contributor
has been advised of the possibility of such damages.
9. Accepting Warranty or Additional Liability. While redistributing
the Work or Derivative Works thereof, You may choose to offer,
and charge a fee for, acceptance of support, warranty, indemnity,
or other liability obligations and/or rights consistent with this
License. However, in accepting such obligations, You may act only
on Your own behalf and on Your sole responsibility, not on behalf
of any other Contributor, and only if You agree to indemnify,
defend, and hold each Contributor harmless for any liability
incurred by, or claims asserted against, such Contributor by reason
of your accepting any such warranty or additional liability.
END OF TERMS AND CONDITIONS
APPENDIX: How to apply the Apache License to your work.
To apply the Apache License to your work, attach the following
boilerplate notice, with the fields enclosed by brackets "[]"
replaced with your own identifying information. (Don't include
the brackets!) The text should be enclosed in the appropriate
comment syntax for the file format. We also recommend that a
file or class name and description of purpose be included on the
same "printed page" as the copyright notice for easier
identification within third-party archives.
Copyright [yyyy] [name of copyright owner]
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.

View File

@@ -0,0 +1,31 @@
# ruby-lsp
Ruby language server for Claude Code, providing code intelligence and analysis.
## Supported Extensions
`.rb`, `.rake`, `.gemspec`, `.ru`, `.erb`
## Installation
### Via gem (recommended)
```bash
gem install ruby-lsp
```
### Via Bundler
Add to your Gemfile:
```ruby
gem 'ruby-lsp', group: :development
```
Then run:
```bash
bundle install
```
## Requirements
- Ruby 3.0 or later
## More Information
- [Ruby LSP Website](https://shopify.github.io/ruby-lsp/)
- [GitHub Repository](https://github.com/Shopify/ruby-lsp)