Track how 350+ AI models mention your brand — manage projects, run scans, analyze results
Config is the same across clients — only the file and path differ.
{
"mcpServers": {
"io-github-serpstatglobal-llm-brand-monitor": {
"command": "<see-readme>",
"args": []
}
}
}Are you the author?
Add this badge to your README to show your security score and help users find safe servers.
Track how 350+ AI models mention your brand — manage projects, run scans, analyze results
No automated test available for this server. Check the GitHub README for setup instructions.
Five weighted categories — click any category to see the underlying evidence.
No known CVEs.
No package registry to scan.
This server is missing a description. Tools and install config are also missing.If you've used it, help the community.
Add informationBe the first to review
Have you used this server?
Share your experience — it helps other developers decide.
Sign in to write a review.
Others in ai-ml / marketing
Dynamic problem-solving through sequential thought chains
A Model Context Protocol (MCP) server and CLI that provides tools for agent use when working on iOS and macOS projects.
The official Python SDK for Model Context Protocol servers and clients
An open-source AI agent that brings the power of Gemini directly into your terminal.
MCP Security Weekly
Get CVE alerts and security updates for io.github.SerpstatGlobal/llm-brand-monitor and similar servers.
Start a conversation
Ask a question, share a tip, or report an issue.
Sign in to join the discussion.
MCP server for LLM Brand Monitor — a platform that tracks how AI models mention your brand.
Connect Claude, Cursor, Windsurf, or any MCP-compatible client to manage brand monitoring projects, run scans across 350+ LLMs, and analyze results — all through natural language.
Add to your claude_desktop_config.json:
{
"mcpServers": {
"lbm": {
"command": "npx",
"args": ["-y", "@serpstat/llm-brand-monitor-mcp"],
"env": {
"LBM_API_KEY": "lbm_your_key_here"
}
}
}
}
claude mcp add lbm-mcp -e LBM_API_KEY=lbm_your_key_here -- npx -y @serpstat/llm-brand-monitor-mcp
git clone https://github.com/SerpstatGlobal/llm-brand-monitor-mcp.git
cd llm-brand-monitor-mcp
npm install && npm run build
Then add to your MCP client config:
{
"mcpServers": {
"lbm": {
"command": "node",
"args": ["/path/to/llm-brand-monitor-mcp/dist/index.js"],
"env": {
"LBM_API_KEY": "lbm_your_key_here"
}
}
}
}
LBM_API_KEY=lbm_your_key_here npx @modelcontextprotocol/inspector node dist/index.js
lbm_)
17 tools across 4 categories:
| Tool | What it does |
|---|---|
lbm_list_projects | List all brand monitoring projects |
lbm_get_project | Get project details with prompts and models |
lbm_create_project | Create a new project |
lbm_update_project | Update project name, models, or monitoring settings |
lbm_archive_project | Archive a project |
lbm_add_prompts | Add monitoring prompts to a project |
lbm_delete_prompt | Remove a prompt from a project |
| Tool | What it does |
|---|---|
lbm_run_scan | Start a scan — sends prompts to LLMs and collects responses |
lbm_get_scan_status | Check scan progress |
lbm_list_scans | View scan history |
| Tool | What it does |
|---|---|
lbm_list_results | Browse monitoring results (brand mentions, status) |
lbm_get_transcript | Read the full LLM response for a specific result |
lbm_list_competitors | See which competitor brands LLMs mention |
lbm_list_links | See which URLs and domains LLMs cite |
lbm_get_history | Competitor mention trends over time |
| Tool | What it does |
|---|---|
lbm_list_models | List 350+ available LLM models |
lbm_get_usage | Check credit balance and usage stats |
You: "What brand monitoring projects do I have?"
Claude: → lbm_list_projects
You: "Run a scan on the Serpstat project"
Claude: → lbm_run_scan (asks you to confirm — scans spend credits)
→ lbm_get_scan_status (polls until complete)
You: "Show me the results — which models mentioned my brand?"
Claude: → lbm_list_results
You: "What did GPT-5 say exactly?"
Claude: → lbm_get_transcript
You: "Who are my competitors according to AI models?"
Claude: → lbm_list_competitors
All list tools return compact CSV by default instead of verbose JSON. This reduces token usage by 80–96%, k