developer-tools

LLM Code Context

by cyberchitta

LLM Code Context boosts code reviews and documentation with smart file selection, code outlining, and multi-language sup

Streamlines code context sharing with LLMs by implementing smart file selection, code outlining, and multi-language support for efficient code reviews and documentation generation.

github stars

295

Multi-language supportRule-based intelligent file filteringClipboard integration for chat interfaces

best for

  • / Developers doing code reviews with AI assistants
  • / Sharing project context efficiently in LLM conversations
  • / AI agents needing focused codebase access
  • / Managing development session context

capabilities

  • / Find files modified since a specific timestamp
  • / Generate code outlines highlighting important sections
  • / Retrieve missing context files on demand
  • / Create custom file selection rules
  • / Extract excerpted content from codebases

what it does

Provides smart file selection and code context management for sharing relevant project files with LLMs, avoiding token limits while ensuring complete coverage.

about

LLM Code Context is a community-built MCP server published by cyberchitta that provides AI assistants with tools and capabilities via the Model Context Protocol. LLM Code Context boosts code reviews and documentation with smart file selection, code outlining, and multi-language sup It is categorized under developer tools. This server exposes 4 tools that AI clients can invoke during conversations and coding sessions.

how to install

You can install LLM Code Context in your AI client of choice. Use the install panel on this page to get one-click setup for Cursor, Claude Desktop, VS Code, and other MCP-compatible clients. This server runs locally on your machine via the stdio transport.

license

Apache-2.0

LLM Code Context is released under the Apache-2.0 license. This is a permissive open-source license, meaning you can freely use, modify, and distribute the software.

readme

LLM Context

License PyPI version Downloads

Smart context management for LLM development workflows. Share relevant project files instantly through intelligent selection and rule-based filtering.

The Problem

Getting the right context into LLM conversations is friction-heavy:

  • Manually finding and copying relevant files wastes time
  • Too much context hits token limits, too little misses important details
  • AI requests for additional files require manual fetching
  • Hard to track what changed during development sessions

The Solution

llm-context provides focused, task-specific project context through composable rules.

For humans using chat interfaces:

lc-select   # Smart file selection
lc-context  # Copy formatted context to clipboard
# Paste and work - AI can access additional files via MCP

For AI agents with CLI access:

lc-preview tmp-prm-auth    # Validate rule selects right files
lc-context tmp-prm-auth    # Get focused context for sub-agent

For AI agents in chat (MCP tools):

  • lc_outlines - Generate excerpted context from current rule
  • lc_preview - Validate rule effectiveness before use
  • lc_missing - Fetch specific files/implementations on demand

Note: This project was developed in collaboration with several Claude Sonnets (3.5, 3.6, 3.7, 4.0) and Groks (3, 4), using LLM Context itself to share code during development. All code is heavily human-curated by @restlessronin.

Installation

uv tool install "llm-context>=0.6.0"

Quick Start

Human Workflow (Clipboard)

# One-time setup
cd your-project
lc-init

# Daily usage
lc-select
lc-context
# Paste into your LLM chat

MCP Integration (Recommended)

Add to Claude Desktop config (~/Library/Application Support/Claude/claude_desktop_config.json):

{
  "mcpServers": {
    "llm-context": {
      "command": "uvx",
      "args": ["--from", "llm-context", "lc-mcp"]
    }
  }
}

Restart Claude Desktop. Now AI can access additional files during conversations without manual copying.

Agent Workflow (CLI)

AI agents with shell access use llm-context to create focused contexts:

# Agent explores codebase
lc-outlines

# Agent creates focused rule for specific task
# (via Skill or lc-rule-instructions)

# Agent validates rule
lc-preview tmp-prm-oauth-task

# Agent uses context for sub-task
lc-context tmp-prm-oauth-task

Agent Workflow (MCP)

AI agents in chat environments use MCP tools:

# Explore codebase structure
lc_outlines(root_path, rule_name)

# Validate rule effectiveness  
lc_preview(root_path, rule_name)

# Fetch specific files/implementations
lc_missing(root_path, param_type, data, timestamp)

Core Concepts

Rules: Task-Specific Context Descriptors

Rules are YAML+Markdown files that describe what context to provide for a task:

---
description: "Debug API authentication"
compose:
  filters: [lc/flt-no-files]
  excerpters: [lc/exc-base]
also-include:
  full-files: ["/src/auth/**", "/tests/auth/**"]
---
Focus on authentication system and related tests.

Five Rule Categories

  • Prompt Rules (prm-): Generate project contexts (e.g., lc/prm-developer)
  • Filter Rules (flt-): Control file inclusion (e.g., lc/flt-base, lc/flt-no-files)
  • Instruction Rules (ins-): Provide guidelines (e.g., lc/ins-developer)
  • Style Rules (sty-): Enforce coding standards (e.g., lc/sty-python)
  • Excerpt Rules (exc-): Configure content extraction (e.g., lc/exc-base)

Rule Composition

Build complex rules from simpler ones:

---
instructions: [lc/ins-developer, lc/sty-python]
compose:
  filters: [lc/flt-base, project-filters]
  excerpters: [lc/exc-base]
---

Essential Commands

CommandPurpose
lc-initInitialize project configuration
lc-selectSelect files based on current rule
lc-contextGenerate and copy context
lc-context -pInclude prompt instructions
lc-context -mFormat as separate message
lc-context -ntNo tools (manual workflow)
lc-set-rule <name>Switch active rule
lc-preview <rule>Validate rule selection and size
lc-outlinesGet code structure excerpts
lc-missingFetch files/implementations (manual MCP)

AI-Assisted Rule Creation

Let AI help create focused, task-specific rules. Two approaches depending on your environment:

Claude Skill (Interactive, Claude Desktop/Code)

How it works: Global skill guides you through creating rules interactively. Examines your codebase as needed using MCP tools.

Setup:

lc-init  # Installs skill to ~/.claude/skills/
# Restart Claude Desktop or Claude Code

Usage:

# 1. Share project context
lc-context  # Any rule - overview included

# 2. Paste into Claude, then ask:
# "Create a rule for refactoring authentication to JWT"
# "I need a rule to debug the payment processing"

Claude will:

  1. Use project overview already in context
  2. Examine specific files via lc-missing as needed
  3. Ask clarifying questions about scope
  4. Generate optimized rule (tmp-prm-<task>.md)
  5. Provide validation instructions

Skill documentation (progressively disclosed):

  • Skill.md - Quick workflow, decision patterns
  • PATTERNS.md - Common rule patterns
  • SYNTAX.md - Detailed reference
  • EXAMPLES.md - Complete walkthroughs
  • TROUBLESHOOTING.md - Problem solving

Instruction Rules (Works Anywhere)

How it works: Load comprehensive rule-creation documentation into context, work with any LLM.

Usage:

# 1. Load framework
lc-set-rule lc/prm-rule-create
lc-select
lc-context -nt

# 2. Paste into any LLM
# "I need a rule for adding OAuth integration"

# 3. LLM generates focused rule using framework

# 4. Use the new rule
lc-set-rule tmp-prm-oauth
lc-select
lc-context

Included documentation:

  • lc/ins-rule-intro - Introduction and overview
  • lc/ins-rule-framework - Complete decision framework

Comparison

AspectSkillInstruction Rules
SetupAutomatic with lc-initAlready available
InteractionInteractive, uses lc-missingStatic documentation
File examinationAutomatic via MCPManual or via AI
Best forClaude Desktop/CodeAny LLM, any environment
UpdatesAutomatic with version upgradesBuilt-in to rules

Both require sharing project context first. Both produce equivalent results.

Project Customization

Create Base Filters

cat > .llm-context/rules/flt-repo-base.md << 'EOF'
---
description: "Repository-specific exclusions"
compose:
  filters: [lc/flt-base]
gitignores:
  full-files: ["*.md", "/tests", "/node_modules"]
  excerpted-files: ["*.md", "/tests"]
---
EOF

Create Development Rule

cat > .llm-context/rules/prm-code.md << 'EOF'
---
description: "Main development rule"
instructions: [lc/ins-developer, lc/sty-python]
compose:
  filters: [flt-repo-base]
  excerpters: [lc/exc-base]
---
Additional project-specific guidelines and context.
EOF

lc-set-rule prm-code

Deployment Patterns

Choose format based on your LLM environment:

PatternCommandUse Case
System Messagelc-context -pAI Studio, etc.
Single User Messagelc-context -p -mGrok, etc.
Separate Messageslc-prompt + lc-context -mFlexible placement
Project Files (included)lc-contextClaude Projects, etc.
Project Files (searchable)lc-context -mForce into context

See Deployment Patterns for details.

Key Features

  • Intelligent Selection: Rules automatically include/exclude appropriate files
  • Context Validation: Preview size and selection before generation
  • Code Excerpting: Extract structure while reducing tokens (15+ languages)
  • MCP Integration: AI accesses additional files without manual intervention
  • Composable Rules: Build complex contexts from reusable patterns
  • AI-Assisted Creation: Interactive skill or documentation-based approaches
  • Agent-Friendly: CLI and MCP interfaces for autonomous operation

Common Workflows

Daily Development (Human)

lc-set-rule prm-code
lc-select
lc-context
# Paste into chat - AI accesses more files via MCP if needed

Focused Task (Human or Agent)

# Share project context first
lc-context

# Then create focused rule:
# Via Skill: "Create a rule for [task]"
# Via Instructions: lc-set-rule lc/prm-rule-create && lc-context -nt

# Validate and use
lc-preview tmp-prm-task
lc-context tmp-prm-task

Agent Context Provisioning (CLI)

# Agent validates rule effectiveness
lc-preview tmp-prm-refactor-auth

# Agent generates context for sub-agent
lc-context tmp-prm-refactor-auth > /tmp/context.md
# Sub-agent reads context and executes task

Agent Context Provisioning (MCP)

# Agent validates rule
preview

---

FAQ

What is the LLM Code Context MCP server?
LLM Code Context is a Model Context Protocol (MCP) server profile on explainx.ai. MCP lets AI hosts (e.g. Claude Desktop, Cursor) call tools and resources through a standard interface; this page summarizes categories, install hints, and community ratings.
How do MCP servers relate to agent skills?
Skills are reusable instruction packages (often SKILL.md); MCP servers expose live capabilities. Teams frequently combine both—skills for workflows, MCP for APIs and data. See explainx.ai/skills and explainx.ai/mcp-servers for parallel directories.
How are reviews shown for LLM Code Context?
This profile displays 10 aggregated ratings (sample rows for discoverability plus signed-in user reviews). Average score is about 4.5 out of 5—verify behavior in your own environment before production use.
MCP server reviews

Ratings

4.510 reviews
  • Shikha Mishra· Oct 10, 2024

    LLM Code Context is among the better-indexed MCP projects we tried; the explainx.ai summary tracks the official description.

  • Piyush G· Sep 9, 2024

    We evaluated LLM Code Context against two servers with overlapping tools; this profile had the clearer scope statement.

  • Chaitanya Patil· Aug 8, 2024

    Useful MCP listing: LLM Code Context is the kind of server we cite when onboarding engineers to host + tool permissions.

  • Sakshi Patil· Jul 7, 2024

    LLM Code Context reduced integration guesswork — categories and install configs on the listing matched the upstream repo.

  • Ganesh Mohane· Jun 6, 2024

    I recommend LLM Code Context for teams standardizing on MCP; the explainx.ai page compares cleanly with sibling servers.

  • Oshnikdeep· May 5, 2024

    Strong directory entry: LLM Code Context surfaces stars and publisher context so we could sanity-check maintenance before adopting.

  • Dhruvi Jain· Apr 4, 2024

    LLM Code Context has been reliable for tool-calling workflows; the MCP profile page is a good permalink for internal docs.

  • Rahul Santra· Mar 3, 2024

    According to our notes, LLM Code Context benefits from clear Model Context Protocol framing — fewer ambiguous “AI plugin” claims.

  • Pratham Ware· Feb 2, 2024

    We wired LLM Code Context into a staging workspace; the listing’s GitHub and npm pointers saved time versus hunting across READMEs.

  • Yash Thakker· Jan 1, 2024

    LLM Code Context is a well-scoped MCP server in the explainx.ai directory — install snippets and categories matched our Claude Code setup.