Installation
Requirements
- Python 3.11+ (3.12 recommended)
- pip or uv package manager
- Git (for version-controlled projects)
Install from PyPI
pip install codexaThis gives you the full CLI (39 commands), tree-sitter parsing (11 languages), quality analysis, grep, explain, and all non-ML features.
ML Extras (Semantic Search & Vector Indexing)
For embedding-based semantic search, install with the [ml] extra:
pip install "codexa[ml]"This adds sentence-transformers, torch, and faiss-cpu. Requires ~2 GB RAM for model loading.
TIP
If you only need structural commands (explain, grep, quality, languages), the lightweight pip install codexa is sufficient. Install [ml] when you want codexa search with vector similarity.
Install from Source
git clone https://github.com/M9nx/CodexA.git
cd CodexA
pip install -e "."Install via Docker
docker build -t codexa .
docker run --rm -v /path/to/project:/workspace codexa search "auth"The Docker image includes ripgrep, git, and a pre-loaded embedding model.
Install via Homebrew (macOS)
brew install --formula Formula/codexa.rbOptional Extras
Install with optional dependencies for additional features:
pip install "codexa[ml]"
# Includes: sentence-transformers, torch, faiss-cpupip install -e ".[dev]"
# Includes: pytest, pytest-covpip install -e ".[tui]"
# Includes: textual (interactive terminal UI)pip install -e ".[dev,tui,ml]"Verify Installation
codexa --version
# codexa 0.4.4
codexa doctor
# Checks environment healthFirst-Time Setup
Initialize CodexA in your project:
cd /path/to/your-project
codexa initThis creates a .codexa/ directory with:
.codexa/
├── config.json # Configuration (embedding, search, LLM settings)
├── index/ # FAISS vector index (after indexing)
├── cache/ # Query and embedding caches
└── plugins/ # Custom plugins directoryLLM Configuration (Optional)
For AI-powered commands (ask, review, refactor, chat, etc.), configure an LLM provider in .codexa/config.json:
{
"llm": {
"provider": "openai",
"model": "gpt-4",
"api_key": "sk-...",
"temperature": 0.2,
"max_tokens": 2048
}
}Supported providers: openai, ollama (local), mock (testing).
Next Steps
- Quick Start — Index a project and start searching
- Configuration — Full configuration reference
- AI Agent Setup — Integrate with VS Code Copilot