AI Code Review Model

Multi-language code review model optimized for automated code review in CI/CD pipelines.

Model Details

  • Base Model: Qwen/Qwen2.5-Coder-1.5B-Instruct
  • Training Method: LoRA fine-tuning with MLX
  • Format: GGUF (Q4_K_M quantization)
  • Purpose: Automated code review for CI/CD pipelines

Usage

Docker (Recommended)

docker pull ghcr.io/iq2i/ai-code-review:latest

# Review your codebase
docker run --rm -v $(pwd):/workspace ghcr.io/iq2i/ai-code-review:latest /workspace/src

llama.cpp

# Download the model
wget https://huggingface.co/iq2i/ai-code-review/resolve/main/model-Q4_K_M.gguf

# Run inference
./llama-cli -m model-Q4_K_M.gguf -p "Review this code: ..."

Python (llama-cpp-python)

from llama_cpp import Llama

llm = Llama(model_path="model-Q4_K_M.gguf")
output = llm("Review this code: ...", max_tokens=512)
print(output)

Output Format

The model outputs concise text-based code reviews:

**SQL injection vulnerability**

User input is concatenated directly into a raw SQL query without parameterization or escaping.

Impact: An attacker can execute arbitrary SQL commands, potentially dumping the entire database, deleting data, or escalating privileges. For example: keyword=' OR '1'='1' -- would return all products.

Suggestion:
Use parameter binding: DB::select("SELECT * FROM products WHERE name LIKE ?", ['%' . $keyword . '%']) or better, use Eloquent: Product::where('name', 'like', '%' . $keyword . '%')->get()

Training

  • Training examples: 100+ real-world code issues
  • Format: ChatML conversation format with concise reviews
  • Framework: MLX for Apple Silicon acceleration
  • Method: LoRA adapters (r=4, alpha=8)
  • Iterations: 625

For training details, see the GitHub repository.

Limitations

  • Should be used as a supplementary tool, not a replacement for human review
  • May not catch all edge cases or security vulnerabilities
  • Best results on common programming patterns and frameworks

License

Apache 2.0

Citation

@software{ai_code_review,
  title = {AI Code Review Model},
  author = {IQ2i Team},
  year = {2025},
  url = {https://github.com/iq2i/ai-code-review}
}
Downloads last month
93
GGUF
Model size
2B params
Architecture
qwen2
Hardware compatibility
Log In to view the estimation

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for iq2i/ai-code-review

Base model

Qwen/Qwen2.5-1.5B
Quantized
(83)
this model