Compare commits
2 Commits
main
...
issue-102-
Author | SHA1 | Date | |
---|---|---|---|
5595c0bbca | |||
f9a13896f9 |
|
@ -1,7 +1,3 @@
|
|||
# WARNING!
|
||||
# THIS IS AN AUTOGENERATED FILE!
|
||||
# MANUAL CHANGES CAN AND WILL BE OVERWRITTEN!
|
||||
|
||||
name: Build Python Container
|
||||
on:
|
||||
push:
|
||||
|
@ -10,72 +6,13 @@ on:
|
|||
paths-ignore: ['README.md', '.gitignore', 'LICENSE', 'CONVENTIONS.md', 'ruff.toml']
|
||||
|
||||
jobs:
|
||||
release-image:
|
||||
runs-on: ubuntu-latest
|
||||
container:
|
||||
image: catthehacker/ubuntu:act-latest
|
||||
env:
|
||||
RUNNER_TOOL_CACHE: /toolcache
|
||||
steps:
|
||||
- run: apt-get update
|
||||
- name: Checkout
|
||||
uses: actions/checkout@v3
|
||||
- name: Setting up SSH
|
||||
if: ${{ hashFiles('requirements_private.txt') != '' }}
|
||||
uses: https://github.com/shimataro/ssh-key-action@v2.5.1
|
||||
Package-Container:
|
||||
uses: jmaa/workflows/.gitea/workflows/container.yaml@v6.21
|
||||
with:
|
||||
key: ${{ secrets.PIPELINE_WORKER_SSH_KEY }}
|
||||
name: id_rsa
|
||||
known_hosts: ${{ secrets.PIPELINE_WORKER_KNOWN_HOSTS }}
|
||||
config: |
|
||||
Host gitfub
|
||||
HostName gitfub.space
|
||||
User ${{ secrets.PIPY_REPO_USER }}
|
||||
|
||||
- name: Download private dependencies
|
||||
if: ${{ hashFiles('requirements_private.txt') != '' }}
|
||||
shell: bash
|
||||
run: |
|
||||
set -e
|
||||
mkdir -p private_deps
|
||||
cd private_deps
|
||||
while IFS=$" " read -r -a dependency_spec
|
||||
do
|
||||
if test -n "${dependency_spec[1]}"
|
||||
then
|
||||
git clone -v --single-branch --no-tags "${dependency_spec[0]}" --branch "${dependency_spec[1]}"
|
||||
else
|
||||
git clone -v --single-branch --no-tags "${dependency_spec[0]}"
|
||||
fi
|
||||
done < ../requirements_private.txt
|
||||
|
||||
- name: Set up QEMU
|
||||
uses: docker/setup-qemu-action@v2
|
||||
|
||||
- name: Set up Docker BuildX
|
||||
uses: docker/setup-buildx-action@v2
|
||||
|
||||
- name: Login to Docker Registry
|
||||
uses: docker/login-action@v2
|
||||
with:
|
||||
registry: gitfub.space
|
||||
username: ${{ secrets.PIPY_REPO_USER }}
|
||||
password: ${{ secrets.PIPY_REPO_PASS }}
|
||||
|
||||
- name: Get Meta
|
||||
id: meta
|
||||
run: |
|
||||
echo REPO_NAME=$(echo ${GITHUB_REPOSITORY} | awk -F"/" '{print $2}') >> $GITHUB_OUTPUT
|
||||
echo REPO_VERSION=$(git describe --tags --always | sed 's/^v//') >> $GITHUB_OUTPUT
|
||||
|
||||
- name: Build and push
|
||||
uses: docker/build-push-action@v4
|
||||
with:
|
||||
context: .
|
||||
file: ./Dockerfile
|
||||
platforms: |
|
||||
linux/amd64
|
||||
push: true
|
||||
tags: |
|
||||
gitfub.space/jmaa/${{ steps.meta.outputs.REPO_NAME }}:${{ steps.meta.outputs.REPO_VERSION }}
|
||||
gitfub.space/jmaa/${{ steps.meta.outputs.REPO_NAME }}:latest
|
||||
REGISTRY_DOMAIN: gitfub.space
|
||||
REGISTRY_ORGANIZATION: jmaa
|
||||
secrets:
|
||||
DOCKER_USERNAME: ${{ secrets.PIPY_REPO_USER }}
|
||||
DOCKER_PASSWORD: ${{ secrets.PIPY_REPO_PASS }}
|
||||
PIPELINE_WORKER_SSH_KEY: ${{ secrets.PIPELINE_WORKER_SSH_KEY }}
|
||||
PIPELINE_WORKER_KNOWN_HOSTS: ${{ secrets.PIPELINE_WORKER_KNOWN_HOSTS }}
|
||||
|
|
|
@ -1,7 +1,3 @@
|
|||
# WARNING!
|
||||
# THIS IS AN AUTOGENERATED FILE!
|
||||
# MANUAL CHANGES CAN AND WILL BE OVERWRITTEN!
|
||||
|
||||
name: Package Python
|
||||
on:
|
||||
push:
|
||||
|
@ -10,24 +6,11 @@ on:
|
|||
paths-ignore: ['README.md', '.gitignore', 'LICENSE', 'CONVENTIONS.md', 'ruff.toml']
|
||||
|
||||
jobs:
|
||||
Package-Python-And-Publish:
|
||||
runs-on: ubuntu-latest
|
||||
container:
|
||||
image: node:21-bookworm
|
||||
steps:
|
||||
- name: Setting up Python ${{ env.PYTHON_VERSION }} for ${{runner.arch}} ${{runner.os}}
|
||||
run: |
|
||||
apt-get update
|
||||
apt-get install -y python3 python3-pip
|
||||
- name: Check out repository code
|
||||
if: success()
|
||||
uses: actions/checkout@v3
|
||||
- name: Installing Python Dependencies
|
||||
if: success()
|
||||
run: python3 -m pip install --upgrade pip setuptools wheel build twine pytest --break-system-packages
|
||||
- name: Build
|
||||
if: success()
|
||||
run: python3 -m build
|
||||
- name: Publish
|
||||
if: success()
|
||||
run: python3 -m twine upload --repository-url "https://gitfub.space/api/packages/jmaa/pypi" -u ${{ secrets.PIPY_REPO_USER }} -p ${{ secrets.PIPY_REPO_PASS }} dist/*
|
||||
Package:
|
||||
uses: jmaa/workflows/.gitea/workflows/python-package.yaml@v6.21
|
||||
with:
|
||||
REGISTRY_DOMAIN: gitfub.space
|
||||
REGISTRY_ORGANIZATION: jmaa
|
||||
secrets:
|
||||
PIPY_REPO_USER: ${{ secrets.PIPY_REPO_USER }}
|
||||
PIPY_REPO_PASS: ${{ secrets.PIPY_REPO_PASS }}
|
||||
|
|
|
@ -1,7 +1,3 @@
|
|||
# WARNING!
|
||||
# THIS IS AN AUTOGENERATED FILE!
|
||||
# MANUAL CHANGES CAN AND WILL BE OVERWRITTEN!
|
||||
|
||||
name: Run Python tests (through Pytest)
|
||||
|
||||
on:
|
||||
|
|
|
@ -1,7 +1,3 @@
|
|||
# WARNING!
|
||||
# THIS IS AN AUTOGENERATED FILE!
|
||||
# MANUAL CHANGES CAN AND WILL BE OVERWRITTEN!
|
||||
|
||||
name: Verify Python project can be installed, loaded and have version checked
|
||||
|
||||
on:
|
||||
|
|
|
@ -1,7 +1,3 @@
|
|||
<!-- WARNING! -->
|
||||
<!-- THIS IS AN AUTOGENERATED FILE! -->
|
||||
<!-- MANUAL CHANGES CAN AND WILL BE OVERWRITTEN! -->
|
||||
|
||||
# Conventions
|
||||
|
||||
When contributing code to this project, you MUST follow the requirements
|
||||
|
|
101
README.md
101
README.md
|
@ -1,29 +1,23 @@
|
|||
<!-- WARNING! -->
|
||||
<!-- THIS IS AN AUTOGENERATED FILE! -->
|
||||
<!-- MANUAL CHANGES CAN AND WILL BE OVERWRITTEN! -->
|
||||
<!--- WARNING --->
|
||||
<!--- THIS IS AN AUTO-GENERATED FILE --->
|
||||
<!--- MANUAL CHANGES CAN AND WILL BE OVERWRITTEN --->
|
||||
|
||||
|
||||
|
||||
# Aider Gitea
|
||||
|
||||

|
||||
|
||||
A code automation tool that integrates Gitea with AI assistants to automatically solve issues.
|
||||
A code automation tool that integrates Gitea with Aider to automatically solve issues.
|
||||
|
||||
This program monitors your [Gitea](https://about.gitea.com/) repository for issues with the 'aider' label.
|
||||
When such an issue is found, it:
|
||||
|
||||
1. Creates a new branch.
|
||||
2. Invokes an AI assistant (Aider or Claude Code) to solve the issue using a Large-Language Model.
|
||||
2. Invokes [Aider](https://aider.chat/) to solve the issue using a Large-Language Model.
|
||||
3. Runs tests and code quality checks.
|
||||
4. Creates a pull request with the solution.
|
||||
|
||||
The tool automatically selects the appropriate AI assistant based on the specified model:
|
||||
- **Aider**: Used for non-Anthropic models (e.g., GPT, Ollama, Gemini)
|
||||
- **Claude Code**: Used for Anthropic models (e.g., Claude, Sonnet, Haiku, Opus)
|
||||
|
||||
Inspired by [the AI workflows](https://github.com/oscoreio/ai-workflows/)
|
||||
project.
|
||||
|
||||
## Usage
|
||||
|
||||
An application token must be supplied for the `gitea_token` secret. This must
|
||||
|
@ -36,109 +30,48 @@ have the following permissions:
|
|||
### Command Line
|
||||
|
||||
```bash
|
||||
# Run with default settings (uses Aider)
|
||||
python -m aider_gitea --aider-model gpt-4
|
||||
|
||||
# Use Claude Code with Anthropic models
|
||||
python -m aider_gitea --aider-model claude-3-sonnet
|
||||
python -m aider_gitea --aider-model claude-3-haiku
|
||||
python -m aider_gitea --aider-model anthropic/claude-3-opus
|
||||
|
||||
# Use Aider with various models
|
||||
python -m aider_gitea --aider-model gpt-4
|
||||
python -m aider_gitea --aider-model ollama/llama3
|
||||
python -m aider_gitea --aider-model gemini-pro
|
||||
# Run with default settings
|
||||
python -m aider_gitea
|
||||
|
||||
# Specify custom repository and owner
|
||||
python -m aider_gitea --owner myorg --repo myproject --aider-model claude-3-sonnet
|
||||
python -m aider_gitea --owner myorg --repo myproject
|
||||
|
||||
# Use a custom Gitea URL
|
||||
python -m aider_gitea --gitea-url https://gitea.example.com --aider-model gpt-4
|
||||
python -m aider_gitea --gitea-url https://gitea.example.com
|
||||
|
||||
# Specify a different base branch
|
||||
python -m aider_gitea --base-branch develop --aider-model claude-3-haiku
|
||||
python -m aider_gitea --base-branch develop
|
||||
```
|
||||
|
||||
### AI Assistant Selection
|
||||
|
||||
The tool automatically routes to the appropriate AI assistant based on the model name:
|
||||
|
||||
**Claude Code Integration (Anthropic Models):**
|
||||
- Model names containing: `claude`, `anthropic`, `sonnet`, `haiku`, `opus`
|
||||
- Examples: `claude-3-sonnet`, `claude-3-haiku`, `anthropic/claude-3-opus`
|
||||
- Requires: `ANTHROPIC_API_KEY` environment variable
|
||||
|
||||
**Aider Integration (All Other Models):**
|
||||
- Any model not matching Anthropic patterns
|
||||
- Examples: `gpt-4`, `ollama/llama3`, `gemini-pro`, `mistral-7b`
|
||||
- Requires: `LLM_API_KEY` environment variable
|
||||
|
||||
### Python API
|
||||
|
||||
```python
|
||||
from aider_gitea import solve_issue_in_repository, create_code_solver
|
||||
from aider_gitea import solve_issue_in_repository
|
||||
from pathlib import Path
|
||||
import argparse
|
||||
|
||||
# Solve an issue programmatically with automatic AI assistant selection
|
||||
repository_config = RepositoryConfig(
|
||||
# Solve an issue programmatically
|
||||
args = argparse.Namespace(
|
||||
gitea_url="https://gitea.example.com",
|
||||
owner="myorg",
|
||||
repo="myproject",
|
||||
base_branch="main"
|
||||
)
|
||||
|
||||
# Set the model to control which AI assistant is used
|
||||
import aider_gitea
|
||||
aider_gitea.CODE_MODEL = "claude-3-sonnet" # Will use Claude Code
|
||||
# aider_gitea.CODE_MODEL = "gpt-4" # Will use Aider
|
||||
|
||||
code_solver = create_code_solver() # Automatically selects based on model
|
||||
|
||||
solve_issue_in_repository(
|
||||
repository_config,
|
||||
args,
|
||||
Path("/path/to/repo"),
|
||||
"issue-123-fix-bug",
|
||||
"Fix critical bug",
|
||||
"The application crashes when processing large files",
|
||||
"123",
|
||||
gitea_client,
|
||||
code_solver
|
||||
"123"
|
||||
)
|
||||
```
|
||||
|
||||
### Environment Configuration
|
||||
|
||||
The tool uses environment variables for sensitive information:
|
||||
|
||||
**Required for all setups:**
|
||||
- `GITEA_TOKEN`: Your Gitea API token
|
||||
|
||||
**For Aider (non-Anthropic models):**
|
||||
- `LLM_API_KEY`: API key for the language model (OpenAI, Ollama, etc.)
|
||||
|
||||
**For Claude Code (Anthropic models):**
|
||||
- `ANTHROPIC_API_KEY`: Your Anthropic API key for Claude models
|
||||
|
||||
### Model Examples
|
||||
|
||||
**Anthropic Models (→ Claude Code):**
|
||||
```bash
|
||||
--aider-model claude-3-sonnet
|
||||
--aider-model claude-3-haiku
|
||||
--aider-model claude-3-opus
|
||||
--aider-model anthropic/claude-3-sonnet
|
||||
```
|
||||
|
||||
**Non-Anthropic Models (→ Aider):**
|
||||
```bash
|
||||
--aider-model gpt-4
|
||||
--aider-model gpt-3.5-turbo
|
||||
--aider-model ollama/llama3
|
||||
--aider-model ollama/codellama
|
||||
--aider-model gemini-pro
|
||||
--aider-model mistral-7b
|
||||
```
|
||||
- `LLM_API_KEY`: API key for the language model used by Aider
|
||||
```
|
||||
|
||||
## Dependencies
|
||||
|
|
|
@ -1,19 +1,15 @@
|
|||
"""Aider Gitea.
|
||||
|
||||
A code automation tool that integrates Gitea with AI assistants to automatically solve issues.
|
||||
A code automation tool that integrates Gitea with Aider to automatically solve issues.
|
||||
|
||||
This program monitors your [Gitea](https://about.gitea.com/) repository for issues with the 'aider' label.
|
||||
When such an issue is found, it:
|
||||
|
||||
1. Creates a new branch.
|
||||
2. Invokes an AI assistant (Aider or Claude Code) to solve the issue using a Large-Language Model.
|
||||
2. Invokes [Aider](https://aider.chat/) to solve the issue using a Large-Language Model.
|
||||
3. Runs tests and code quality checks.
|
||||
4. Creates a pull request with the solution.
|
||||
|
||||
The tool automatically selects the appropriate AI assistant based on the specified model:
|
||||
- **Aider**: Used for non-Anthropic models (e.g., GPT, Ollama, Gemini)
|
||||
- **Claude Code**: Used for Anthropic models (e.g., Claude, Sonnet, Haiku, Opus)
|
||||
|
||||
Inspired by [the AI workflows](https://github.com/oscoreio/ai-workflows/)
|
||||
project.
|
||||
|
||||
|
@ -29,109 +25,48 @@ have the following permissions:
|
|||
### Command Line
|
||||
|
||||
```bash
|
||||
# Run with default settings (uses Aider)
|
||||
python -m aider_gitea --aider-model gpt-4
|
||||
|
||||
# Use Claude Code with Anthropic models
|
||||
python -m aider_gitea --aider-model claude-3-sonnet
|
||||
python -m aider_gitea --aider-model claude-3-haiku
|
||||
python -m aider_gitea --aider-model anthropic/claude-3-opus
|
||||
|
||||
# Use Aider with various models
|
||||
python -m aider_gitea --aider-model gpt-4
|
||||
python -m aider_gitea --aider-model ollama/llama3
|
||||
python -m aider_gitea --aider-model gemini-pro
|
||||
# Run with default settings
|
||||
python -m aider_gitea
|
||||
|
||||
# Specify custom repository and owner
|
||||
python -m aider_gitea --owner myorg --repo myproject --aider-model claude-3-sonnet
|
||||
python -m aider_gitea --owner myorg --repo myproject
|
||||
|
||||
# Use a custom Gitea URL
|
||||
python -m aider_gitea --gitea-url https://gitea.example.com --aider-model gpt-4
|
||||
python -m aider_gitea --gitea-url https://gitea.example.com
|
||||
|
||||
# Specify a different base branch
|
||||
python -m aider_gitea --base-branch develop --aider-model claude-3-haiku
|
||||
python -m aider_gitea --base-branch develop
|
||||
```
|
||||
|
||||
### AI Assistant Selection
|
||||
|
||||
The tool automatically routes to the appropriate AI assistant based on the model name:
|
||||
|
||||
**Claude Code Integration (Anthropic Models):**
|
||||
- Model names containing: `claude`, `anthropic`, `sonnet`, `haiku`, `opus`
|
||||
- Examples: `claude-3-sonnet`, `claude-3-haiku`, `anthropic/claude-3-opus`
|
||||
- Requires: `ANTHROPIC_API_KEY` environment variable
|
||||
|
||||
**Aider Integration (All Other Models):**
|
||||
- Any model not matching Anthropic patterns
|
||||
- Examples: `gpt-4`, `ollama/llama3`, `gemini-pro`, `mistral-7b`
|
||||
- Requires: `LLM_API_KEY` environment variable
|
||||
|
||||
### Python API
|
||||
|
||||
```python
|
||||
from aider_gitea import solve_issue_in_repository, create_code_solver
|
||||
from aider_gitea import solve_issue_in_repository
|
||||
from pathlib import Path
|
||||
import argparse
|
||||
|
||||
# Solve an issue programmatically with automatic AI assistant selection
|
||||
repository_config = RepositoryConfig(
|
||||
# Solve an issue programmatically
|
||||
args = argparse.Namespace(
|
||||
gitea_url="https://gitea.example.com",
|
||||
owner="myorg",
|
||||
repo="myproject",
|
||||
base_branch="main"
|
||||
)
|
||||
|
||||
# Set the model to control which AI assistant is used
|
||||
import aider_gitea
|
||||
aider_gitea.CODE_MODEL = "claude-3-sonnet" # Will use Claude Code
|
||||
# aider_gitea.CODE_MODEL = "gpt-4" # Will use Aider
|
||||
|
||||
code_solver = create_code_solver() # Automatically selects based on model
|
||||
|
||||
solve_issue_in_repository(
|
||||
repository_config,
|
||||
args,
|
||||
Path("/path/to/repo"),
|
||||
"issue-123-fix-bug",
|
||||
"Fix critical bug",
|
||||
"The application crashes when processing large files",
|
||||
"123",
|
||||
gitea_client,
|
||||
code_solver
|
||||
"123"
|
||||
)
|
||||
```
|
||||
|
||||
### Environment Configuration
|
||||
|
||||
The tool uses environment variables for sensitive information:
|
||||
|
||||
**Required for all setups:**
|
||||
- `GITEA_TOKEN`: Your Gitea API token
|
||||
|
||||
**For Aider (non-Anthropic models):**
|
||||
- `LLM_API_KEY`: API key for the language model (OpenAI, Ollama, etc.)
|
||||
|
||||
**For Claude Code (Anthropic models):**
|
||||
- `ANTHROPIC_API_KEY`: Your Anthropic API key for Claude models
|
||||
|
||||
### Model Examples
|
||||
|
||||
**Anthropic Models (→ Claude Code):**
|
||||
```bash
|
||||
--aider-model claude-3-sonnet
|
||||
--aider-model claude-3-haiku
|
||||
--aider-model claude-3-opus
|
||||
--aider-model anthropic/claude-3-sonnet
|
||||
```
|
||||
|
||||
**Non-Anthropic Models (→ Aider):**
|
||||
```bash
|
||||
--aider-model gpt-4
|
||||
--aider-model gpt-3.5-turbo
|
||||
--aider-model ollama/llama3
|
||||
--aider-model ollama/codellama
|
||||
--aider-model gemini-pro
|
||||
--aider-model mistral-7b
|
||||
```
|
||||
- `LLM_API_KEY`: API key for the language model used by Aider
|
||||
```
|
||||
"""
|
||||
|
||||
|
@ -168,11 +103,7 @@ class RepositoryConfig:
|
|||
class IssueResolution:
|
||||
success: bool
|
||||
pull_request_url: str | None = None
|
||||
pull_request_id: int | None = None
|
||||
|
||||
def __post_init__(self):
|
||||
assert self.pull_request_id is None or isinstance(self.pull_request_id, int)
|
||||
assert self.pull_request_url is None or isinstance(self.pull_request_url, str)
|
||||
pull_request_id: str | None = None
|
||||
|
||||
|
||||
def generate_branch_name(issue_number: str, issue_title: str) -> str:
|
||||
|
@ -199,21 +130,17 @@ def bash_cmd(*commands: str) -> str:
|
|||
|
||||
|
||||
AIDER_TEST = bash_cmd(
|
||||
'echo "Setting up virtual environment"',
|
||||
'virtualenv venv',
|
||||
'echo "Activating virtual environment"',
|
||||
'source venv/bin/activate',
|
||||
'echo "Installing package"',
|
||||
'pip install -e .',
|
||||
'echo "Testing package"',
|
||||
'pytest test',
|
||||
)
|
||||
|
||||
RUFF_FORMAT_AND_AUTO_FIX = bash_cmd(
|
||||
'ruff format --silent',
|
||||
'ruff check --fix --ignore RUF022 --ignore PGH004 --silent',
|
||||
'ruff format --silent',
|
||||
'ruff check --fix --ignore RUF022 --ignore PGH004 --silent',
|
||||
'ruff format',
|
||||
'ruff check --fix --ignore RUF022 --ignore PGH004',
|
||||
'ruff format',
|
||||
'ruff check --fix --ignore RUF022 --ignore PGH004',
|
||||
)
|
||||
|
||||
AIDER_LINT = bash_cmd(
|
||||
|
@ -223,80 +150,23 @@ AIDER_LINT = bash_cmd(
|
|||
)
|
||||
|
||||
|
||||
LLM_MESSAGE_FORMAT = """{issue}
|
||||
|
||||
Go ahead with the changes you deem appropriate without waiting for explicit approval.
|
||||
|
||||
Do not draft changes beforehand; produce changes only once prompted for a specific file.
|
||||
"""
|
||||
|
||||
CLAUDE_CODE_MESSAGE_FORMAT = """{issue}
|
||||
|
||||
Please fix this issue by making the necessary code changes. Follow these guidelines:
|
||||
1. Run tests after making changes to ensure they pass
|
||||
2. Follow existing code style and conventions
|
||||
3. Make minimal, focused changes to address the issue
|
||||
4. Commit your changes with a descriptive message
|
||||
|
||||
The test command for this project is: {test_command}
|
||||
The lint command for this project is: {lint_command}
|
||||
"""
|
||||
|
||||
CODE_MODEL = None
|
||||
EVALUATOR_MODEL = None
|
||||
|
||||
MODEL_EDIT_MODES = {
|
||||
'ollama/qwen3:32b': 'diff',
|
||||
'ollama/hf.co/unsloth/Qwen3-30B-A3B-GGUF:Q4_K_M': 'diff',
|
||||
}
|
||||
|
||||
|
||||
def run_post_solver_cleanup(repository_path: Path, solver_name: str) -> None:
|
||||
"""Run standard code quality fixes and commit changes after a code solver.
|
||||
|
||||
Args:
|
||||
repository_path: Path to the repository
|
||||
solver_name: Name of the solver (for commit message)
|
||||
"""
|
||||
# Auto-fix standard code quality stuff
|
||||
run_cmd(['bash', '-c', RUFF_FORMAT_AND_AUTO_FIX], repository_path, check=False)
|
||||
run_cmd(['git', 'add', '.'], repository_path)
|
||||
run_cmd(
|
||||
['git', 'commit', '-m', f'Ruff after {solver_name}'],
|
||||
repository_path,
|
||||
check=False,
|
||||
LLM_MESSAGE_FORMAT = (
|
||||
"""{issue}\nDo not wait for explicit approval before working on code changes."""
|
||||
)
|
||||
|
||||
|
||||
@dataclasses.dataclass(frozen=True)
|
||||
class CodeSolverStrategy:
|
||||
"""Base interface for code solving strategies."""
|
||||
|
||||
def solve_issue_round(self, repository_path: Path, issue_content: str) -> bool:
|
||||
"""Attempt to solve an issue in a single round.
|
||||
|
||||
Args:
|
||||
repository_path: Path to the repository
|
||||
issue_content: The issue description to solve
|
||||
|
||||
Returns:
|
||||
True if the solution round completed without crashing, False otherwise
|
||||
"""
|
||||
raise NotImplementedError
|
||||
# CODE_MODEL = 'ollama/gemma3:4b'
|
||||
CODE_MODEL = 'o4-mini'
|
||||
EVALUATOR_MODEL = 'ollama/gemma3:27b'
|
||||
|
||||
|
||||
@dataclasses.dataclass(frozen=True)
|
||||
class AiderCodeSolver(CodeSolverStrategy):
|
||||
"""Code solver that uses Aider for issue resolution."""
|
||||
|
||||
def _create_aider_command(self, issue: str) -> list[str]:
|
||||
"""Create the Aider command with all necessary flags."""
|
||||
def create_aider_command(issue: str) -> list[str]:
|
||||
l = [
|
||||
'aider',
|
||||
'--chat-language',
|
||||
'english',
|
||||
'--no-stream',
|
||||
'--no-analytics',
|
||||
#'--no-check-update',
|
||||
'--test-cmd',
|
||||
AIDER_TEST,
|
||||
'--lint-cmd',
|
||||
|
@ -304,16 +174,8 @@ class AiderCodeSolver(CodeSolverStrategy):
|
|||
'--auto-test',
|
||||
'--no-auto-lint',
|
||||
'--yes',
|
||||
'--disable-playwright',
|
||||
'--timeout',
|
||||
str(10_000),
|
||||
]
|
||||
|
||||
if edit_format := MODEL_EDIT_MODES.get(CODE_MODEL):
|
||||
l.append('--edit-format')
|
||||
l.append(edit_format)
|
||||
del edit_format
|
||||
|
||||
for key in secrets.llm_api_keys():
|
||||
l += ['--api-key', key]
|
||||
|
||||
|
@ -331,7 +193,7 @@ class AiderCodeSolver(CodeSolverStrategy):
|
|||
l.append('--model')
|
||||
l.append(CODE_MODEL)
|
||||
|
||||
if CODE_MODEL.startswith('ollama/') and False:
|
||||
if CODE_MODEL.startswith('ollama/'):
|
||||
l.append('--auto-lint')
|
||||
|
||||
if True:
|
||||
|
@ -340,96 +202,6 @@ class AiderCodeSolver(CodeSolverStrategy):
|
|||
|
||||
return l
|
||||
|
||||
def solve_issue_round(self, repository_path: Path, issue_content: str) -> bool:
|
||||
"""Solve an issue using Aider."""
|
||||
# Primary Aider command
|
||||
aider_command = self._create_aider_command(issue_content)
|
||||
aider_did_not_crash = run_cmd(
|
||||
aider_command,
|
||||
cwd=repository_path,
|
||||
check=False,
|
||||
)
|
||||
if not aider_did_not_crash:
|
||||
return aider_did_not_crash
|
||||
|
||||
# Run post-solver cleanup
|
||||
run_post_solver_cleanup(repository_path, 'aider')
|
||||
|
||||
return True
|
||||
|
||||
|
||||
@dataclasses.dataclass(frozen=True)
|
||||
class ClaudeCodeSolver(CodeSolverStrategy):
|
||||
"""Code solver that uses Claude Code for issue resolution."""
|
||||
|
||||
def _create_claude_command(self, issue: str) -> list[str]:
|
||||
"""Create the Claude Code command for programmatic use."""
|
||||
cmd = [
|
||||
'claude',
|
||||
'-p',
|
||||
'--output-format',
|
||||
'stream-json',
|
||||
#'--max-turns', '100',
|
||||
'--debug',
|
||||
'--verbose',
|
||||
'--dangerously-skip-permissions',
|
||||
]
|
||||
|
||||
if CODE_MODEL:
|
||||
cmd.extend(['--model', CODE_MODEL])
|
||||
|
||||
cmd.append(issue)
|
||||
return cmd
|
||||
|
||||
def solve_issue_round(self, repository_path: Path, issue_content: str) -> bool:
|
||||
"""Solve an issue using Claude Code."""
|
||||
# Prepare the issue prompt for Claude Code
|
||||
enhanced_issue = CLAUDE_CODE_MESSAGE_FORMAT.format(
|
||||
issue=issue_content,
|
||||
test_command=AIDER_TEST,
|
||||
lint_command=AIDER_LINT,
|
||||
)
|
||||
|
||||
# Create Claude Code command
|
||||
claude_command = self._create_claude_command(enhanced_issue)
|
||||
|
||||
# Run Claude Code
|
||||
run_cmd(
|
||||
claude_command,
|
||||
cwd=repository_path,
|
||||
check=False,
|
||||
)
|
||||
|
||||
# Run post-solver cleanup
|
||||
run_post_solver_cleanup(repository_path, 'Claude Code')
|
||||
|
||||
return True
|
||||
|
||||
|
||||
def is_anthropic_model(model: str) -> bool:
|
||||
"""Check if the model string indicates an Anthropic/Claude model."""
|
||||
if not model:
|
||||
return False
|
||||
|
||||
anthropic_indicators = [
|
||||
'claude',
|
||||
'anthropic',
|
||||
'sonnet',
|
||||
'haiku',
|
||||
'opus',
|
||||
]
|
||||
|
||||
model_lower = model.lower()
|
||||
return any(indicator in model_lower for indicator in anthropic_indicators)
|
||||
|
||||
|
||||
def create_code_solver() -> CodeSolverStrategy:
|
||||
"""Create the appropriate code solver based on the configured model."""
|
||||
if is_anthropic_model(CODE_MODEL):
|
||||
return ClaudeCodeSolver()
|
||||
else:
|
||||
return AiderCodeSolver()
|
||||
|
||||
|
||||
def get_commit_messages(cwd: Path, base_branch: str, current_branch: str) -> list[str]:
|
||||
"""Get commit messages between base branch and current branch.
|
||||
|
@ -511,8 +283,8 @@ def push_changes(
|
|||
# Extract PR number and URL if available
|
||||
return IssueResolution(
|
||||
True,
|
||||
str(pr_response.get('number')),
|
||||
pr_response.get('html_url'),
|
||||
int(pr_response.get('number')),
|
||||
)
|
||||
|
||||
|
||||
|
@ -550,19 +322,29 @@ def run_cmd(cmd: list[str], cwd: Path | None = None, check=True) -> bool:
|
|||
return result.returncode == 0
|
||||
|
||||
|
||||
def remove_thinking_tokens(text: str) -> str:
|
||||
text = re.sub(r'^\s*<think>.*?</think>', '', text, flags=re.MULTILINE | re.DOTALL)
|
||||
text = text.strip()
|
||||
return text
|
||||
def issue_solution_round(repository_path, issue_content):
|
||||
# Primary Aider command
|
||||
aider_command = create_aider_command(issue_content)
|
||||
print(aider_command)
|
||||
aider_did_not_crash = run_cmd(
|
||||
aider_command,
|
||||
repository_path,
|
||||
check=False,
|
||||
)
|
||||
if not aider_did_not_crash:
|
||||
return aider_did_not_crash
|
||||
|
||||
# Auto-fix standard code quality stuff after aider
|
||||
run_cmd(['bash', '-c', RUFF_FORMAT_AND_AUTO_FIX], repository_path, check=False)
|
||||
run_cmd(['git', 'add', '.'], repository_path)
|
||||
run_cmd(['git', 'commit', '-m', 'Ruff after aider'], repository_path, check=False)
|
||||
|
||||
assert remove_thinking_tokens('<think>Hello</think>\nWorld\n') == 'World'
|
||||
assert remove_thinking_tokens('<think>\nHello\n</think>\nWorld\n') == 'World'
|
||||
assert remove_thinking_tokens('\n<think>\nHello\n</think>\nWorld\n') == 'World'
|
||||
return True
|
||||
|
||||
|
||||
def run_ollama(cwd: Path, texts: list[str]) -> str:
|
||||
cmd = ['ollama', 'run', EVALUATOR_MODEL.removeprefix('ollama/')]
|
||||
print(cmd)
|
||||
process = subprocess.Popen(
|
||||
cmd,
|
||||
cwd=cwd,
|
||||
|
@ -572,14 +354,14 @@ def run_ollama(cwd: Path, texts: list[str]) -> str:
|
|||
text=True,
|
||||
)
|
||||
stdout, stderr = process.communicate('\n'.join(texts))
|
||||
stdout = remove_thinking_tokens(stdout)
|
||||
print(stdout)
|
||||
return stdout
|
||||
|
||||
|
||||
def parse_yes_no_answer(text: str) -> bool | None:
|
||||
interword = '\n \t.,?-'
|
||||
text = text.lower().strip(interword)
|
||||
words = text.split(interword)
|
||||
text = text.lower().strip()
|
||||
words = text.split('\n \t.,?-')
|
||||
print(words)
|
||||
if words[-1] in {'yes', 'agree'}:
|
||||
return True
|
||||
if words[-1] in {'no', 'disagree'}:
|
||||
|
@ -587,10 +369,6 @@ def parse_yes_no_answer(text: str) -> bool | None:
|
|||
return None
|
||||
|
||||
|
||||
assert parse_yes_no_answer('Yes.') == True
|
||||
assert parse_yes_no_answer('no') == False
|
||||
|
||||
|
||||
def run_ollama_and_get_yes_or_no(cwd, initial_texts: list[str]) -> bool:
|
||||
texts = list(initial_texts)
|
||||
texts.append('Think through your answer.')
|
||||
|
@ -605,9 +383,6 @@ def run_ollama_and_get_yes_or_no(cwd, initial_texts: list[str]) -> bool:
|
|||
|
||||
|
||||
def verify_solution(repository_path: Path, issue_content: str) -> bool:
|
||||
if not EVALUATOR_MODEL:
|
||||
return True
|
||||
|
||||
summary = run_ollama(
|
||||
repository_path,
|
||||
[
|
||||
|
@ -646,7 +421,6 @@ def solve_issue_in_repository(
|
|||
issue_description: str,
|
||||
issue_number: str,
|
||||
gitea_client,
|
||||
code_solver: CodeSolverStrategy,
|
||||
) -> IssueResolution:
|
||||
logger.info('### %s #####', issue_title)
|
||||
|
||||
|
@ -656,31 +430,28 @@ def solve_issue_in_repository(
|
|||
run_cmd(['git', 'checkout', repository_config.base_branch], repository_path)
|
||||
run_cmd(['git', 'checkout', '-b', branch_name], repository_path)
|
||||
|
||||
# Run initial ruff pass before code solver
|
||||
# Run initial ruff pass before aider
|
||||
run_cmd(['bash', '-c', RUFF_FORMAT_AND_AUTO_FIX], repository_path, check=False)
|
||||
run_cmd(['git', 'add', '.'], repository_path)
|
||||
run_cmd(['git', 'commit', '-m', 'Initial ruff pass'], repository_path, check=False)
|
||||
|
||||
# Run code solver
|
||||
# Run aider
|
||||
issue_content = f'# {issue_title}\n{issue_description}'
|
||||
|
||||
while True:
|
||||
# Save the commit hash after ruff but before code solver
|
||||
# Save the commit hash after ruff but before aider
|
||||
pre_aider_commit = get_head_commit_hash(repository_path)
|
||||
|
||||
# Run code solver
|
||||
solver_did_not_crash = code_solver.solve_issue_round(
|
||||
repository_path,
|
||||
issue_content,
|
||||
)
|
||||
if not solver_did_not_crash:
|
||||
logger.error('Code solver invocation failed for issue #%s', issue_number)
|
||||
# Run aider
|
||||
aider_did_not_crash = issue_solution_round(repository_path, issue_content)
|
||||
if not aider_did_not_crash:
|
||||
logger.error('Aider invocation failed for issue #%s', issue_number)
|
||||
return IssueResolution(False)
|
||||
|
||||
# Check if solver made any changes beyond the initial ruff pass
|
||||
# Check if aider made any changes beyond the initial ruff pass
|
||||
if not has_commits_on_branch(repository_path, pre_aider_commit, 'HEAD'):
|
||||
logger.error(
|
||||
'Code solver did not make any changes beyond the initial ruff pass for issue #%s',
|
||||
'Aider did not make any changes beyond the initial ruff pass for issue #%s',
|
||||
issue_number,
|
||||
)
|
||||
return IssueResolution(False)
|
||||
|
@ -731,9 +502,9 @@ def solve_issues_in_repository(
|
|||
title = issue.get('title', f'Issue {issue_number}')
|
||||
if seen_issues_db.has_seen(issue_url):
|
||||
logger.info('Skipping already processed issue #%s: %s', issue_number, title)
|
||||
else:
|
||||
continue
|
||||
|
||||
branch_name = generate_branch_name(issue_number, title)
|
||||
code_solver = create_code_solver()
|
||||
with tempfile.TemporaryDirectory() as repository_path:
|
||||
issue_resolution = solve_issue_in_repository(
|
||||
repository_config,
|
||||
|
@ -743,8 +514,9 @@ def solve_issues_in_repository(
|
|||
issue_description,
|
||||
issue_number,
|
||||
client,
|
||||
code_solver,
|
||||
)
|
||||
|
||||
if issue_resolution.success:
|
||||
seen_issues_db.mark_as_seen(issue_url, str(issue_number))
|
||||
seen_issues_db.update_pr_info(
|
||||
issue_url,
|
||||
|
@ -756,9 +528,6 @@ def solve_issues_in_repository(
|
|||
issue_resolution.pull_request_id,
|
||||
issue_number,
|
||||
)
|
||||
|
||||
# TODO: PR comment handling disabled for now due to missing functionality
|
||||
if False:
|
||||
# Handle unresolved pull request comments
|
||||
handle_pr_comments(
|
||||
repository_config,
|
||||
|
@ -768,31 +537,19 @@ def solve_issues_in_repository(
|
|||
client,
|
||||
seen_issues_db,
|
||||
issue_url,
|
||||
code_solver,
|
||||
)
|
||||
|
||||
# Handle failing pipelines
|
||||
handle_failing_pipelines(
|
||||
repository_config,
|
||||
issue_resolution.pull_request_id,
|
||||
branch_name,
|
||||
Path(repository_path),
|
||||
client,
|
||||
code_solver,
|
||||
)
|
||||
|
||||
|
||||
def handle_pr_comments(
|
||||
repository_config,
|
||||
pr_number: int,
|
||||
pr_number,
|
||||
branch_name,
|
||||
repository_path,
|
||||
client,
|
||||
seen_issues_db,
|
||||
issue_url,
|
||||
code_solver: CodeSolverStrategy,
|
||||
):
|
||||
"""Fetch unresolved PR comments and resolve them via code solver."""
|
||||
"""Fetch unresolved PR comments and resolve them via aider."""
|
||||
comments = client.get_pull_request_comments(
|
||||
repository_config.owner,
|
||||
repository_config.repo,
|
||||
|
@ -814,8 +571,8 @@ def handle_pr_comments(
|
|||
f'Resolve the following reviewer comment:\n{body}\n\n'
|
||||
f'File: {path}\n\nContext:\n{context}'
|
||||
)
|
||||
# invoke code solver on the comment context
|
||||
code_solver.solve_issue_round(repository_path, issue)
|
||||
# invoke aider on the comment context
|
||||
issue_solution_round(repository_path, issue)
|
||||
# commit and push changes for this comment
|
||||
run_cmd(['git', 'add', path], repository_path, check=False)
|
||||
run_cmd(
|
||||
|
@ -824,43 +581,3 @@ def handle_pr_comments(
|
|||
check=False,
|
||||
)
|
||||
run_cmd(['git', 'push', 'origin', branch_name], repository_path, check=False)
|
||||
|
||||
|
||||
def handle_failing_pipelines(
|
||||
repository_config: RepositoryConfig,
|
||||
pr_number: str,
|
||||
branch_name: str,
|
||||
repository_path: Path,
|
||||
client,
|
||||
code_solver: CodeSolverStrategy,
|
||||
) -> None:
|
||||
"""Fetch failing pipelines for the given PR and resolve them via code solver."""
|
||||
while True:
|
||||
failed_runs = client.get_failed_pipelines(
|
||||
repository_config.owner,
|
||||
repository_config.repo,
|
||||
pr_number,
|
||||
)
|
||||
if not failed_runs:
|
||||
break
|
||||
for run_id in failed_runs:
|
||||
log = client.get_pipeline_log(
|
||||
repository_config.owner,
|
||||
repository_config.repo,
|
||||
run_id,
|
||||
)
|
||||
lines = log.strip().split('\n')
|
||||
context = '\n'.join(lines[-100:])
|
||||
issue = f'Resolve the following failing pipeline run {run_id}:\n\n{context}'
|
||||
code_solver.solve_issue_round(repository_path, issue)
|
||||
run_cmd(['git', 'add', '.'], repository_path, check=False)
|
||||
run_cmd(
|
||||
['git', 'commit', '-m', f'Resolve pipeline {run_id}'],
|
||||
repository_path,
|
||||
check=False,
|
||||
)
|
||||
run_cmd(
|
||||
['git', 'push', 'origin', branch_name],
|
||||
repository_path,
|
||||
check=False,
|
||||
)
|
||||
|
|
|
@ -42,19 +42,9 @@ def parse_args():
|
|||
parser.add_argument(
|
||||
'--interval',
|
||||
type=int,
|
||||
default=30,
|
||||
default=300,
|
||||
help='Interval in seconds between checks in daemon mode (default: 300)',
|
||||
)
|
||||
parser.add_argument(
|
||||
'--aider-model',
|
||||
help='Model to use for generating code (overrides default)',
|
||||
required=True,
|
||||
)
|
||||
parser.add_argument(
|
||||
'--evaluator-model',
|
||||
help='Model to use for evaluating code (overrides default)',
|
||||
default=None,
|
||||
)
|
||||
return parser.parse_args()
|
||||
|
||||
|
||||
|
@ -62,12 +52,6 @@ def main():
|
|||
logging.basicConfig(level='INFO')
|
||||
args = parse_args()
|
||||
|
||||
# Override default models if provided
|
||||
import aider_gitea as core
|
||||
|
||||
core.CODE_MODEL = args.aider_model
|
||||
core.EVALUATOR_MODEL = args.evaluator_model
|
||||
|
||||
seen_issues_db = SeenIssuesDB()
|
||||
client = GiteaClient(args.gitea_url, secrets.gitea_token())
|
||||
|
||||
|
|
|
@ -1 +1 @@
|
|||
__version__ = '0.1.10'
|
||||
__version__ = '0.1.6'
|
||||
|
|
|
@ -167,7 +167,7 @@ class GiteaClient:
|
|||
|
||||
response = self.session.post(url, json=json_data)
|
||||
# If a pull request for this head/base already exists, return it instead of crashing
|
||||
if response.status_code == 409:
|
||||
if response.status_code == 422:
|
||||
logger.warning(
|
||||
'Pull request already exists for head %s and base %s',
|
||||
head,
|
||||
|
@ -185,28 +185,19 @@ class GiteaClient:
|
|||
response.raise_for_status()
|
||||
return response.json()
|
||||
|
||||
def get_failed_pipelines(self, owner: str, repo: str, pr_number: str) -> list[int]:
|
||||
"""Fetch pipeline runs for a PR and return IDs of failed runs."""
|
||||
url = f'{self.gitea_url}/repos/{owner}/{repo}/actions/runs'
|
||||
def get_pull_request_comments(
|
||||
self,
|
||||
owner: str,
|
||||
repo: str,
|
||||
pr_number: str,
|
||||
) -> list[dict]:
|
||||
"""
|
||||
Fetch comments for a pull request.
|
||||
"""
|
||||
url = f'{self.gitea_url}/repos/{owner}/{repo}/pulls/{pr_number}/comments'
|
||||
response = self.session.get(url)
|
||||
response.raise_for_status()
|
||||
runs = response.json().get('workflow_runs', [])
|
||||
failed = []
|
||||
for run in runs:
|
||||
if any(
|
||||
pr.get('number') == int(pr_number)
|
||||
for pr in run.get('pull_requests', [])
|
||||
):
|
||||
if run.get('conclusion') not in ('success',):
|
||||
failed.append(run.get('id'))
|
||||
return failed
|
||||
|
||||
def get_pipeline_log(self, owner: str, repo: str, run_id: int) -> str:
|
||||
"""Download the logs for a pipeline run."""
|
||||
url = f'{self.gitea_url}/repos/{owner}/{repo}/actions/runs/{run_id}/logs'
|
||||
response = self.session.get(url)
|
||||
response.raise_for_status()
|
||||
return response.text
|
||||
return response.json()
|
||||
|
||||
def get_pull_requests(
|
||||
self,
|
||||
|
|
|
@ -9,7 +9,3 @@ def llm_api_keys() -> list[str]:
|
|||
|
||||
def gitea_token() -> str:
|
||||
return SECRETS.load_or_fail('GITEA_TOKEN')
|
||||
|
||||
|
||||
def anthropic_api_key() -> str:
|
||||
return SECRETS.load_or_fail('ANTHROPIC_API_KEY')
|
||||
|
|
123
setup.py
123
setup.py
|
@ -1,9 +1,10 @@
|
|||
# WARNING!
|
||||
# THIS IS AN AUTOGENERATED FILE!
|
||||
# MANUAL CHANGES CAN AND WILL BE OVERWRITTEN!
|
||||
# WARNING
|
||||
#
|
||||
# THIS IS AN AUTOGENERATED FILE.
|
||||
#
|
||||
# MANUAL CHANGES CAN AND WILL BE OVERWRITTEN.
|
||||
|
||||
import re
|
||||
from pathlib import Path
|
||||
|
||||
from setuptools import setup
|
||||
|
||||
|
@ -12,23 +13,16 @@ PACKAGE_NAME = 'aider_gitea'
|
|||
PACKAGE_DESCRIPTION = """
|
||||
Aider Gitea.
|
||||
|
||||
A code automation tool that integrates Gitea with AI assistants to automatically solve issues.
|
||||
A code automation tool that integrates Gitea with Aider to automatically solve issues.
|
||||
|
||||
This program monitors your [Gitea](https://about.gitea.com/) repository for issues with the 'aider' label.
|
||||
When such an issue is found, it:
|
||||
|
||||
1. Creates a new branch.
|
||||
2. Invokes an AI assistant (Aider or Claude Code) to solve the issue using a Large-Language Model.
|
||||
2. Invokes [Aider](https://aider.chat/) to solve the issue using a Large-Language Model.
|
||||
3. Runs tests and code quality checks.
|
||||
4. Creates a pull request with the solution.
|
||||
|
||||
The tool automatically selects the appropriate AI assistant based on the specified model:
|
||||
- **Aider**: Used for non-Anthropic models (e.g., GPT, Ollama, Gemini)
|
||||
- **Claude Code**: Used for Anthropic models (e.g., Claude, Sonnet, Haiku, Opus)
|
||||
|
||||
Inspired by [the AI workflows](https://github.com/oscoreio/ai-workflows/)
|
||||
project.
|
||||
|
||||
## Usage
|
||||
|
||||
An application token must be supplied for the `gitea_token` secret. This must
|
||||
|
@ -41,138 +35,63 @@ have the following permissions:
|
|||
### Command Line
|
||||
|
||||
```bash
|
||||
# Run with default settings (uses Aider)
|
||||
python -m aider_gitea --aider-model gpt-4
|
||||
|
||||
# Use Claude Code with Anthropic models
|
||||
python -m aider_gitea --aider-model claude-3-sonnet
|
||||
python -m aider_gitea --aider-model claude-3-haiku
|
||||
python -m aider_gitea --aider-model anthropic/claude-3-opus
|
||||
|
||||
# Use Aider with various models
|
||||
python -m aider_gitea --aider-model gpt-4
|
||||
python -m aider_gitea --aider-model ollama/llama3
|
||||
python -m aider_gitea --aider-model gemini-pro
|
||||
# Run with default settings
|
||||
python -m aider_gitea
|
||||
|
||||
# Specify custom repository and owner
|
||||
python -m aider_gitea --owner myorg --repo myproject --aider-model claude-3-sonnet
|
||||
python -m aider_gitea --owner myorg --repo myproject
|
||||
|
||||
# Use a custom Gitea URL
|
||||
python -m aider_gitea --gitea-url https://gitea.example.com --aider-model gpt-4
|
||||
python -m aider_gitea --gitea-url https://gitea.example.com
|
||||
|
||||
# Specify a different base branch
|
||||
python -m aider_gitea --base-branch develop --aider-model claude-3-haiku
|
||||
python -m aider_gitea --base-branch develop
|
||||
```
|
||||
|
||||
### AI Assistant Selection
|
||||
|
||||
The tool automatically routes to the appropriate AI assistant based on the model name:
|
||||
|
||||
**Claude Code Integration (Anthropic Models):**
|
||||
- Model names containing: `claude`, `anthropic`, `sonnet`, `haiku`, `opus`
|
||||
- Examples: `claude-3-sonnet`, `claude-3-haiku`, `anthropic/claude-3-opus`
|
||||
- Requires: `ANTHROPIC_API_KEY` environment variable
|
||||
|
||||
**Aider Integration (All Other Models):**
|
||||
- Any model not matching Anthropic patterns
|
||||
- Examples: `gpt-4`, `ollama/llama3`, `gemini-pro`, `mistral-7b`
|
||||
- Requires: `LLM_API_KEY` environment variable
|
||||
|
||||
### Python API
|
||||
|
||||
```python
|
||||
from aider_gitea import solve_issue_in_repository, create_code_solver
|
||||
from aider_gitea import solve_issue_in_repository
|
||||
from pathlib import Path
|
||||
import argparse
|
||||
|
||||
# Solve an issue programmatically with automatic AI assistant selection
|
||||
repository_config = RepositoryConfig(
|
||||
# Solve an issue programmatically
|
||||
args = argparse.Namespace(
|
||||
gitea_url="https://gitea.example.com",
|
||||
owner="myorg",
|
||||
repo="myproject",
|
||||
base_branch="main"
|
||||
)
|
||||
|
||||
# Set the model to control which AI assistant is used
|
||||
import aider_gitea
|
||||
aider_gitea.CODE_MODEL = "claude-3-sonnet" # Will use Claude Code
|
||||
# aider_gitea.CODE_MODEL = "gpt-4" # Will use Aider
|
||||
|
||||
code_solver = create_code_solver() # Automatically selects based on model
|
||||
|
||||
solve_issue_in_repository(
|
||||
repository_config,
|
||||
args,
|
||||
Path("/path/to/repo"),
|
||||
"issue-123-fix-bug",
|
||||
"Fix critical bug",
|
||||
"The application crashes when processing large files",
|
||||
"123",
|
||||
gitea_client,
|
||||
code_solver
|
||||
"123"
|
||||
)
|
||||
```
|
||||
|
||||
### Environment Configuration
|
||||
|
||||
The tool uses environment variables for sensitive information:
|
||||
|
||||
**Required for all setups:**
|
||||
- `GITEA_TOKEN`: Your Gitea API token
|
||||
|
||||
**For Aider (non-Anthropic models):**
|
||||
- `LLM_API_KEY`: API key for the language model (OpenAI, Ollama, etc.)
|
||||
|
||||
**For Claude Code (Anthropic models):**
|
||||
- `ANTHROPIC_API_KEY`: Your Anthropic API key for Claude models
|
||||
|
||||
### Model Examples
|
||||
|
||||
**Anthropic Models (→ Claude Code):**
|
||||
```bash
|
||||
--aider-model claude-3-sonnet
|
||||
--aider-model claude-3-haiku
|
||||
--aider-model claude-3-opus
|
||||
--aider-model anthropic/claude-3-sonnet
|
||||
```
|
||||
|
||||
**Non-Anthropic Models (→ Aider):**
|
||||
```bash
|
||||
--aider-model gpt-4
|
||||
--aider-model gpt-3.5-turbo
|
||||
--aider-model ollama/llama3
|
||||
--aider-model ollama/codellama
|
||||
--aider-model gemini-pro
|
||||
--aider-model mistral-7b
|
||||
```
|
||||
- `LLM_API_KEY`: API key for the language model used by Aider
|
||||
```
|
||||
""".strip()
|
||||
|
||||
PACKAGE_DESCRIPTION_SHORT = """
|
||||
A code automation tool that integrates Gitea with AI assistants to automatically solve issues.""".strip()
|
||||
A code automation tool that integrates Gitea with Aider to automatically solve issues.""".strip()
|
||||
|
||||
|
||||
def parse_version_file(text: str) -> str:
|
||||
text = re.sub('^#.*', '', text, flags=re.MULTILINE)
|
||||
match = re.match(r'^\s*__version__\s*=\s*(["\'])([\d\.]+)\1$', text)
|
||||
match = re.match(r'^__version__\s*=\s*(["\'])([\d\.]+)\1$', text)
|
||||
if match is None:
|
||||
msg = 'Malformed _version.py file!'
|
||||
raise Exception(msg)
|
||||
return match.group(2)
|
||||
|
||||
|
||||
def find_python_packages() -> list[str]:
|
||||
"""
|
||||
Find all python packages. (Directories containing __init__.py files.)
|
||||
"""
|
||||
root_path = Path(PACKAGE_NAME)
|
||||
packages: set[str] = set([PACKAGE_NAME])
|
||||
|
||||
# Search recursively
|
||||
for init_file in root_path.rglob('__init__.py'):
|
||||
packages.add(str(init_file.parent).replace('/', '.'))
|
||||
|
||||
return sorted(packages)
|
||||
|
||||
with open(PACKAGE_NAME + '/_version.py') as f:
|
||||
version = parse_version_file(f.read())
|
||||
|
||||
|
@ -192,7 +111,7 @@ setup(
|
|||
author='Jon Michael Aanes',
|
||||
author_email='jonjmaa@gmail.com',
|
||||
url='https://gitfub.space/Jmaa/' + PACKAGE_NAME,
|
||||
packages=find_python_packages(),
|
||||
packages=[PACKAGE_NAME],
|
||||
install_requires=REQUIREMENTS_MAIN,
|
||||
extras_require={
|
||||
'test': REQUIREMENTS_TEST,
|
||||
|
|
|
@ -1,122 +0,0 @@
|
|||
import pytest
|
||||
|
||||
from aider_gitea import (
|
||||
AIDER_LINT,
|
||||
AIDER_TEST,
|
||||
CLAUDE_CODE_MESSAGE_FORMAT,
|
||||
AiderCodeSolver,
|
||||
ClaudeCodeSolver,
|
||||
create_code_solver,
|
||||
is_anthropic_model,
|
||||
)
|
||||
|
||||
|
||||
class TestClaudeCodeIntegration:
|
||||
"""Test Claude Code integration and model routing logic."""
|
||||
|
||||
def test_is_anthropic_model_detection(self):
|
||||
"""Test that Anthropic models are correctly detected."""
|
||||
# Anthropic models should return True
|
||||
assert is_anthropic_model('claude-3-sonnet')
|
||||
assert is_anthropic_model('claude-3-haiku')
|
||||
assert is_anthropic_model('claude-3-opus')
|
||||
assert is_anthropic_model('anthropic/claude-3-sonnet')
|
||||
assert is_anthropic_model('Claude-3-Sonnet') # Case insensitive
|
||||
assert is_anthropic_model('ANTHROPIC/CLAUDE')
|
||||
assert is_anthropic_model('some-sonnet-model')
|
||||
assert is_anthropic_model('haiku-variant')
|
||||
|
||||
# Non-Anthropic models should return False
|
||||
assert not is_anthropic_model('gpt-4')
|
||||
assert not is_anthropic_model('gpt-3.5-turbo')
|
||||
assert not is_anthropic_model('ollama/llama')
|
||||
assert not is_anthropic_model('gemini-pro')
|
||||
assert not is_anthropic_model('mistral-7b')
|
||||
assert not is_anthropic_model('')
|
||||
assert not is_anthropic_model(None)
|
||||
|
||||
def test_create_code_solver_routing(self, monkeypatch):
|
||||
"""Test that the correct solver is created based on model."""
|
||||
import aider_gitea
|
||||
|
||||
# Test Anthropic model routing
|
||||
monkeypatch.setattr(aider_gitea, 'CODE_MODEL', 'claude-3-sonnet')
|
||||
solver = create_code_solver()
|
||||
assert isinstance(solver, ClaudeCodeSolver)
|
||||
|
||||
# Test non-Anthropic model routing
|
||||
monkeypatch.setattr(aider_gitea, 'CODE_MODEL', 'gpt-4')
|
||||
solver = create_code_solver()
|
||||
assert isinstance(solver, AiderCodeSolver)
|
||||
|
||||
# Test None model routing (should default to Aider)
|
||||
monkeypatch.setattr(aider_gitea, 'CODE_MODEL', None)
|
||||
solver = create_code_solver()
|
||||
assert isinstance(solver, AiderCodeSolver)
|
||||
|
||||
def test_claude_code_solver_command_creation(self):
|
||||
"""Test that Claude Code commands are created correctly."""
|
||||
import aider_gitea
|
||||
|
||||
solver = ClaudeCodeSolver()
|
||||
issue = 'Fix the bug in the code'
|
||||
|
||||
# Test without model
|
||||
with pytest.MonkeyPatch().context() as m:
|
||||
m.setattr(aider_gitea, 'CODE_MODEL', None)
|
||||
cmd = solver._create_claude_command(issue)
|
||||
expected = [
|
||||
'claude',
|
||||
'-p',
|
||||
'--output-format',
|
||||
'json',
|
||||
'--max-turns',
|
||||
'10',
|
||||
issue,
|
||||
]
|
||||
assert cmd == expected
|
||||
|
||||
# Test with model
|
||||
with pytest.MonkeyPatch().context() as m:
|
||||
m.setattr(aider_gitea, 'CODE_MODEL', 'claude-3-sonnet')
|
||||
cmd = solver._create_claude_command(issue)
|
||||
expected = [
|
||||
'claude',
|
||||
'-p',
|
||||
'--output-format',
|
||||
'json',
|
||||
'--max-turns',
|
||||
'10',
|
||||
'--model',
|
||||
'claude-3-sonnet',
|
||||
issue,
|
||||
]
|
||||
assert cmd == expected
|
||||
|
||||
def test_claude_code_message_format(self):
|
||||
"""Test that Claude Code message format works correctly."""
|
||||
issue_content = 'Fix the authentication bug'
|
||||
|
||||
formatted_message = CLAUDE_CODE_MESSAGE_FORMAT.format(
|
||||
issue=issue_content,
|
||||
test_command=AIDER_TEST,
|
||||
lint_command=AIDER_LINT,
|
||||
)
|
||||
|
||||
# Verify the issue content is included
|
||||
assert issue_content in formatted_message
|
||||
|
||||
# Verify the test and lint commands are included
|
||||
assert AIDER_TEST in formatted_message
|
||||
assert AIDER_LINT in formatted_message
|
||||
|
||||
# Verify the guidelines are present
|
||||
assert 'Run tests after making changes' in formatted_message
|
||||
assert 'Follow existing code style' in formatted_message
|
||||
assert 'Make minimal, focused changes' in formatted_message
|
||||
assert 'Commit your changes' in formatted_message
|
||||
|
||||
# Verify the structure contains placeholders that got replaced
|
||||
assert '{issue}' not in formatted_message
|
||||
assert '{test_command}' not in formatted_message
|
||||
assert '{lint_command}' not in formatted_message
|
Loading…
Reference in New Issue
Block a user