Patchwork automates development gruntwork like PR reviews, bug fixing, security patching, and more using a self-hosted CLI agent and your preferred LLMs. Try the hosted version here.
Patchflows can be run locally in your CLI and IDE, or as part of your CI/CD pipeline. There are several patchflows available out of the box, and you can always create your own.
Patchwork is available on PyPI and can be installed using pip:
pip install 'patchwork-cli[all]' --upgrade
The following optional dependency groups are available.
security
: Installs semgrep
and depscan
with pip install 'patchwork-cli[security]'
and is required for AutoFix and DependencyUpgrade patchflows.rag
: Installs chromadb
with pip install 'patchwork-cli[rag]'
and is required for the ResolveIssue patchflow.notifications
: Used by steps sending notifications, e.g. slack messages.all
: installs everything.pip install patchwork-cli
) will install a core set of dependencies that are sufficient to run the GenerateDocstring, PRReview and GenerateREADME patchflows.If you'd like to build from source using poetry, please see detailed documentation here .
The CLI runs Patchflows, as follows:
patchwork <PatchFlow> <?Arguments>
Where
key=value
. If key
does not have any value, it is considered a boolean True
flag.For an AutoFix patchflow which patches vulnerabilities based on a scan using Semgrep:
patchwork AutoFix openai_api_key=<YOUR_OPENAI_API_KEY> github_api_key=<YOUR_GITHUB_TOKEN>
The above command defaults to patching code in the current directory by running Semgrep to identify the vulnerabilities. You can view the default.yml
file for the list of configurations you can set to manage the AutoFix patchflow. For more details on how you can use a personal access token from GitHub on CLI, can read this.
You can replace the OpenAI key with a key from our managed service by signing in at https://app.patched.codes/signin and generating an API key from the integrations tab. You can then call the patchflow with the key as follows:
patchwork AutoFix patched_api_key=<YOUR_PATCHED_API_KEY> github_api_key=<YOUR_GITHUB_TOKEN>
To use Google's models you can set the google_api_key
and model
, this is useful if you want to work with large contexts as the gemini-pro-1.5
model supports an input context length of 1 million tokens.
The patchwork-template repository contains the default configuration and prompts for all the patchflows. You can clone that repo and pass it as a flag to the CLI:
patchwork AutoFix --config /path/to/patchwork-configs/patchflows
Patchwork supports any OpenAI compatible endpoint, allowing use of any LLM from various providers like Groq, Together AI, or Hugging Face.
E.g. to use Llama 3.1 405B from Groq.com run:
patchwork AutoFix client_base_url=https://api.groq.com/openai/v1 openai_api_key=your_groq_key model=llama-3.1-405b-reasoning
You can also use a config file to do the same. To use Llama 3.1 405B from Hugging Face, create a config.yml file:
openai_api_key: your_hf_token
client_base_url: https://api-inference.huggingface.co/models/meta-llama/Meta-Llama-3.1-405B-Instruct-FP8/v1
model: Meta-Llama-3.1-405B-Instruct-FP8
And run as:
patchwork AutoFix --config=/path/to/config.yml
This allows you to run local models via llama.cpp
, ollama
, vllm
or tgi
. For instance, you can run Llama 3.1 8B locally using llama_cpp.server
:
python -m llama_cpp.server --hf_model_repo_id bullerwins/Meta-Llama-3.1-8B-Instruct-GGUF --model 'Meta-Llama-3.1-8B-Instruct-Q4_K_M.gguf' --chat_format chatml
Then run your patchflow:
patchwork AutoFix client_base_url=http://localhost:8080/v1 openai_api_key=no_key_local_model
Patchwork comes with predefined patchflows, with more added over time. Sample patchflows include:
Prompt templates are used by patchflows and passed as queries to LLMs. Templates contain prompts with placeholder variables enclosed by {{}}
which are replaced by the data from the steps or inputs on every run.
Below is a sample prompt template:
{
"id": "diffreview_summary",
"prompts": [
{
"role": "user",
"content": "Summarize the following code change descriptions in 1 paragraph. {{diffreviews}}"
}
]
}
Each patchflow comes with an optimized default prompt template. But you can specify your own using the prompt_template_file=/path/to/prompt/template/file
option.
Contributions for new patchflows and steps, or to the core framework are welcome. Please look at open issues for details.
We also provide a chat assistant to help you create new steps and patchflows easily.
Patchwork is licensed under AGPL-3.0 terms. However, custom patchflows and steps can be created and shared using the patchwork template repository which is licensed under Apache-2.0 terms.