Gitbot uses GitHub Actions to query an LLM (Large Language Model) and respond with useful insights or code snippets. The action is designed to run specific models from Ollama, with different commands based on the model specifications.
Phi 3 Medium
).- **Tip.1**: Free AI resources can sometimes be slow, so consider watching a quick video or taking a break while you wait!
- **Tip.2**: Keep your specific use case in mind and select a model that aligns best with it.
- **Tip.3**: The structure and content of your query matter. Provide clear instructions and context for the response you’re looking for.
-Tip.4: Treat your query as a prompt. Be precise and technical about your needs, and remember to include any relevant context that the AI might not inherently know. This `concept of "inherently knowing" is crucial, as general knowledge can vary widely based on context and specific domains. What might seem like common knowledge to you may `not be recognized by the AI, especially if it lacks specific details or nuances related to your topic. Therefore, providing additional information can significantly `enhance the AI's ability to generate accurate and relevant responses.
`For instance, if you're asking for help with a technical problem, include relevant details like the software version, specific error messages, or the steps you've already `tried. The more context you provide, the better the AI can tailor its response to your unique situation.
Model | Parameters | Size | Command |
---|---|---|---|
Llama 3 | 8B | 4.7GB | ollama run llama3 |
Llama 3 | 70B | 40GB | ollama run llama3:70b |
Phi 3 Mini | 3.8B | 2.3GB | ollama run phi3 |
Phi 3 Medium | 14B | 7.9GB | ollama run phi3:medium |
Gemma | 2B | 1.4GB | ollama run gemma:2b |
Gemma | 7B | 4.8GB | ollama run gemma:7b |
Mistral | 7B | 4.1GB | ollama run mistral |
Moondream 2 | 1.4B | 829MB | ollama run moondream |
Neural Chat | 7B | 4.1GB | ollama run neural-chat |
Starling | 7B | 4.1GB | ollama run starling-lm |
Code Llama | 7B | 3.8GB | ollama run codellama |
Llama 2 Uncensored | 7B | 3.8GB | ollama run llama2-uncensored |
LLaVA | 7B | 4.5GB | ollama run llava |
Solar | 10.7B | 6.1GB | ollama run solar |
Below is a table showing simple benchmarks for how long each model takes to complete a query. Note that performance may vary based on server load and the complexity of your request.
Model | Completion Time | Notes | Benchmarks | Pass/Fail |
---|---|---|---|---|
Phi 3 Mini | 2m:37s | Fast response, suitable for simpler queries | Phi 3 Mini Benchmark | Pass |
Phi 3 Medium | 5m:20s | Great for more complex code explanations | Phi 3 Medium Benchmark | Pass |
Llama 3 (8B) | 3m:20s | Good balance between speed and depth | Llama 3 8B Benchmark | Pass |
Mistral | 2m:41s | Very fast but less detailed | Mistral Benchmark | Pass |
Moondream 2 | 1m:24s | Quickest but limited in complexity | Unable to test without GPU | Fail |
Neural Chat | 3m:06s | Effective for conversational queries | Neural Chat Benchmark | Pass |
Code Llama | 3m:03s | Optimized for coding tasks | Code Llama Benchmark | Pass |
Llama 3 (70B) | 1m:38s | Large model by Meta, good for code generation | Too large for Codespaces | Fail |
Solar | 4m | Takes longer, but high detail | Solar Benchmark | Pass |
LLaVA | 2m:31s | Ideal for visual language tasks | LLaVA Benchmark | Pass |
Llama 2 Uncensored | 2m:48s | Uncensored Llama 2 model | Llama 2 Uncensored Benchmark | Pass |
Gemma (2B) | 1m:32s | High-performing and efficient | Gemma (2B) Benchmark | Pass |
Gemma (7B) | 3m | High-performing and efficient | Gemma (7B) Benchmark | Pass |
Starling | 2m:50s | Large language model trained by reinforcement learning | Starling Benchmark | Pass |
For a completed query example, see this issue. It ran Phi 3 Mini
and completed in 2 minutes and 37 seconds, providing correct code with a detailed explanation. The same code will be repeated for all models over time, and results will be added to the table.
all times in this section are related to the benchmark from Benchmark information on how the test is configured this test will not be changed and will be used to test all llm's that are added to the system
Simple Queries:
Complex Code Explanations:
Conversational Tasks:
Coding Tasks:
Visual Language Tasks:
High Detail Requirements:
Uncensored Outputs:
Reinforcement Learning Applications: