Open mickitty0511 opened 4 months ago
got the same
same too
same.
It seems doesn't supported llama3.1
Maybe you can run llama run llama3
and use interpreter --model ollama/llama3
to use
This PR should fix this issue. https://github.com/OpenInterpreter/open-interpreter/pull/1400
I'm still facing this issue on Windows 11, even running the latest OI, Ollama and Llama 3.1 versions.
PS C:\Users\User> interpreter --version
Open Interpreter 0.3.7 The Beginning (Ty and Victor)
PS C:\Users\User> ollama --version
ollama version is 0.3.6
PS C:\Users\User> interpreter --model ollama/llama3.1
▌ Model set to ollama/llama3.1
Loading llama3.1...
Traceback (most recent call last):
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\runpy.py", line 197, in _run_module_as_main
return _run_code(code, main_globals, None,
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\runpy.py", line 87, in _run_code
exec(code, run_globals)
File "C:\Users\User\AppData\Local\Programs\Python\Python39\Scripts\interpreter.exe\__main__.py", line 7, in <module>
sys.exit(main())
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\site-packages\interpreter\terminal_interface\start_terminal_interface.py", line 586, in main
start_terminal_interface(interpreter)
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\site-packages\interpreter\terminal_interface\start_terminal_interface.py", line 540, in start_terminal_interface
validate_llm_settings(
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\site-packages\interpreter\terminal_interface\validate_llm_settings.py", line 110, in validate_llm_settings
interpreter.llm.load()
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\site-packages\interpreter\core\llm\llm.py", line 358, in load
self.interpreter.computer.ai.chat("ping")
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\site-packages\interpreter\core\computer\ai\ai.py", line 130, in chat
for chunk in self.computer.interpreter.llm.run(messages):
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\site-packages\interpreter\core\llm\llm.py", line 291, in run
yield from run_tool_calling_llm(self, params)
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\site-packages\interpreter\core\llm\run_tool_calling_llm.py", line 177, in run_tool_calling_llm
for chunk in llm.completions(**request_params):
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\site-packages\interpreter\core\llm\llm.py", line 420, in fixed_litellm_completions
raise first_error # If all attempts fail, raise the first error
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\site-packages\interpreter\core\llm\llm.py", line 400, in fixed_litellm_completions
yield from litellm.completion(**params)
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\site-packages\litellm\llms\ollama.py", line 370, in ollama_completion_stream
raise e
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\site-packages\litellm\llms\ollama.py", line 348, in ollama_completion_stream
function_call = json.loads(response_content)
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\json\__init__.py", line 346, in loads
return _default_decoder.decode(s)
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\json\decoder.py", line 337, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "C:\Users\User\AppData\Local\Programs\Python\Python39\lib\json\decoder.py", line 353, in raw_decode
obj, end = self.scan_once(s, idx)
json.decoder.JSONDecodeError: Unterminated string starting at: line 1 column 2 (char 1)
PS C:\Users\User>
I'm not sure about what I'm doing wrong.
@mickitty0511 @leafarilongamor
I brought this up internally and it's being worked on! Thanks for raising the issue
Hi @MikeBirdTech, Same issue is on macOS also.
goku@192 ~ % interpreter --version
Open Interpreter 0.3.7 The Beginning (Ty and Victor)
goku@192 ~ % ollama -v
ollama version is 0.3.6
I am on macOS Version 15.0 Beta (24A5309e) if that makes any difference for you. Best!
same issue
interpreter --version
Open Interpreter 0.4.3 Developer Preview
ollama -v
ollama version is 0.3.13
@MikeBirdTech @leafarilongamor Did ya'll got the fix? I'm trying to load llama 3.2 and I'm getting the same error
This will be fixed with the merge of this PR: https://github.com/OpenInterpreter/open-interpreter/pull/1524
If you want to try it early you can install it like this
pip install --upgrade --force-reinstall git+https://github.com/CyanideByte/open-interpreter.git@local-fixes
Describe the bug
When reading your official doc about using ollama's module, I tried using llama 3.1 for open interpreter. However, errors were produced during the preps made after the module setup. I need a detailed resolution or explanation about what happened in my case. I hope some developers would reproduce this error and then tell me about this case.
Reproduce
Follow your official doc
Used this command
Then, open interpreter asked me if I would have new profile file. I did answer n.
Then error is as follows.
[2024-07-30T03:56:01Z ERROR cached_path::cache] ETAG fetch for https://huggingface.co/llama3.1/resolve/main/tokenizer.json failed with fatal error Traceback (most recent call last):
json.decoder.JSONDecodeError: Unterminated string starting at: line 1 column 2 (char 1)
Expected behavior
I suppose it should complete the preps based on what I read from your official docs.
Screenshots
No response
Open Interpreter version
0.3.4
Python version
3.11.5
Operating System name and version
Windows 11
Additional context
No response