AdieLaine / lnu-ai

Lnu-AI: An Indigenous AI System. Preserving the Mi'kmaq Language in AI.
https://twitter.com/justmadielaine
MIT License
9 stars 0 forks source link

ModuleNotFoundError: No module named 'jsonlines' #1

Open thoughtformer opened 1 year ago

thoughtformer commented 1 year ago

Hi there

Thank you for making this. I'm so curious to dive into this.

I'm running into a small bug though: when I try to run it via the streamlit run src/lnu-ai.py command I get the following error message:

File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/site-packages/streamlit/runtime/scriptrunner/script_runner.py", line 565, in _run_script exec(code, module.dict) File "/Users/voidwalker/Documents/Documents - Starhaven Air/Starhaven Forge/04_Lnu-AI/lnu-ai/src/lnu-ai.py", line 9, in import jsonlines

SCR-20230621-lpsu

Anything you can think of that may cause this issue?

AdieLaine commented 1 year ago

Hey there, it looks like you need to install one of the modules named jsonlines. Ensure that you are in a Python 3.10 environment and run requirements.txt. Then run the command from inside the root folder named lnu-ai.

thoughtformer commented 1 year ago

Thanks, some dependencies were indeed missing. However, I'm now receiving this error message:

ERROR: pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.
argilla 1.9.0 requires numpy<1.24.0, but you have numpy 1.24.3 which is incompatible.
argilla 1.9.0 requires pandas<2.0.0,>=1.0.0, but you have pandas 2.0.1 which is incompatible.
argilla 1.9.0 requires rich<=13.0.1, but you have rich 13.4.2 which is incompatible.
google-auth 2.19.1 requires urllib3<2.0, but you have urllib3 2.0.3 which is incompatible.
mdit-py-plugins 0.3.3 requires markdown-it-py<3.0.0,>=1.0.0, but you have markdown-it-py 3.0.0 which is incompatible.
mediapipe 0.10.1 requires protobuf<4,>=3.11, but you have protobuf 4.23.3 which is incompatible.
open-clip-torch 2.7.0 requires protobuf==3.20.0, but you have protobuf 4.23.3 which is incompatibl

Do I have to downgrade all these individual versions?

Thanks!!

thoughtformer commented 1 year ago

I tried to run

streamlit run src/lnu-ai.py

and get these error messages:

File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/site-packages/streamlit/runtime/scriptrunner/script_runner.py", line 552, in _run_script
    exec(code, module.__dict__)
File "/Users/voidwalker/Documents/Documents - Starhaven Air/Starhaven Forge/04_Lnu-AI/lnu-ai/src/lnu-ai.py", line 134, in <module>
    all_word_details: dict = load_all_word_details(file)
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/site-packages/streamlit/runtime/caching/cache_utils.py", line 209, in wrapper
    return cached_func(*args, **kwargs)
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/site-packages/streamlit/runtime/caching/cache_utils.py", line 238, in __call__
    return self._get_or_create_cached_value(args, kwargs)
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/site-packages/streamlit/runtime/caching/cache_utils.py", line 263, in _get_or_create_cached_value
    return self._handle_cache_miss(cache, value_key, func_args, func_kwargs)
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/site-packages/streamlit/runtime/caching/cache_utils.py", line 317, in _handle_cache_miss
    computed_value = self._info.func(*func_args, **func_kwargs)
File "/Users/voidwalker/Documents/Documents - Starhaven Air/Starhaven Forge/04_Lnu-AI/lnu-ai/src/lnu-ai.py", line 127, in load_all_word_details
    all_word_details = json.load(f)
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/json/__init__.py", line 293, in load
    return loads(fp.read(),
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/json/__init__.py", line 346, in loads
    return _default_decoder.decode(s)
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/json/decoder.py", line 337, in decode
    obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/Library/Frameworks/Python.framework/Versions/3.10/lib/python3.10/json/decoder.py", line 355, in raw_decode
    raise JSONDecodeError("Expecting value", s, err.value) from None

No worries, if you can't solve them for me, just really curious to explore your research 🤗

AdieLaine commented 1 year ago

No worries, let's see if we can figure this out.

The error message Expecting value usually means that the decoder was expecting to find a value for a key in the JSON document but did not find one. This could happen for several reasons:

The JSON document could be empty. There could be a typo or some unexpected formatting in the JSON data. There may be a mismatch between the encoding of the file and what the JSON module expects.

Here's what you can do to debug this issue:

Try pulling the program again and see if the error persists.

thoughtformer commented 1 year ago

Thank you again for taking the time to troubleshoot this issue. Here's the code contained in the all_world_details.json document:

version https://git-lfs.github.com/spec/v1
oid sha256:cf442f1a1d386b785eff9c400a2782f4ac74c63c3a11c3509e56a2dc152972cd
size 2828118

Do I need to add something here? Wela'lin!

AdieLaine commented 1 year ago

I see the issue. I am using Git LFS because some of the files exceed a certain size. I will have to adjust this, but in the interim you have two options.

Option 1:

You can download the zip of the repository via the "Download ZIP" option, the downloaded zip file should contain the actual files, not the LFS pointers. Simply replace the data files.

Option 2:

AdieLaine commented 1 year ago

Removed LFS dependency