-
Dears,
I failed to run Llama-2-7b-chat-hf on NPU, please give me a hand.
1. I converted the mode by below command, and got two models,
a) optimum-cli export openvino --task text-generation -m Meta-…
-
I want to benchmark faster-whisper and some pipeline whisper implementations of whisper in huggingface.
For the sake of fairness I would like to parametrize the models as equally as possible.
In H…
-
**Why**
The ability to continue chats in a Beam would enable users to:
- compare the reasoning of multiple models without necessarily wanting to obtain *one* merged answer
- guide each model toward…
dandv updated
4 months ago
-
I realized that `convert_stills_to_sequence=True` leads to a unified beam model. Is this expected/intended? This is fine with SMX datasets but not adequate for SFX, because XFEL pulses have shot-to-sh…
-
High priority:
- [ ] Generic unknown offworld item. Item sphere with a question mark on it?
- [ ] Generic beam. Someone suggested removing the letter from an existing beam model and perhaps recoloring…
-
This only happens with `BEAM=1`. `BEAM=0`, `BEAM=2`, `BEAM=3` all work fine
This happens because exo runs tinygrad inference on another thread.
Example command to reproduce: `DEBUG=6 BEAM=1 python3 …
-
Hello, How should I set the decoding parameters (e.g., temperature) for Gemma-2? My result is about ~50.0, far from the benchmark of 76.
-
Hi!
Thanks for open-sourcing the codes for Beam Retrieval.
Do you have any plans to share the checkpoints of the models?
Thanks a lot!
-
Extend and improve the Beam feature of Big-AGI.
Possible vectors for improvement are below:
- **UX**:
- [ ] Respond to single chats (chat bubble in the bottom-left) #553
- [ ] Looping Beam (…
-
- Encoder
- Decoder with Beam Search
https://github.com/axinc-ai/ailia-models/pull/1500