-
### System Info
I am experimenting with TRT LLM and `flan-t5` models. My simple goal is to build engines with different configurations and tensor parallelism, then review performance. Have a DGX syst…
-
### Your current environment
The output of `python collect_env.py`
```text
PyTorch version: 2.4.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A…
jgen1 updated
3 weeks ago
-
I run Code on Colab, in this [cell code](https://colab.research.google.com/github/HandsOnLLM/Hands-On-Large-Language-Models/blob/main/chapter09/Chapter%209%20-%20Multimodal%20Large%20Language%20Models…
-
Requires extensive automated and manual testing and code changes (imports) which are part of https://github.com/Chainlit/chainlit/pull/1267.
-
**Describe the bug**
```
Failed to create agent from provided information:
Agent with name AngryReindeer already exists
```
**Please describe your setup**
- [ ] MemGPT version
0.2.12
-…
-
I'd like to run live llava completely locally on Jetson including a web browser.
However, if I turn off wifi before starting live llava, the video won't play on the browser.
If I turn off wifi after…
-
## Goal
Follow up #81 and make the web editor usable for managing models in local/GitHub folders
## Development items
- #473
- [ ] Remove ScalablyTyped plugin, as it can't import duckdb-wasm and ge…
-
I have installed trl
-
### What is the issue?
No issues with any model that fits into a single 3090 but seems to run out of memory when trying to distribute to the second 3090.
```
INFO [wmain] starting c++ runner | ti…
-
**Describe the bug**
Hi, all. Working on a blog article, following a mix of local documentation + Intelligent app workshop, but instead of going Falcon, I've gone with the Mistral 7b model. and at …