-
The model engine is built from llama 3 70b with tensor parallelism tp=2 and pp=2 and deployed by below triton launch script:
python3 scripts/launch_triton_server.py --world_size 4 --model_repo=llama_…
-
### 🚀 The feature, motivation and pitch
### Motivation
SPMD sharding in pytorch/XLA offers model parallelism by sharding tensors within an operator. However, we need a mechanism to integrate thi…
-
Do you know how much memory individual_llama2 took? I saw you mentioned you ran it on 1 80gb A100. For people with smaller GPUs it might be nice to implement allowing the work to be split over multipl…
-
Unfortunately this discussion [started in Error handling thread](https://github.com/keean/zenscript/issues/40#issuecomment-390693316).
And the discussion derived from [a post I made](https://github…
-
Hi,
1st, Thank you very much for an amazing tool. it worked like a charm
2nd, it would be great if the migration of the data can be set with an amount of parallelism. so i can copy X tables at once.
…
-
What changes would it take for cover to instrument and run test modules in parallel?
-
### Apache Iceberg version
1.5.0
### Query engine
Flink (iceberg-flink-1.18:1.5.0)
### Question
Hello, I'm using iceberg-flink-1.18-1.5.0.
I've configured the [flink-operator autoscaler fe…
-
### Description of the problem / feature request:
This is an umbrella issue of problems that arise from using build tools that have their own internal parallelism.
In this Google Groups thread, …
-
I noticed there are some settings about tensor parallelism in `DeepSpeedEngine` and `PipielineEngine`. Can you please provide us with some examples of combinig tensor parallelism with pipeline paralle…
-
https://cython.readthedocs.io/en/latest/src/userguide/parallelism.html