-
Hi! Suppose in my cluster I have 2 nodes with 2 gpus each. What is the better practice for using all 4 gpus:
1. To spawn 4 pods with 1 gpu per each
or
2. To spawn 2 pods with 2 gpus each?
I've …
-
# ✨ Feature Request
### **Description of the Feature**
The Online Complaint Box feature will allow users to submit complaints or feedback directly through the platform. This feature will offer a f…
-
Environment:
Hardware: Power 10 system (PPC64LE)
OS: Red Hat Enterprise Linux release 9.3 (Plow)
kernel: 5.14.0-362.18.1.el9_3.ppc64le
GH repo: https://github.com/foundation-model-stack/found…
-
🚀 The feature, motivation and pitch
# RFC: Multi-Gpu Python Frontend API
This RFC compares and contrasts some ideas for exposing multi-gpu support in the python frontend.
1. The current `multigpu_sc…
-
Creating a multi-leader microservice using Paxos with emojis as identifiers
would require encoding the essential Paxos concepts and processes into your
emoji-based system. Here's a simplified represen…
-
With following command:
```
make prod NO_DOCKER=true
```
the errror messages:
```
ERROR in ./node_modules/@akeneo-pim-community/communication-channel/src/components/index.tsx
Module not fou…
-
### Describe the issue
Issue:
We collect a large-scale instruction dataset, and want to use muti-nodes training. When using the following script, the traing time is too slow and no log about time.
…
-
Given there is already support for nccl, whats the overhead to add support for multi node gpu support for training/inference
-
The high-level architecture diagram seems more up-to-date than the text around it:
> The RChain Network implements direct node-to-node communication, where each node runs the RChain platform and a …
-
Hi,
I was looking for a micro-services tool for Node.js and found Seneca, which seems a really nice fit for this.
However, I want to run those microservices in multiple docker containers spread over …