-
See also: [Topic: MDIB/MDS Modeling for Device Aggregators](https://confluence.hl7.org/x/QSsvBQ)
-
The most obvious way to retrieve model scale (atomic/coarse-grained/multiscale (a mix of atomic and coarse-grained)) is from the `_ihm_model_representation_details` table. However, the scale is also e…
-
This issue tracks the progress on improving the handling and testing of Vision-Language Models. The main goals are to enhance/enable generation tests, handle other generation techniques like assisted …
-
When a computed is a simple alias over another property or link (even a backlink), we should allow mutating it. This will remove one of the inconveniences of modeling one-to-many as many-to-one with a…
-
-
**Describe the question(问题描述)**
Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts
In the MOE method does expert have to learn and can the frozen model be us…
-
请提供下述完整信息以便快速定位问题/Please provide the following information to quickly locate the problem
- 系统环境/System Environment:Ubuntu 20.04 环境为paddle官方docker
- 版本号/Version:paddlepaddle/paddle:2.5.2-gpu…
-
Hi @matteofrigo @rutgerfick We are trying to follow this example: (https://nbviewer.jupyter.org/github/AthenaEPI/dmipy/blob/master/examples/example_multi_compartment_spherical_mean_technique.ipynb) bu…
-
I tryed to modify your example code to run this model on lowvram card by BNB 4bit or 8bit quantization config.
While use bnb 4bit config like below:
```python
qnt_config = BitsAndBytesConfig(load…
-
Hi all,
I am trying to follow the instruction for [INT8 weight only + INT8 KV cache](https://github.com/NVIDIA/TensorRT-LLM/tree/release/0.5.0/examples/llama#int8-weight-only--int8-kv-cache) for Ll…