-
System Info
GPU: NVIDIA RTX 4090
TensorRT-LLM 0.13
quest 1: How can I use the OpenAPI to perform inference on a TensorRT engine model?
root@docker-desktop:/llm/tensorrt-llm-0.13.0/examples/apps# pyt…
-
Hi, I'm new to active learning and modAL. I want to fine-tune a RoBERTa model on a (yet) unannotated dataset with active learning. Has anyone tried applying modAL to huggingface models like RoBERTa or…
-
Does ace-compiler support current large language models such as BERT? Thank you.
-
**Is your feature request related to a problem?**
It is not really a problem but it would make the code working in a more efficiently way, in particular in other languages.
**What is the expected…
-
Thanks for releasing the code!
I am trying to run `QuestionAnsweringSquadDiffMaskAnalysis.ipynb`. but I get this error when load_from_checkpoint()
`RuntimeError: Error(s) in loading state_dict for B…
-
Hi, I was trying to adapt K-BERT for RoBERTa and tried using the pre-trained model for RoBERTa from Huggingface for that. But somehow, the model never seems to converge at all and gives very poor scor…
-
I propose adding a Model Evaluation and Benchmarking System to ML Nexus to help users assess their model performance on standardized datasets and compare it against benchmarked scores. This feature wo…
-
Hello, I was wondering if we can replace the BERT for current supported models
```python
Languages.JP: BASE_DIR / "bert" / "deberta-v2-large-japanese-char-wwm",
Languages.EN: BASE_DIR / "…
-
Hi,
I'd like to compile `projects/pt1/examples/torchscript_stablehlo_backend_tinybert.py` into torch-mlir, so I did the modification:
```
--- a/projects/pt1/examples/torchscript_stablehlo_backend…
-
I am not sure how to pass the inputs to the BERT Squad model session in onnx web runtime.
these are the inputs:
unique_ids_raw_output___9:0
name: unique_ids_raw_output___9:0
type: int64[unk__475]…