Closed aklemen closed 4 months ago
Update: We observed couple of code changes required with this script due to recent updates during the model and transcription refactoring. @karpov-nick is working to provide a fix for this.
There is a work in progress in the PR https://github.com/NVIDIA/NeMo/pull/8428
Thank you both!
You can try decoding without N-gram at the branch karpnv/beamsearch with parameters
python3 ./scripts/asr_language_modeling/ngram_lm/eval_beamsearch_ngram_ctc.py \
model_path=./am_model.nemo \
dataset_manifest=./manifest.json \
preds_output_folder=/tmp \
ctc_decoding.strategy=flashlight \
ctc_decoding.beam.nemo_kenlm_path="" \
ctc_decoding.beam.beam_size=[4] \
ctc_decoding.beam.beam_beta=[0.5]
This issue is stale because it has been open for 30 days with no activity. Remove stale label or comment or this will be closed in 7 days.
This issue was closed because it has been inactive for 7 days since being marked as stale.
Describe the bug
I am trying to use an external LLM to rescore the results of beam search from Conformer-CTC model.
When trying to get the beam search results with the
eval_beamsearch_ngram_ctc.py
without passing the N-gram LM, I get the following error:Steps/Code to reproduce bug
Expected behavior
I would expect the error to not be thrown as
BeamSearchDecoderWithLM
actually handles the case when the path to N-gram LM is not passed:When I removed the check for the KenLM file path from
nemo/collections/asr/parts/submodules/ctc_beam_decoding.py
, it worked:Environment overview
python -m pip install git+https://github.com/NVIDIA/NeMo.git@v1.23.0#egg=nemo_toolkit[all]
Environment details
Additional context
GPU: T4