Open albanoandrea opened 4 years ago
Hi @albanoandrea Sorry about the late response. The solution available on GitHub is for evaluation purpose and uses Espressif's custom WW, which is not certified by Amazon. Usually in a commercial product, we get this WW length from an external DSP (Amazon certified). Hence here we are not sending WW length as it will anyway be rejected by the cloud. DSP's certified firmwares requires an NDA.
Hi Amit, don't worry, we have addressed and fixed the issue with your team.
Hi, we are building our custom device using aia_beta branch.
When WW is recognized, amazon asks to send the audio to the server from a starting point before the Alexa WW is detected, so that the server can verify the WW to and decide if "Alexa" was good enough and maybe not commercial on TV.
I've noticed that in the state machine in va_dsp, there is the WW event that should make the work, but the phrase_length is missing.
So I store somewhere the ww length and when requested I give it to the state machine using the get_ww_length e function:
And here the conresponding log:
When we call the TAP_TO_TALK event this is the log of a call:
It seems to me that the server always reject the request, while with tap to talk it Thinks and then give the answer.
Do you have ever used that feature and you know it's working? There is some advice we should follow in order to use it? (e.g. we pass as phrase_length the duration in number of samples estimated by the WuW engine, but maybe we need to pass the size in byte).
Thanks, Andrea