Open mknw opened 4 weeks ago
Very interesting use!
If you are running on MacOS you can do some of this today with the built-in speech to text. Make sure dictation is enabled under System Preferences -> Keyboard -> Dictation (toggle to on) and set a dictation keyboard shortcut if desired (defaults to Fn+F5).
A more integrated approach would likely leverage one or more of the following:
Check for existing issues
Describe the feature
I started using the inline assistant tool more and more, for both longer and shorter tasks. There are a lot of advantages in it.
I would like to be able to dictate to the agent, while seeing what I type and being able to approve/edit the input before submitting the instructions to the LLM model.
This could be visualised in the inline assistant element (⌃ctrl+↩return) as a microphone icon next to the "configure" icon to the left. It would have the benefit to shorten the time to give directions to the assistant, while providing one with the ability to adjust the input to correct variable naming and similar mistakes.
Additionally, I would suggest adding keyboard shortcuts to:
Nice to have
It would be nice if dictation itself was aware of the context (instead of only submitting the context while submitting the prompt). This would make it easier for the Voice-to-Text model to pick up variable names and other code specific keywords.