-
Notifications
You must be signed in to change notification settings - Fork 283
Open
Description
I am working on a project that involves animating a chatbot, and wish to use rhubarb to animate the lip-syncing. However, after testing with both the PocketSphinx and Phonetic recognizers, neither seem to register individual phonemes, instead only resulting a rest output. This is the output I have consistently gotten testing on AI voices:
0.00 X
It should be said that in the same environment, rhubarb does work for recognizing recordings of my own voice, so I assume it has to do with the structure of AI generated audio files themselves. I wonder if support could possibly be added for tweaking sensitivity on the user's end, or if adding support for AI voices may require supporting an entirely new speech recognizer.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels