This has become my problem for a while now. After tons of google search on how to implement Speech Recognition in Unity 3 or 4, I end up finding no answers. So I’m asking for the last time if my research is still lacking and I’ve missed the recent updates, is it possible to implement Speech Recognition into Unity?
On the cheap side, there isn’t much. I think someone has a more expensive solution, but I forgot their name right now. Basically you have the phoneme problem which looks better if it’s vertex animated rather than bone, and the speech recognition, which isn’t that advanced. Sapi-lipsync is the only free version I know about. I don’t think it’s fast enough for real time, but it might be. I think I’m going to experiment with the volume based solution that I’ve seen around with random bone animated mouth motion. Your eye really doesn’t care that much as long as the mouth closes when it should. I did some tests a while back and I couldn’t tell the difference as long as the mouth was shut during the quiet times. You can’t just do that open/close thing with volume, because that looks pretty bad, but if you add random lip motion, it looks all right. I guess the thing is, how much money or time do you want to put into it. I don’t want to put too much in either one.
To make myself clear, what I want is the Speech-to-text function. My game is just for school project purposes and will be designed just for Windows OS. Can I use the .NET Framework Speech-to-text capability and import it into Unity since my game will just run in Windows OS?