As of today a large change of using SWIG-generated python bindings has been merged into pocketsphinx and sphinxbase trunk.
SWIG is an interface compiler that connects programs written in C and C++ with scripting languages such as Perl, Python, Ruby, and Tcl. It works by taking the declarations found in C/C++ header files and using them to generate the wrapper code that scripting languages need to access the underlying C/C++ code. In addition, SWIG provides a variety of customization features that let you tailor the wrapping process to suit your application.
By this port we hope to increase coverage of pocketsphinx bindings and provide a uniform and documented interface in various language: Python, Ruby, Java.
To test the change checkout sphinxbase and pocketsphinx from trunk and see the examples in pocketsphinx/swig/python/test.
It is an old idea to implement an open source dictation tool everyone could use. Without servers, networking, without the need to share your private speech with someone else. This is certainly not a trivial project which was started many times, but it's something really world-changing. Now, it's live again, powered by CMUSphinx.
Consider details about ongoing efforts of Simon project to implement open source dictation.
It has been a long dream to voice-enable websites. However, no good technology existed for this either because speech recognition on the web required a connection to a server or due to the requirement to install binary plugin.
Great news is that you can now use CMUSphinx in any modern browser completely on the client side. No need for installation, no need to maintain voice recognition server farm. This is a really cool technology.
Sylvain Chevalier has been working on a port of Pocketsphinx to JavaScript using emscripten. Combined with the Web Audio API, it works great as a real-time recognizer for web applications, running entirely in the browser, without plug-in.
It's on Github (https://github.com/syl22-00/pocketsphinx.js),
comments, suggestions and contributions are more than welcome!
Months ago, Mark Shuttleworth announced and explained Ubuntu's converged vision, where a singular OS is to power phones, tablets, desktops, TVs, etc.
Presently, the Ubuntu developers are working on strengthening Ubuntu for phones (Ubuntu Touch), development where speech recognition is to probably play a relevant role.
Among the usage-cases, the speech recognition was demoed (at a mockup level) as part of the HUD 2.0, basically, allowing the user to trigger commands by pressing a button and speaking into the phone's microphone, oral command translated and applied similarly to a regular command.
Great news of today is that PocketSphinx has just landed in Ubuntu 13.10 by default being shifted from its previous universe availability (via Ubuntu Software Center) directly into main (landing via the regular updates). This means PocketSphinx is to be utilized for the upcoming Unity 8's release on the desktop, probably to allow users to fully grasp the Unity 8's features via a full spectrum of functionalities.
It's definitely just a beginning of the work but it's really great to see CMUSphinx on its way to the desktop. Definitely there will be many problems on the way since a proper implementation of the speech recognition system is not a trivial task and needs certain expertise. Your help is needed here otherwise all the issues will be assigned to Pocketsphinx.