FreeTTS - Speech Synthesizer in Java
FreeTTS is a speech synthesis system written entirely in the Java. It is based upon Flite, a small run-time speech synthesis engine developed at Carnegie Mellon University. Flite is derived from the Festival Speech Synthesis System from the University of Edinburgh and the FestVox project from Carnegie Mellon University. FreeTTS supports a subset of the JSAPI 1.0 java speech synthesis specification.
comments powered by Disqus
Dragonfly is a speech recognition framework. It is a Python package which offers a high-level object model and allows its users to easily write scripts, macros, and programs which use speech recognition. It currently supports the following speech recognition engines: Dragon NaturallySpeaking (DNS), a product of Nuance, Window Speech Recognition (WSR), as included in Microsoft Windows Vista.
Wiki-to-Speech is the former name of the SlideSpeech project. Roughly speaking, the original project was the Open Allure Dialog System which focused on a desktop application, this project, Wiki-to-Speech, focused on a mobile application and finally SlideSpeech focused on a web application. IntroductionCompanion Android/Python wiki-to-speech project. Drive text-to-speech interaction using wiki-based scripts. Scripts can include statements, question/answer/response (multiple choice), links to webs
Voxforge - Free GPL Speech Corpus and Acoustic Model Repository for Open Source Speech Recognition E
VoxForge was set up to collect speech audio files to create a GPL Speech Corpus for use with Free and Open Source Speech Recognition Engines (on Linux and Windows). The transcribed speech will be 'compiled' into acoustic models for use with Open Source speech recognition engines such as Julius, ISIP, and Sphinx, and HTK (note that HTK has distribution restrictions). Why Do We Need Free GPL Speech Audio?Most acoustic models used by 'Open Source' Speech Recognition engines are closed source. They
SpeakRight is an Java framework for writing speech recognition applications in VoiceXML. Dynamic generation of VoiceXML is done using the popular StringTemplate templating framework. Although VoiceXML uses a similar web architecture as HTML, the needs of a speech app are very different. SpeakRight lives in application code layer, typically in a servlet. The SpeakRight runtime dynamically generates VoiceXML pages, one per HTTP request.
Project Name: VOICE ENABLED INTERACTIVE LEARNING MADE BY: AJAY KUMAR PROJECT DESCRIPTION: This project includes 3 parts:- 1:Speech Synthesis 2:Speech Recognition 3:Speech Analysis 1: Speech Synthesis: In this part it takes text as input and voice as output.You can open any text file or doc file,it will read for u. 2: Speech Recognition: In this part it takes speech as input and text as an ouput.Whatever you speak it will print on the screen. 3.Speech Analysis: It has two parts a. waveform creati
speech.pyspeech.py is a Python module that provides a clean interface to Windows's voice recognition and text-to-speech capabilities. It's very easy to use within a program that needs to listen for specific phrases or general speech, or that needs to speak. It is available on PyPI at http://pypi.python.org/pypi/speech/ . Questions? You can contact Michael Gundlach, the developer, at gundlach at gmail. Or, just file a bug under "Issues" above. Example codeHere's a very simple program that repeats
eSpeak is a compact open source software speech synthesizer for English and other languages. eSpeak uses a formant synthesis method. This allows many languages to be provided in a small size. It supports SAPI5 version for Windows, so it can be used with screen-readers and other programs that support the Windows SAPI5 interface. It can translate text into phoneme codes, so it could be adapted as a front end for another speech synthesis engine.
Festival offers a general framework for building speech synthesis systems as well as including examples of various modules. It offers full text to speech through a APIs via shell and though a Scheme command interpreter. It has native support for Apple OS. It supports English and Spanish languages.
AxTk is a toolkit for building highly accessible applications with speech output. AxTk is built on top of wxWidgets and so is cross-platform. The developer can opt to speech-enable an existing wxWidgets UI, or use a new menu-based interface which is easier for a vision impaired user. AxTk also contains a text to speech wrapper class, wxTextToSpeech, with handlers for a variety of speech engines including SAPI, Mac Speech Synthesis Manager, eSpeak and Cepstral. wxTextToSpeech can be used independ
The Jave speech toolkit (JSTK) provides a native implementation of both library/API and applications for speech recognition, speaker verification, speech visualization (including transcription tools), and evaluation of related human rater tasks. It is maintained by the speech group at the Univ. Erlangen-Nuremberg (www5.cs.fau.de) JSTK - The Java Speech ToolkitThe Java Speech Tooklit (JSTK) is developed and maintained by the Speech Group at the University of Erlangen-Nuremberg. It is designed to