By definition, synthesized speech is "synthetic" sound. This is"electro - mechanical", and not natural. In reality it is "sound" that is synthesized to resemble speech. Just as sound can be synthesized to resemble musical instruments. Accomplished by mathematic programming because recorded voice WAV Based Systems take up disk storage space and processor power.
Synthesis To Sound.
To the human ear and expectations music instruments are simple wave forms compared to the human voice. Music synthesis can synthesize the generic sound of a music instrument. Music synthesizers still can only make an approximation to reproduce the individual sounds of a specific music instrument ; or a specific musician. In speech the complication is with the human voice which is both instrument and musician.
Synthesis To Speech.
The human voice is complex and further complicated by the human factors of individual preference and perception. Even before pronunciation and enunciation, there are numerous factors, and sub - factors, that go into what we "perceive" as making speech natural.
For example even with today's imperfections in synthesized speech:
For now, like it or not, there is no Natural Sounding Text To Speech available for PCs. If you want to improve the currently imperfect text -to- speech applications your choices are:
To Train Or Not To Train.
In Computer Speech Command and Control, Recognition and Dictation, both you and/or your computer program will need to "adjust or train". Even the newest out of the box "self training" programs require user input.
Self training systems require user input to enhance the program's learning process. Adding text vocabulary as well as spell checking and correcting are functions which enhance greater program accuracy.
Out of the box only the simplest single purpose programs with limited and preset vocabularies require no further training.
A few will allow you to add a limited number of your own voice command controls. The simplest may add your commands phonetically and require no further voice training.
2. The simplest single purpose discrete speech dictation [Note Pads] programs with a limited preset vocabulary may require you to train your voice pronunciation to the program's expectations. As they have preset, and or limited, voice dictation word vocabularies.
A few will allow you to add a limited number of your own voice vocabulary words. The simplest may add your vocabulary words phonetically and require no further voice training. Others will require the addition of your own voice pronunciation to train the program.
Not All Dictation Programs Are Created Equal.
All dictation programs will enter spoken text into your computer. Not all enter spoken text to the same place or application in your computer. Some will only enter to Clipboard, Note pad, Write, Word, or WordPerfect. The most versatile will enter text directly into any Windows compatible application.
Not all dictation programs will provide Speech Command and Control over your computer applications and operations. This an important option that can substantially reduce or eliminate the necessity of keyboard and mouse input. Full voice input [without the necessity of keyboard or mouse] is currently available to: * IBM ViaVoice Gold *, and * Dragon Systems Dragon Dictate *.
In speech dictation you talk text into your computer. To read text out of your computer you need a text -to- speech [screen reader] synthesizer. Compatibility between programs is critical as these are two different multi-tasking programs, with two different purposes. Together they form a feedback loop with the user.
* Top * | * ACSP Home * | * SuperAdaptoid Column *