In addition to Weibo, there is also WeChat
Please pay attention
WeChat public account
Shulou
2025-02-20 Update From: SLTechnology News&Howtos shulou NAV: SLTechnology News&Howtos > Development >
Share
Shulou(Shulou.com)06/03 Report--
This article is about the usefulness of XML in speech synthesis. The editor thinks it is very practical, so share it with you as a reference and follow the editor to have a look.
The Internet and everything associated with it now seems to be everywhere. You may have tried to get a voice call from a night telemarketer or a prescription notice from your local drugstore. Now, there is a new technology that can use speech synthesis combined with xml technology to transmit voice messages.
The means of sending messages by voice is nothing new. It is a method of communication that we have been using for thousands of years. Moreover, receiving a telephone from a computer is not a new invention. Many voice technologies have become popular today, from fax machines and automatic dialers to integrated voice response systems (IVR). The telephone is, of course, its most common application.
Traditional voice systems use pre-recorded samples, dictionaries and phonemes to create the sounds we hear. However, there are many problems with using this pre-recording method. One of the most common problems is the lack of coherence and change. If there is only one recorded voice version, in which there is only one sample for each word or sound, it is difficult for the computer to issue questions that are different from ordinary statements. It is also difficult for the computer to know when to pronounce in a certain tone or what tone to pronounce.
To help solve the problem of speech synthesis, W3C has created a new working draft for speech Synthesis markup language (Speech Synthesis Markup Language). This new XML vocabulary allows voice browser developers to control how a voice synthesizer is created. For example, developers can include commands in the volume and use it when synthesizing voice patterns.
The SSML specification is based on an early research work by Sun called jspeeck Markup Language (JSML). JSML is based on java Speech API Markup Language. SSML is now the working paper of the W3C Voice Research working Group.
The basic goal of the SSML language is a text-to-speech (Text-To-Speech for short TTS) processor. A TTS engine takes a collection of text and converts it into speech. There are already several TTS applications, such as telephone voice synthesis and reply systems, as well as more advanced systems designed for the blind. The inherent uncertainty of the pronunciation of a specific text set is one of the main problems faced by existing TTS systems. Other common problems focus on the pronunciation of parts of speech such as abbreviations of words (such as HTML) and words with different spelling and pronunciation (such as subpoena).
The basic elements of the SSML language specify the format of the text. For example, it provides a paragraph element for the HTML,SSML language and goes further. Because it also provides sentence elements. By specifying the address of the sentence like a paragraph, including the start address and the end address, the TTS engine can generate speech more accurately.
In addition to the basic format, SSML also provides functionality to specify how to pronounce a predetermined word or collection of words. This function is implemented by the "say-as" element. It is a very useful component in SSML. It allows you to specify a template that describes how to pronounce a word or collection of words. With "say-as", we can specify how to pronounce abbreviated words, or we can specify pronunciation for words whose spelling and pronunciation are different. We can also list the differences between numbers and dates. The "say-as" element contains support for email addresses, currencies, phone numbers, and so on.
We can also provide a phonological way of expression for the text. For example, we can use this means to point out the differences in pronunciation of potato words between American English and British English.
Several advanced attributes of the SSML language can help us make TTS systems generate more human sounds. We can use the "voice" element to specify a male, female, or neutral voice, and we can also specify the age of the voice. We can use this element to specify any sound from a 4-year-old boy to a 75-year-old woman.
We can also use the "emphasis" element to surround text that needs to be emphasized or compared. We can also use the "break" element to tell the system that voice should be paused somewhere.
One of the most advanced features of the SSML language is reflected in its "PRosody" element. Through it, we can generate the voice of a certain set of text in a specified way. We can specify the intonation, range and speed of the sound (words per minute). We can even specify something more detailed by using the "contour" element, which integrates intonation and speed. By specifying the value of the "contour" element of a text collection, we can define how speech is generated more precisely.
Thank you for reading! This is the end of the article on "what is the use of XML in speech synthesis". I hope the above content can be of some help to you, so that you can learn more knowledge. if you think the article is good, you can share it for more people to see!
Welcome to subscribe "Shulou Technology Information " to get latest news, interesting things and hot topics in the IT industry, and controls the hottest and latest Internet news, technology news and IT industry trends.
Views: 0
*The comments in the above article only represent the author's personal views and do not represent the views and positions of this website. If you have more insights, please feel free to contribute and share.
Continue with the installation of the previous hadoop.First, install zookooper1. Decompress zookoope
"Every 5-10 years, there's a rare product, a really special, very unusual product that's the most un
© 2024 shulou.com SLNews company. All rights reserved.