/external/chromium_org/chrome/browser/speech/ |
H A D | tts_platform.cc | 26 void TtsPlatformImpl::WillSpeakUtteranceWithVoice(const Utterance* utterance, argument
|
H A D | tts_android.cc | 42 const std::string& utterance, 50 base::android::ConvertUTF8ToJavaString(env, utterance).Release(), 56 utterance_ = utterance; 40 Speak( int utterance_id, const std::string& utterance, const std::string& lang, const VoiceData& voice, const UtteranceContinuousParameters& params) argument
|
H A D | tts_message_filter.cc | 86 scoped_ptr<Utterance> utterance(new Utterance(browser_context_)); 87 utterance->set_src_id(request.id); 88 utterance->set_text(request.text); 89 utterance->set_lang(request.lang); 90 utterance->set_voice_name(request.voice); 91 utterance->set_can_enqueue(true); 97 utterance->set_continuous_parameters(params); 99 utterance->set_event_delegate(weak_ptr_factory_.GetWeakPtr()); 101 TtsController::GetInstance()->SpeakOrEnqueue(utterance.release()); 119 void TtsMessageFilter::OnTtsEvent(Utterance* utterance, argument [all...] |
H A D | tts_controller_impl.cc | 135 void TtsControllerImpl::SpeakOrEnqueue(Utterance* utterance) { argument 136 // If we're paused and we get an utterance that can't be queued, 138 if (paused_ && !utterance->can_enqueue()) { 141 delete utterance; 145 if (paused_ || (IsSpeaking() && utterance->can_enqueue())) { 146 utterance_queue_.push(utterance); 149 SpeakNow(utterance); 153 void TtsControllerImpl::SpeakNow(Utterance* utterance) { argument 157 GetPlatformImpl()->LoadBuiltInTtsExtension(utterance->browser_context()); 161 GetVoices(utterance 347 Utterance* utterance = utterance_queue_.front(); local 355 Utterance* utterance = utterance_queue_.front(); local 381 GetMatchingVoice( const Utterance* utterance, std::vector<VoiceData>& voices) argument [all...] |
H A D | tts_linux.cc | 39 const std::string& utterance, 79 // These apply to the current utterance only. 167 const std::string& utterance, 195 utterance_ = utterance; 198 if (libspeechd_loader_.spd_say(conn_, SPD_TEXT, utterance.c_str()) == -1) { 165 Speak( int utterance_id, const std::string& utterance, const std::string& lang, const VoiceData& voice, const UtteranceContinuousParameters& params) argument
|
/external/chromium_org/chrome/browser/resources/network_speech_synthesis/ |
H A D | tts_extension.js | 50 * The arguments passed to the onSpeak event handler for the utterance 54 * @type {?{utterance: string, options: Object, callback: Function}} 108 * url for the requested speech utterance. Sets that url as the source 110 * @param {string} utterance The text to be spoken. 115 onSpeak_: function(utterance, options, callback) { 116 // Truncate the utterance if it's too long. Both Chrome's tts 118 // maximum limit for an utterance. 119 if (utterance.length > 32768) 120 utterance = utterance [all...] |
/external/chromium_org/content/public/common/ |
H A D | speech_recognition_result.h | 17 base::string16 utterance; member in struct:content::SpeechRecognitionHypothesis 24 : utterance(utterance_value),
|
/external/chromium_org/third_party/WebKit/Source/modules/speech/testing/ |
H A D | PlatformSpeechSynthesizerMock.cpp | 77 void PlatformSpeechSynthesizerMock::speak(PlatformSpeechSynthesisUtterance* utterance) argument 80 m_utterance = utterance;
|
/external/chromium_org/third_party/WebKit/Source/platform/exported/ |
H A D | WebSpeechSynthesisUtterance.cpp | 36 WebSpeechSynthesisUtterance::WebSpeechSynthesisUtterance(PlatformSpeechSynthesisUtterance* utterance) argument 37 : m_private(utterance) 41 WebSpeechSynthesisUtterance& WebSpeechSynthesisUtterance::operator=(PlatformSpeechSynthesisUtterance* utterance) argument 43 m_private = utterance;
|
H A D | WebSpeechSynthesizerClientImpl.cpp | 52 void WebSpeechSynthesizerClientImpl::didStartSpeaking(const WebSpeechSynthesisUtterance& utterance) argument 54 m_client->didStartSpeaking(utterance); 57 void WebSpeechSynthesizerClientImpl::didFinishSpeaking(const WebSpeechSynthesisUtterance& utterance) argument 59 m_client->didFinishSpeaking(utterance); 62 void WebSpeechSynthesizerClientImpl::didPauseSpeaking(const WebSpeechSynthesisUtterance& utterance) argument 64 m_client->didPauseSpeaking(utterance); 67 void WebSpeechSynthesizerClientImpl::didResumeSpeaking(const WebSpeechSynthesisUtterance& utterance) argument 69 m_client->didResumeSpeaking(utterance); 72 void WebSpeechSynthesizerClientImpl::speakingErrorOccurred(const WebSpeechSynthesisUtterance& utterance) argument 74 m_client->speakingErrorOccurred(utterance); 77 wordBoundaryEventOccurred(const WebSpeechSynthesisUtterance& utterance, unsigned charIndex) argument 82 sentenceBoundaryEventOccurred(const WebSpeechSynthesisUtterance& utterance, unsigned charIndex) argument [all...] |
/external/chromium_org/third_party/WebKit/Source/platform/speech/ |
H A D | PlatformSpeechSynthesizer.cpp | 57 void PlatformSpeechSynthesizer::speak(PlatformSpeechSynthesisUtterance* utterance) argument 62 m_webSpeechSynthesizer->speak(blink::WebSpeechSynthesisUtterance(utterance));
|
/external/chromium_org/chrome/browser/chromeos/accessibility/ |
H A D | speech_monitor.cc | 42 const std::string& utterance, 49 static_cast<int>(utterance.size()), 74 void SpeechMonitor::WillSpeakUtteranceWithVoice(const Utterance* utterance, argument 76 utterance_queue_.push_back(utterance->text()); 40 Speak( int utterance_id, const std::string& utterance, const std::string& lang, const VoiceData& voice, const UtteranceContinuousParameters& params) argument
|
H A D | spoken_feedback_browsertest.cc | 147 const std::string& utterance = speech_monitor_.GetNextUtterance(); local 148 if (utterance == expected_utterance) 201 std::string utterance = speech_monitor_.GetNextUtterance(); local 202 VLOG(0) << "Got utterance: " << utterance; 203 if (utterance == "Bookmarks,") 278 std::string utterance = speech_monitor_.GetNextUtterance(); local 279 VLOG(0) << "Got utterance: " << utterance; 280 if (utterance 301 std::string utterance = speech_monitor_.GetNextUtterance(); local 317 std::string utterance = speech_monitor_.GetNextUtterance(); local 337 std::string utterance = speech_monitor_.GetNextUtterance(); local [all...] |
/external/chromium_org/chrome/browser/speech/extension_api/ |
H A D | tts_engine_extension_api.cc | 140 void TtsExtensionEngine::Speak(Utterance* utterance, argument 143 // utterance around and track it. If not, we're finished with this 144 // utterance now. 148 args->AppendString(utterance->text()); 153 utterance->options()->DeepCopy())); 176 args->AppendInteger(utterance->id()); 180 Profile* profile = Profile::FromBrowserContext(utterance->browser_context()); 183 ->DispatchEventToExtension(utterance->extension_id(), event.Pass()); 186 void TtsExtensionEngine::Stop(Utterance* utterance) { argument 190 Profile* profile = Profile::FromBrowserContext(utterance 196 Pause(Utterance* utterance) argument 208 Resume(Utterance* utterance) argument [all...] |
H A D | tts_extension_apitest.cc | 57 const std::string& utterance, 100 const std::string& utterance, 108 false, utterance_id, TTS_EVENT_END, utterance.size(), 115 const std::string& utterance, 123 true, utterance_id, TTS_EVENT_END, utterance.size(), std::string()), 128 const std::string& utterance, 132 for (int i = 0; i < static_cast<int>(utterance.size()); i++) { 133 if (i == 0 || utterance[i - 1] == ' ') { 239 // One utterance starts speaking, and then a second interrupts. 245 // Expect the second utterance an 99 SendEndEvent(int utterance_id, const std::string& utterance, const std::string& lang, const VoiceData& voice, const UtteranceContinuousParameters& params) argument 113 SendEndEventWhenQueueNotEmpty( int utterance_id, const std::string& utterance, const std::string& lang, const VoiceData& voice, const UtteranceContinuousParameters& params) argument 127 SendWordEvents(int utterance_id, const std::string& utterance, const std::string& lang, const VoiceData& voice, const UtteranceContinuousParameters& params) argument [all...] |
H A D | tts_extension_api.cc | 83 // One of these is constructed for each utterance, and deleted 84 // when the utterance gets any final event. 89 virtual void OnTtsEvent(Utterance* utterance, 95 void TtsExtensionEventHandler::OnTtsEvent(Utterance* utterance, argument 99 if (utterance->src_id() < 0) { 100 if (utterance->finished()) 106 utterance->desired_event_types(); 109 if (utterance->finished()) 122 details->SetInteger(constants::kSrcIdKey, utterance->src_id()); 123 details->SetBoolean(constants::kIsFinalEventKey, utterance 267 Utterance* utterance = new Utterance(GetProfile()); local [all...] |
/external/chromium_org/chrome/renderer/ |
H A D | tts_dispatcher.cc | 65 TtsUtteranceRequest utterance; local 66 utterance.id = id; 67 utterance.text = web_utterance.text().utf8(); 68 utterance.lang = web_utterance.lang().utf8(); 69 utterance.voice = web_utterance.voice().utf8(); 70 utterance.volume = web_utterance.volume(); 71 utterance.rate = web_utterance.rate(); 72 utterance.pitch = web_utterance.pitch(); 73 RenderThread::Get()->Send(new TtsHostMsg_Speak(utterance)); 113 WebSpeechSynthesisUtterance utterance 121 WebSpeechSynthesisUtterance utterance = FindUtterance(utterance_id); local 130 WebSpeechSynthesisUtterance utterance = FindUtterance(utterance_id); local 138 WebSpeechSynthesisUtterance utterance = FindUtterance(utterance_id); local 148 WebSpeechSynthesisUtterance utterance = FindUtterance(utterance_id); local 159 WebSpeechSynthesisUtterance utterance = FindUtterance(utterance_id); local 172 WebSpeechSynthesisUtterance utterance = FindUtterance(utterance_id); local 182 WebSpeechSynthesisUtterance utterance = FindUtterance(utterance_id); local 193 WebSpeechSynthesisUtterance utterance = FindUtterance(utterance_id); local [all...] |
/external/chromium_org/third_party/WebKit/Source/modules/speech/ |
H A D | SpeechSynthesis.cpp | 82 // If we have a current speech utterance, then that means we're assumed to be in a speaking state. 83 // This state is independent of whether the utterance happens to be paused. 101 SpeechSynthesisUtterance* utterance = currentSpeechUtterance(); local 102 ASSERT(utterance); 104 utterance->setStartTime(monotonicallyIncreasingTime()); 106 m_platformSpeechSynthesizer->speak(utterance->platformUtterance()); 109 void SpeechSynthesis::speak(SpeechSynthesisUtterance* utterance, ExceptionState& exceptionState) argument 111 if (!utterance) { 112 exceptionState.throwTypeError("Invalid utterance argument"); 116 m_utteranceQueue.append(utterance); 145 fireEvent(const AtomicString& type, SpeechSynthesisUtterance* utterance, unsigned long charIndex, const String& name) argument 151 handleSpeakingCompleted(SpeechSynthesisUtterance* utterance, bool errorOccurred) argument 174 boundaryEventOccurred(PlatformSpeechSynthesisUtterance* utterance, SpeechBoundary boundary, unsigned charIndex) argument 191 didStartSpeaking(PlatformSpeechSynthesisUtterance* utterance) argument 197 didPauseSpeaking(PlatformSpeechSynthesisUtterance* utterance) argument 204 didResumeSpeaking(PlatformSpeechSynthesisUtterance* utterance) argument 211 didFinishSpeaking(PlatformSpeechSynthesisUtterance* utterance) argument 217 speakingErrorOccurred(PlatformSpeechSynthesisUtterance* utterance) argument [all...] |
/external/chromium_org/content/browser/speech/ |
H A D | google_one_shot_remote_engine.cc | 32 const char* const kUtteranceString = "utterance"; 125 base::string16 utterance; local 127 if (!hypothesis_value->GetString(kUtteranceString, &utterance)) { 128 LOG(WARNING) << "ParseServerResponse: Missing utterance value."; 135 result->hypotheses.push_back(SpeechRecognitionHypothesis(utterance,
|
/external/srec/srec/Recognizer/include/ |
H A D | SR_RecognizerImpl.h | 183 CA_Utterance* utterance; member in struct:SR_RecognizerImpl_t
|
/external/chromium_org/third_party/android_platform/webview/ |
H A D | frameworks.jar | META-INF/ META-INF/MANIFEST.MF android/ android/location/ android/location/Address$1.class ... |