blob: b9ca7065fb3fd1690ff0497a598261f39b7bb360 [file] [log] [blame]
/*
* Copyright (C) 2012 Google Inc. All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* * Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* * Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS ``AS IS'' AND ANY
* EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
* CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
* EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
* PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
* OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#include "modules/speech/SpeechRecognition.h"
#include "bindings/core/v8/ExceptionState.h"
#include "core/dom/Document.h"
#include "core/dom/ExceptionCode.h"
#include "core/page/Page.h"
#include "modules/mediastream/MediaStreamTrack.h"
#include "modules/speech/SpeechRecognitionController.h"
#include "modules/speech/SpeechRecognitionError.h"
#include "modules/speech/SpeechRecognitionEvent.h"
namespace blink {
SpeechRecognition* SpeechRecognition::create(ExecutionContext* context) {
ASSERT(context && context->isDocument());
Document* document = toDocument(context);
ASSERT(document);
SpeechRecognition* speechRecognition =
new SpeechRecognition(document->page(), context);
speechRecognition->suspendIfNeeded();
return speechRecognition;
}
void SpeechRecognition::start(ExceptionState& exceptionState) {
if (!m_controller)
return;
if (m_started) {
exceptionState.throwDOMException(InvalidStateError,
"recognition has already started.");
return;
}
m_finalResults.clear();
m_controller->start(this, m_grammars, m_lang, m_continuous, m_interimResults,
m_maxAlternatives, m_audioTrack);
m_started = true;
}
void SpeechRecognition::stopFunction() {
if (!m_controller)
return;
if (m_started && !m_stopping) {
m_stopping = true;
m_controller->stop(this);
}
}
void SpeechRecognition::abort() {
if (!m_controller)
return;
if (m_started && !m_stopping) {
m_stopping = true;
m_controller->abort(this);
}
}
void SpeechRecognition::didStartAudio() {
dispatchEvent(Event::create(EventTypeNames::audiostart));
}
void SpeechRecognition::didStartSound() {
dispatchEvent(Event::create(EventTypeNames::soundstart));
}
void SpeechRecognition::didStartSpeech() {
dispatchEvent(Event::create(EventTypeNames::speechstart));
}
void SpeechRecognition::didEndSpeech() {
dispatchEvent(Event::create(EventTypeNames::speechend));
}
void SpeechRecognition::didEndSound() {
dispatchEvent(Event::create(EventTypeNames::soundend));
}
void SpeechRecognition::didEndAudio() {
dispatchEvent(Event::create(EventTypeNames::audioend));
}
void SpeechRecognition::didReceiveResults(
const HeapVector<Member<SpeechRecognitionResult>>& newFinalResults,
const HeapVector<Member<SpeechRecognitionResult>>& currentInterimResults) {
size_t resultIndex = m_finalResults.size();
for (size_t i = 0; i < newFinalResults.size(); ++i)
m_finalResults.append(newFinalResults[i]);
HeapVector<Member<SpeechRecognitionResult>> results = m_finalResults;
for (size_t i = 0; i < currentInterimResults.size(); ++i)
results.append(currentInterimResults[i]);
dispatchEvent(SpeechRecognitionEvent::createResult(resultIndex, results));
}
void SpeechRecognition::didReceiveNoMatch(SpeechRecognitionResult* result) {
dispatchEvent(SpeechRecognitionEvent::createNoMatch(result));
}
void SpeechRecognition::didReceiveError(SpeechRecognitionError* error) {
dispatchEvent(error);
m_started = false;
}
void SpeechRecognition::didStart() {
dispatchEvent(Event::create(EventTypeNames::start));
}
void SpeechRecognition::didEnd() {
m_started = false;
m_stopping = false;
// If m_controller is null, this is being aborted from the ExecutionContext
// being detached, so don't dispatch an event.
if (m_controller)
dispatchEvent(Event::create(EventTypeNames::end));
}
const AtomicString& SpeechRecognition::interfaceName() const {
return EventTargetNames::SpeechRecognition;
}
ExecutionContext* SpeechRecognition::getExecutionContext() const {
return ActiveDOMObject::getExecutionContext();
}
void SpeechRecognition::contextDestroyed() {
m_controller = nullptr;
if (hasPendingActivity())
abort();
}
bool SpeechRecognition::hasPendingActivity() const {
return m_started;
}
SpeechRecognition::SpeechRecognition(Page* page, ExecutionContext* context)
: ActiveScriptWrappable(this),
ActiveDOMObject(context),
m_grammars(SpeechGrammarList::create()), // FIXME: The spec is not clear
// on the default value for the
// grammars attribute.
m_audioTrack(nullptr),
m_continuous(false),
m_interimResults(false),
m_maxAlternatives(1),
m_controller(SpeechRecognitionController::from(page)),
m_started(false),
m_stopping(false) {
// FIXME: Need to hook up with Page to get notified when the visibility
// changes.
}
SpeechRecognition::~SpeechRecognition() {}
DEFINE_TRACE(SpeechRecognition) {
visitor->trace(m_grammars);
visitor->trace(m_audioTrack);
visitor->trace(m_controller);
visitor->trace(m_finalResults);
EventTargetWithInlineData::trace(visitor);
ActiveDOMObject::trace(visitor);
}
} // namespace blink