Add two new parameters that are used when synthesizing text and
playing it back directly to control the volume and left-right
panning of the output.
Panning is applied using a balance law, which is not energy-preserving
but which doesn't lower the volume when not panning / panning to
center (legacy behavior).
Reduced amount of logs, and removed spoken text.
In TextToSpeech.java: added convenience method to handle the setting
of the cached synthesis parameters.
Change-Id: I235d3d3193283ccc1891e2065d43787e3f63304d
* commit '0655b31d758ce8fe054c98c6d288e424f2d273e2':
Unhide RecognizerResultsIntent. This API was reviewed for Froyo, but we didn't want to make it public then because it wasn't until our first Market release of Voice Search that the APIs would be in use by our app.
reviewed for Froyo, but we didn't want to make it
public then because it wasn't until our first
Market release of Voice Search that the APIs would
be in use by our app.
http://b/3135351
Change-Id: I49053717cac08e3976c22e3a105139b6755aadb8
avoid triggering any non-web-search actions from the speech recognizer
(such as Voice Actions in the Google implementation). The browser will set
this to true to avoid triggering non-web-search actions when speech
recognition is requested from the browser.
Change-Id: I9b410ef8696e12f22dbad325b367f0f187a2dce8
http://b/3289997http://b/3176717
interfere with one another.
Root cause was not caching all the params that were needed.
This change fixes that bug as well as related bugs for remembering
the default engine and for making sure that the right engine is
loaded when checking for language availability.
Change-Id: I2a76da8faec8112036e68d27539db444c53a1509
This fix is part of addressing bug #2548048
Making it possible to determine which TTS engine is currently set
as the default by the user.
Change-Id: Ifa422f43a33498b94d96811078c0794d64668357
bug fix for: http://b/issue?id=2522063
RecognitionManager does not throw an exception in startListening() was called without the preceding setListener()
Change-Id: I8e303d0c53c7593fbf1dc7033b2c0f82eebfb081
It is a prt of the following bug fix: http://b/issue?id=2478548
The CL adds the ability to pass HTTP headers to the Browser
Change-Id: Ibf0ad8f678fc5aeef4ac098e5dfbcaed9ada8600
This is the framework part, moving classes around so the framework
no longer needs to link to android-common. Makes some APIs public,
others that didn't need to be public are private in the framework,
some small things are copied.
* Allow activities satisfying RecognizerIntent.ACTION_WEB_SEARCH to
point to the class name of a broadcast receiver which returns details
about the voice search implementation in an ordered broadcast response.
Provide a convenience method for getting the intent to fire for this
info. This can be used to get the current language preference and the
list of supported languages, and is extensible for future uses.
* When creating a RecognitionManager, allow the caller to optionally specify
a specific component of a voice recognition service on the device that they
want to use. This way, an app can still use its own service through
RecognitionManager, even if it's not the one chosen the user in settings.
to be used by anyone implementing a voice recognition service. Also define
a new <recognition-service> styleable to be used in such service's metadata
xml.
Still to do: Change VoiceSearch's GoogleRecognitionService to respond to this
intent, and actually use this intent instead of ACTION_RECOGNIZE_SPEECH here
in RecognitionService.
for voice recognition on the device. Right now this just queries
the package manager at boot and finds the (hopefully) single
available recognizer.
TODO: Add an attribute to let recognition services expose a settings
activity, and expose the settings activity of the chosen recognition
service in the system settings for voice input & output.
Reinstalling VoiceIME created a problem because RecognitionService expected the first command to be setListener, in this version, such command is added
if the connection is broken.
Change-Id: Ia102fc1843053e2bdd330b380c2685a1227081b2
Specifically point out that startActivity() is not supported for
ACTION_RECOGNIZE_SPEECH, and make the documentation on EXTRA_RESULTS more
clear to point out that this is a part of the results, not the request.
from system settings. For now it'll just be triggered from within the
voice search app if you choose the settings menu item.
Need to unhide this before we can be fully unbundled for voice search.