RecognizerIntent.java revision 72da1e6f788ad2ca27518b630fa7d251c31d9a8b
1/*
2 * Copyright (C) 2008 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 *      http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17package android.speech;
18
19import java.util.ArrayList;
20
21import android.app.Activity;
22import android.content.ActivityNotFoundException;
23import android.content.BroadcastReceiver;
24import android.content.ComponentName;
25import android.content.Context;
26import android.content.Intent;
27import android.content.pm.PackageManager;
28import android.content.pm.ResolveInfo;
29import android.os.Bundle;
30
31/**
32 * Constants for supporting speech recognition through starting an {@link Intent}
33 */
34public class RecognizerIntent {
35    /**
36     * The extra key used in an intent to the speech recognizer for voice search. Not
37     * generally to be used by developers. The system search dialog uses this, for example,
38     * to set a calling package for identification by a voice search API. If this extra
39     * is set by anyone but the system process, it should be overridden by the voice search
40     * implementation.
41     */
42    public final static String EXTRA_CALLING_PACKAGE = "calling_package";
43
44    private RecognizerIntent() {
45        // Not for instantiating.
46    }
47
48    /**
49     * Starts an activity that will prompt the user for speech and send it through a
50     * speech recognizer.  The results will be returned via activity results (in
51     * {@link Activity#onActivityResult}, if you start the intent using
52     * {@link Activity#startActivityForResult(Intent, int)}), or forwarded via a PendingIntent
53     * if one is provided.
54     *
55     * <p>Starting this intent with just {@link Activity#startActivity(Intent)} is not supported.
56     * You must either use {@link Activity#startActivityForResult(Intent, int)}, or provide a
57     * PendingIntent, to receive recognition results.
58     *
59     * <p>The implementation of this API is likely to stream audio to remote servers to perform
60     * speech recognition which can use a substantial amount of bandwidth.
61     *
62     * <p>Required extras:
63     * <ul>
64     *   <li>{@link #EXTRA_LANGUAGE_MODEL}
65     * </ul>
66     *
67     * <p>Optional extras:
68     * <ul>
69     *   <li>{@link #EXTRA_PROMPT}
70     *   <li>{@link #EXTRA_LANGUAGE}
71     *   <li>{@link #EXTRA_MAX_RESULTS}
72     *   <li>{@link #EXTRA_RESULTS_PENDINGINTENT}
73     *   <li>{@link #EXTRA_RESULTS_PENDINGINTENT_BUNDLE}
74     * </ul>
75     *
76     * <p> Result extras (returned in the result, not to be specified in the request):
77     * <ul>
78     *   <li>{@link #EXTRA_RESULTS}
79     * </ul>
80     *
81     * <p>NOTE: There may not be any applications installed to handle this action, so you should
82     * make sure to catch {@link ActivityNotFoundException}.
83     */
84    public static final String ACTION_RECOGNIZE_SPEECH = "android.speech.action.RECOGNIZE_SPEECH";
85
86    /**
87     * Starts an activity that will prompt the user for speech, send it through a
88     * speech recognizer, and either display a web search result or trigger
89     * another type of action based on the user's speech.
90     *
91     * <p>If you want to avoid triggering any type of action besides web search, you can use
92     * the {@link #EXTRA_WEB_SEARCH_ONLY} extra.
93     *
94     * <p>Required extras:
95     * <ul>
96     *   <li>{@link #EXTRA_LANGUAGE_MODEL}
97     * </ul>
98     *
99     * <p>Optional extras:
100     * <ul>
101     *   <li>{@link #EXTRA_PROMPT}
102     *   <li>{@link #EXTRA_LANGUAGE}
103     *   <li>{@link #EXTRA_MAX_RESULTS}
104     *   <li>{@link #EXTRA_PARTIAL_RESULTS}
105     *   <li>{@link #EXTRA_WEB_SEARCH_ONLY}
106     *   <li>{@link #EXTRA_ORIGIN}
107     * </ul>
108     *
109     * <p> Result extras (returned in the result, not to be specified in the request):
110     * <ul>
111     *   <li>{@link #EXTRA_RESULTS}
112     *   <li>{@link #EXTRA_CONFIDENCE_SCORES} (optional)
113     * </ul>
114     *
115     * <p>NOTE: There may not be any applications installed to handle this action, so you should
116     * make sure to catch {@link ActivityNotFoundException}.
117     */
118    public static final String ACTION_WEB_SEARCH = "android.speech.action.WEB_SEARCH";
119
120    /**
121     * Starts an activity that will prompt the user for speech without requiring the user's
122     * visual attention or touch input. It will send it through a speech recognizer,
123     * and either synthesize speech for a web search result or trigger
124     * another type of action based on the user's speech.
125     *
126     * This activity may be launched while device is locked in a secure mode.
127     * Special care must be taken to ensure that the voice actions that are performed while
128     * hands free cannot compromise the device's security.
129     * The activity should check the value of the {@link #EXTRA_SECURE} extra to determine
130     * whether the device has been securely locked. If so, the activity should either restrict
131     * the set of voice actions that are permitted or require some form of secure
132     * authentication before proceeding.
133     *
134     * To ensure that the activity's user interface is visible while the lock screen is showing,
135     * the activity should set the
136     * {@link android.view.WindowManager.LayoutParams#FLAG_SHOW_WHEN_LOCKED} window flag.
137     * Otherwise the activity's user interface may be hidden by the lock screen. The activity
138     * should take care not to leak private information when the device is securely locked.
139     *
140     * <p>Optional extras:
141     * <ul>
142     *   <li>{@link #EXTRA_SECURE}
143     * </ul>
144     *
145     * <p class="note">
146     * In some cases, a matching Activity may not exist, so ensure you
147     * safeguard against this.
148     */
149    public static final String ACTION_VOICE_SEARCH_HANDS_FREE =
150            "android.speech.action.VOICE_SEARCH_HANDS_FREE";
151
152    /**
153     * Optional boolean to indicate that a "hands free" voice search was performed while the device
154     * was in a secure mode. An example of secure mode is when the device's screen lock is active,
155     * and it requires some form of authentication to be unlocked.
156     *
157     * When the device is securely locked, the voice search activity should either restrict
158     * the set of voice actions that are permitted, or require some form of secure authentication
159     * before proceeding.
160     */
161    public static final String EXTRA_SECURE = "android.speech.extras.EXTRA_SECURE";
162
163    /**
164     * The minimum length of an utterance. We will not stop recording before this amount of time.
165     *
166     * Note that it is extremely rare you'd want to specify this value in an intent. If you don't
167     * have a very good reason to change these, you should leave them as they are. Note also that
168     * certain values may cause undesired or unexpected results - use judiciously! Additionally,
169     * depending on the recognizer implementation, these values may have no effect.
170     */
171    public static final String EXTRA_SPEECH_INPUT_MINIMUM_LENGTH_MILLIS =
172            "android.speech.extras.SPEECH_INPUT_MINIMUM_LENGTH_MILLIS";
173
174    /**
175     * The amount of time that it should take after we stop hearing speech to consider the input
176     * complete.
177     *
178     * Note that it is extremely rare you'd want to specify this value in an intent. If
179     * you don't have a very good reason to change these, you should leave them as they are. Note
180     * also that certain values may cause undesired or unexpected results - use judiciously!
181     * Additionally, depending on the recognizer implementation, these values may have no effect.
182     */
183    public static final String EXTRA_SPEECH_INPUT_COMPLETE_SILENCE_LENGTH_MILLIS =
184            "android.speech.extras.SPEECH_INPUT_COMPLETE_SILENCE_LENGTH_MILLIS";
185
186    /**
187     * The amount of time that it should take after we stop hearing speech to consider the input
188     * possibly complete. This is used to prevent the endpointer cutting off during very short
189     * mid-speech pauses.
190     *
191     * Note that it is extremely rare you'd want to specify this value in an intent. If
192     * you don't have a very good reason to change these, you should leave them as they are. Note
193     * also that certain values may cause undesired or unexpected results - use judiciously!
194     * Additionally, depending on the recognizer implementation, these values may have no effect.
195     */
196    public static final String EXTRA_SPEECH_INPUT_POSSIBLY_COMPLETE_SILENCE_LENGTH_MILLIS =
197            "android.speech.extras.SPEECH_INPUT_POSSIBLY_COMPLETE_SILENCE_LENGTH_MILLIS";
198
199    /**
200     * Informs the recognizer which speech model to prefer when performing
201     * {@link #ACTION_RECOGNIZE_SPEECH}. The recognizer uses this
202     * information to fine tune the results. This extra is required. Activities implementing
203     * {@link #ACTION_RECOGNIZE_SPEECH} may interpret the values as they see fit.
204     *
205     *  @see #LANGUAGE_MODEL_FREE_FORM
206     *  @see #LANGUAGE_MODEL_WEB_SEARCH
207     */
208    public static final String EXTRA_LANGUAGE_MODEL = "android.speech.extra.LANGUAGE_MODEL";
209
210    /**
211     * Use a language model based on free-form speech recognition.  This is a value to use for
212     * {@link #EXTRA_LANGUAGE_MODEL}.
213     * @see #EXTRA_LANGUAGE_MODEL
214     */
215    public static final String LANGUAGE_MODEL_FREE_FORM = "free_form";
216    /**
217     * Use a language model based on web search terms.  This is a value to use for
218     * {@link #EXTRA_LANGUAGE_MODEL}.
219     * @see #EXTRA_LANGUAGE_MODEL
220     */
221    public static final String LANGUAGE_MODEL_WEB_SEARCH = "web_search";
222
223    /** Optional text prompt to show to the user when asking them to speak. */
224    public static final String EXTRA_PROMPT = "android.speech.extra.PROMPT";
225
226    /**
227     * Optional IETF language tag (as defined by BCP 47), for example "en-US". This tag informs the
228     * recognizer to perform speech recognition in a language different than the one set in the
229     * {@link java.util.Locale#getDefault()}.
230     */
231    public static final String EXTRA_LANGUAGE = "android.speech.extra.LANGUAGE";
232
233    /**
234     * Optional value which can be used to indicate the referer url of a page in which
235     * speech was requested. For example, a web browser may choose to provide this for
236     * uses of speech on a given page.
237     */
238    public static final String EXTRA_ORIGIN = "android.speech.extra.ORIGIN";
239
240    /**
241     * Optional limit on the maximum number of results to return. If omitted the recognizer
242     * will choose how many results to return. Must be an integer.
243     */
244    public static final String EXTRA_MAX_RESULTS = "android.speech.extra.MAX_RESULTS";
245
246    /**
247     * Optional boolean, to be used with {@link #ACTION_WEB_SEARCH}, to indicate whether to
248     * only fire web searches in response to a user's speech. The default is false, meaning
249     * that other types of actions can be taken based on the user's speech.
250     */
251    public static final String EXTRA_WEB_SEARCH_ONLY = "android.speech.extra.WEB_SEARCH_ONLY";
252
253    /**
254     * Optional boolean to indicate whether partial results should be returned by the recognizer
255     * as the user speaks (default is false).  The server may ignore a request for partial
256     * results in some or all cases.
257     */
258    public static final String EXTRA_PARTIAL_RESULTS = "android.speech.extra.PARTIAL_RESULTS";
259
260    /**
261     * When the intent is {@link #ACTION_RECOGNIZE_SPEECH}, the speech input activity will
262     * return results to you via the activity results mechanism.  Alternatively, if you use this
263     * extra to supply a PendingIntent, the results will be added to its bundle and the
264     * PendingIntent will be sent to its target.
265     */
266    public static final String EXTRA_RESULTS_PENDINGINTENT =
267            "android.speech.extra.RESULTS_PENDINGINTENT";
268
269    /**
270     * If you use {@link #EXTRA_RESULTS_PENDINGINTENT} to supply a forwarding intent, you can
271     * also use this extra to supply additional extras for the final intent.  The search results
272     * will be added to this bundle, and the combined bundle will be sent to the target.
273     */
274    public static final String EXTRA_RESULTS_PENDINGINTENT_BUNDLE =
275            "android.speech.extra.RESULTS_PENDINGINTENT_BUNDLE";
276
277    /** Result code returned when no matches are found for the given speech */
278    public static final int RESULT_NO_MATCH = Activity.RESULT_FIRST_USER;
279    /** Result code returned when there is a generic client error */
280    public static final int RESULT_CLIENT_ERROR = Activity.RESULT_FIRST_USER + 1;
281    /** Result code returned when the recognition server returns an error */
282    public static final int RESULT_SERVER_ERROR = Activity.RESULT_FIRST_USER + 2;
283    /** Result code returned when a network error was encountered */
284    public static final int RESULT_NETWORK_ERROR = Activity.RESULT_FIRST_USER + 3;
285    /** Result code returned when an audio error was encountered */
286    public static final int RESULT_AUDIO_ERROR = Activity.RESULT_FIRST_USER + 4;
287
288    /**
289     * An ArrayList&lt;String&gt; of the recognition results when performing
290     * {@link #ACTION_RECOGNIZE_SPEECH}. Generally this list should be ordered in
291     * descending order of speech recognizer confidence. (See {@link #EXTRA_CONFIDENCE_SCORES}).
292     * Returned in the results; not to be specified in the recognition request. Only present
293     * when {@link Activity#RESULT_OK} is returned in an activity result. In a PendingIntent,
294     * the lack of this extra indicates failure.
295     */
296    public static final String EXTRA_RESULTS = "android.speech.extra.RESULTS";
297
298    /**
299     * A float array of confidence scores of the recognition results when performing
300     * {@link #ACTION_RECOGNIZE_SPEECH}. The array should be the same size as the ArrayList
301     * returned in {@link #EXTRA_RESULTS}, and should contain values ranging from 0.0 to 1.0,
302     * or -1 to represent an unavailable confidence score.
303     * <p>
304     * Confidence values close to 1.0 indicate high confidence (the speech recognizer is
305     * confident that the recognition result is correct), while values close to 0.0 indicate
306     * low confidence.
307     * <p>
308     * Returned in the results; not to be specified in the recognition request. This extra is
309     * optional and might not be provided. Only present when {@link Activity#RESULT_OK} is
310     * returned in an activity result.
311     */
312    public static final String EXTRA_CONFIDENCE_SCORES = "android.speech.extra.CONFIDENCE_SCORES";
313
314    /**
315     * Returns the broadcast intent to fire with
316     * {@link Context#sendOrderedBroadcast(Intent, String, BroadcastReceiver, android.os.Handler, int, String, Bundle)}
317     * to receive details from the package that implements voice search.
318     * <p>
319     * This is based on the value specified by the voice search {@link Activity} in
320     * {@link #DETAILS_META_DATA}, and if this is not specified, will return null. Also if there
321     * is no chosen default to resolve for {@link #ACTION_WEB_SEARCH}, this will return null.
322     * <p>
323     * If an intent is returned and is fired, a {@link Bundle} of extras will be returned to the
324     * provided result receiver, and should ideally contain values for
325     * {@link #EXTRA_LANGUAGE_PREFERENCE} and {@link #EXTRA_SUPPORTED_LANGUAGES}.
326     * <p>
327     * (Whether these are actually provided is up to the particular implementation. It is
328     * recommended that {@link Activity}s implementing {@link #ACTION_WEB_SEARCH} provide this
329     * information, but it is not required.)
330     *
331     * @param context a context object
332     * @return the broadcast intent to fire or null if not available
333     */
334    public static final Intent getVoiceDetailsIntent(Context context) {
335        Intent voiceSearchIntent = new Intent(ACTION_WEB_SEARCH);
336        ResolveInfo ri = context.getPackageManager().resolveActivity(
337                voiceSearchIntent, PackageManager.GET_META_DATA);
338        if (ri == null || ri.activityInfo == null || ri.activityInfo.metaData == null) return null;
339
340        String className = ri.activityInfo.metaData.getString(DETAILS_META_DATA);
341        if (className == null) return null;
342
343        Intent detailsIntent = new Intent(ACTION_GET_LANGUAGE_DETAILS);
344        detailsIntent.setComponent(new ComponentName(ri.activityInfo.packageName, className));
345        return detailsIntent;
346    }
347
348    /**
349     * Meta-data name under which an {@link Activity} implementing {@link #ACTION_WEB_SEARCH} can
350     * use to expose the class name of a {@link BroadcastReceiver} which can respond to request for
351     * more information, from any of the broadcast intents specified in this class.
352     * <p>
353     * Broadcast intents can be directed to the class name specified in the meta-data by creating
354     * an {@link Intent}, setting the component with
355     * {@link Intent#setComponent(android.content.ComponentName)}, and using
356     * {@link Context#sendOrderedBroadcast(Intent, String, BroadcastReceiver, android.os.Handler, int, String, android.os.Bundle)}
357     * with another {@link BroadcastReceiver} which can receive the results.
358     * <p>
359     * The {@link #getVoiceDetailsIntent(Context)} method is provided as a convenience to create
360     * a broadcast intent based on the value of this meta-data, if available.
361     * <p>
362     * This is optional and not all {@link Activity}s which implement {@link #ACTION_WEB_SEARCH}
363     * are required to implement this. Thus retrieving this meta-data may be null.
364     */
365    public static final String DETAILS_META_DATA = "android.speech.DETAILS";
366
367    /**
368     * A broadcast intent which can be fired to the {@link BroadcastReceiver} component specified
369     * in the meta-data defined in the {@link #DETAILS_META_DATA} meta-data of an
370     * {@link Activity} satisfying {@link #ACTION_WEB_SEARCH}.
371     * <p>
372     * When fired with
373     * {@link Context#sendOrderedBroadcast(Intent, String, BroadcastReceiver, android.os.Handler, int, String, android.os.Bundle)},
374     * a {@link Bundle} of extras will be returned to the provided result receiver, and should
375     * ideally contain values for {@link #EXTRA_LANGUAGE_PREFERENCE} and
376     * {@link #EXTRA_SUPPORTED_LANGUAGES}.
377     * <p>
378     * (Whether these are actually provided is up to the particular implementation. It is
379     * recommended that {@link Activity}s implementing {@link #ACTION_WEB_SEARCH} provide this
380     * information, but it is not required.)
381     */
382    public static final String ACTION_GET_LANGUAGE_DETAILS =
383            "android.speech.action.GET_LANGUAGE_DETAILS";
384
385    /**
386     * Specify this boolean extra in a broadcast of {@link #ACTION_GET_LANGUAGE_DETAILS} to
387     * indicate that only the current language preference is needed in the response. This
388     * avoids any additional computation if all you need is {@link #EXTRA_LANGUAGE_PREFERENCE}
389     * in the response.
390     */
391    public static final String EXTRA_ONLY_RETURN_LANGUAGE_PREFERENCE =
392            "android.speech.extra.ONLY_RETURN_LANGUAGE_PREFERENCE";
393
394    /**
395     * The key to the extra in the {@link Bundle} returned by {@link #ACTION_GET_LANGUAGE_DETAILS}
396     * which is a {@link String} that represents the current language preference this user has
397     * specified - a locale string like "en-US".
398     */
399    public static final String EXTRA_LANGUAGE_PREFERENCE =
400            "android.speech.extra.LANGUAGE_PREFERENCE";
401
402    /**
403     * The key to the extra in the {@link Bundle} returned by {@link #ACTION_GET_LANGUAGE_DETAILS}
404     * which is an {@link ArrayList} of {@link String}s that represents the languages supported by
405     * this implementation of voice recognition - a list of strings like "en-US", "cmn-Hans-CN",
406     * etc.
407     */
408    public static final String EXTRA_SUPPORTED_LANGUAGES =
409            "android.speech.extra.SUPPORTED_LANGUAGES";
410
411    /**
412     * Optional boolean, to be used with {@link #ACTION_RECOGNIZE_SPEECH},
413     * {@link #ACTION_VOICE_SEARCH_HANDS_FREE}, {@link #ACTION_WEB_SEARCH} to indicate whether to
414     * only use an offline speech recognition engine. The default is false, meaning that either
415     * network or offline recognition engines may be used.
416     */
417    public static final String EXTRA_PREFER_OFFLINE = "android.speech.extra.PREFER_OFFLINE";
418}
419