1 /*
2  * Copyright (C) 2008 The Android Open Source Project
3  *
4  * Licensed under the Apache License, Version 2.0 (the "License");
5  * you may not use this file except in compliance with the License.
6  * You may obtain a copy of the License at
7  *
8  *      http://www.apache.org/licenses/LICENSE-2.0
9  *
10  * Unless required by applicable law or agreed to in writing, software
11  * distributed under the License is distributed on an "AS IS" BASIS,
12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13  * See the License for the specific language governing permissions and
14  * limitations under the License.
15  */
16 
17 package android.speech;
18 
19 import java.util.ArrayList;
20 
21 import android.app.Activity;
22 import android.content.ActivityNotFoundException;
23 import android.content.BroadcastReceiver;
24 import android.content.ComponentName;
25 import android.content.Context;
26 import android.content.Intent;
27 import android.content.pm.PackageManager;
28 import android.content.pm.ResolveInfo;
29 import android.os.Bundle;
30 
31 /**
32  * Constants for supporting speech recognition through starting an {@link Intent}
33  */
34 public class RecognizerIntent {
35     /**
36      * The extra key used in an intent to the speech recognizer for voice search. Not
37      * generally to be used by developers. The system search dialog uses this, for example,
38      * to set a calling package for identification by a voice search API. If this extra
39      * is set by anyone but the system process, it should be overridden by the voice search
40      * implementation.
41      */
42     public final static String EXTRA_CALLING_PACKAGE = "calling_package";
43 
RecognizerIntent()44     private RecognizerIntent() {
45         // Not for instantiating.
46     }
47 
48     /**
49      * Starts an activity that will prompt the user for speech and send it through a
50      * speech recognizer.  The results will be returned via activity results (in
51      * {@link Activity#onActivityResult}, if you start the intent using
52      * {@link Activity#startActivityForResult(Intent, int)}), or forwarded via a PendingIntent
53      * if one is provided.
54      *
55      * <p>Starting this intent with just {@link Activity#startActivity(Intent)} is not supported.
56      * You must either use {@link Activity#startActivityForResult(Intent, int)}, or provide a
57      * PendingIntent, to receive recognition results.
58      *
59      * <p>The implementation of this API is likely to stream audio to remote servers to perform
60      * speech recognition which can use a substantial amount of bandwidth.
61      *
62      * <p>Required extras:
63      * <ul>
64      *   <li>{@link #EXTRA_LANGUAGE_MODEL}
65      * </ul>
66      *
67      * <p>Optional extras:
68      * <ul>
69      *   <li>{@link #EXTRA_PROMPT}
70      *   <li>{@link #EXTRA_LANGUAGE}
71      *   <li>{@link #EXTRA_MAX_RESULTS}
72      *   <li>{@link #EXTRA_RESULTS_PENDINGINTENT}
73      *   <li>{@link #EXTRA_RESULTS_PENDINGINTENT_BUNDLE}
74      * </ul>
75      *
76      * <p> Result extras (returned in the result, not to be specified in the request):
77      * <ul>
78      *   <li>{@link #EXTRA_RESULTS}
79      * </ul>
80      *
81      * <p>NOTE: There may not be any applications installed to handle this action, so you should
82      * make sure to catch {@link ActivityNotFoundException}.
83      */
84     public static final String ACTION_RECOGNIZE_SPEECH = "android.speech.action.RECOGNIZE_SPEECH";
85 
86     /**
87      * Starts an activity that will prompt the user for speech, send it through a
88      * speech recognizer, and either display a web search result or trigger
89      * another type of action based on the user's speech.
90      *
91      * <p>If you want to avoid triggering any type of action besides web search, you can use
92      * the {@link #EXTRA_WEB_SEARCH_ONLY} extra.
93      *
94      * <p>Required extras:
95      * <ul>
96      *   <li>{@link #EXTRA_LANGUAGE_MODEL}
97      * </ul>
98      *
99      * <p>Optional extras:
100      * <ul>
101      *   <li>{@link #EXTRA_PROMPT}
102      *   <li>{@link #EXTRA_LANGUAGE}
103      *   <li>{@link #EXTRA_MAX_RESULTS}
104      *   <li>{@link #EXTRA_PARTIAL_RESULTS}
105      *   <li>{@link #EXTRA_WEB_SEARCH_ONLY}
106      *   <li>{@link #EXTRA_ORIGIN}
107      * </ul>
108      *
109      * <p> Result extras (returned in the result, not to be specified in the request):
110      * <ul>
111      *   <li>{@link #EXTRA_RESULTS}
112      *   <li>{@link #EXTRA_CONFIDENCE_SCORES} (optional)
113      * </ul>
114      *
115      * <p>NOTE: There may not be any applications installed to handle this action, so you should
116      * make sure to catch {@link ActivityNotFoundException}.
117      */
118     public static final String ACTION_WEB_SEARCH = "android.speech.action.WEB_SEARCH";
119 
120     /**
121      * Starts an activity that will prompt the user for speech without requiring the user's
122      * visual attention or touch input. It will send it through a speech recognizer,
123      * and either synthesize speech for a web search result or trigger
124      * another type of action based on the user's speech.
125      *
126      * This activity may be launched while device is locked in a secure mode.
127      * Special care must be taken to ensure that the voice actions that are performed while
128      * hands free cannot compromise the device's security.
129      * The activity should check the value of the {@link #EXTRA_SECURE} extra to determine
130      * whether the device has been securely locked. If so, the activity should either restrict
131      * the set of voice actions that are permitted or require some form of secure
132      * authentication before proceeding.
133      *
134      * To ensure that the activity's user interface is visible while the lock screen is showing,
135      * the activity should set the
136      * {@link android.view.WindowManager.LayoutParams#FLAG_SHOW_WHEN_LOCKED} window flag.
137      * Otherwise the activity's user interface may be hidden by the lock screen. The activity
138      * should take care not to leak private information when the device is securely locked.
139      *
140      * <p>Optional extras:
141      * <ul>
142      *   <li>{@link #EXTRA_SECURE}
143      * </ul>
144      *
145      * <p class="note">
146      * In some cases, a matching Activity may not exist, so ensure you
147      * safeguard against this.
148      */
149     public static final String ACTION_VOICE_SEARCH_HANDS_FREE =
150             "android.speech.action.VOICE_SEARCH_HANDS_FREE";
151 
152     /**
153      * Optional boolean to indicate that a "hands free" voice search was performed while the device
154      * was in a secure mode. An example of secure mode is when the device's screen lock is active,
155      * and it requires some form of authentication to be unlocked.
156      *
157      * When the device is securely locked, the voice search activity should either restrict
158      * the set of voice actions that are permitted, or require some form of secure authentication
159      * before proceeding.
160      */
161     public static final String EXTRA_SECURE = "android.speech.extras.EXTRA_SECURE";
162 
163     /**
164      * The minimum length of an utterance. We will not stop recording before this amount of time.
165      *
166      * Note that it is extremely rare you'd want to specify this value in an intent. If you don't
167      * have a very good reason to change these, you should leave them as they are. Note also that
168      * certain values may cause undesired or unexpected results - use judiciously! Additionally,
169      * depending on the recognizer implementation, these values may have no effect.
170      */
171     public static final String EXTRA_SPEECH_INPUT_MINIMUM_LENGTH_MILLIS =
172             "android.speech.extras.SPEECH_INPUT_MINIMUM_LENGTH_MILLIS";
173 
174     /**
175      * The amount of time that it should take after we stop hearing speech to consider the input
176      * complete.
177      *
178      * Note that it is extremely rare you'd want to specify this value in an intent. If
179      * you don't have a very good reason to change these, you should leave them as they are. Note
180      * also that certain values may cause undesired or unexpected results - use judiciously!
181      * Additionally, depending on the recognizer implementation, these values may have no effect.
182      */
183     public static final String EXTRA_SPEECH_INPUT_COMPLETE_SILENCE_LENGTH_MILLIS =
184             "android.speech.extras.SPEECH_INPUT_COMPLETE_SILENCE_LENGTH_MILLIS";
185 
186     /**
187      * The amount of time that it should take after we stop hearing speech to consider the input
188      * possibly complete. This is used to prevent the endpointer cutting off during very short
189      * mid-speech pauses.
190      *
191      * Note that it is extremely rare you'd want to specify this value in an intent. If
192      * you don't have a very good reason to change these, you should leave them as they are. Note
193      * also that certain values may cause undesired or unexpected results - use judiciously!
194      * Additionally, depending on the recognizer implementation, these values may have no effect.
195      */
196     public static final String EXTRA_SPEECH_INPUT_POSSIBLY_COMPLETE_SILENCE_LENGTH_MILLIS =
197             "android.speech.extras.SPEECH_INPUT_POSSIBLY_COMPLETE_SILENCE_LENGTH_MILLIS";
198 
199     /**
200      * Informs the recognizer which speech model to prefer when performing
201      * {@link #ACTION_RECOGNIZE_SPEECH}. The recognizer uses this
202      * information to fine tune the results. This extra is required. Activities implementing
203      * {@link #ACTION_RECOGNIZE_SPEECH} may interpret the values as they see fit.
204      *
205      *  @see #LANGUAGE_MODEL_FREE_FORM
206      *  @see #LANGUAGE_MODEL_WEB_SEARCH
207      */
208     public static final String EXTRA_LANGUAGE_MODEL = "android.speech.extra.LANGUAGE_MODEL";
209 
210     /**
211      * Use a language model based on free-form speech recognition.  This is a value to use for
212      * {@link #EXTRA_LANGUAGE_MODEL}.
213      * @see #EXTRA_LANGUAGE_MODEL
214      */
215     public static final String LANGUAGE_MODEL_FREE_FORM = "free_form";
216     /**
217      * Use a language model based on web search terms.  This is a value to use for
218      * {@link #EXTRA_LANGUAGE_MODEL}.
219      * @see #EXTRA_LANGUAGE_MODEL
220      */
221     public static final String LANGUAGE_MODEL_WEB_SEARCH = "web_search";
222 
223     /** Optional text prompt to show to the user when asking them to speak. */
224     public static final String EXTRA_PROMPT = "android.speech.extra.PROMPT";
225 
226     /**
227      * Optional IETF language tag (as defined by BCP 47), for example "en-US". This tag informs the
228      * recognizer to perform speech recognition in a language different than the one set in the
229      * {@link java.util.Locale#getDefault()}.
230      */
231     public static final String EXTRA_LANGUAGE = "android.speech.extra.LANGUAGE";
232 
233     /**
234      * Optional value which can be used to indicate the referer url of a page in which
235      * speech was requested. For example, a web browser may choose to provide this for
236      * uses of speech on a given page.
237      */
238     public static final String EXTRA_ORIGIN = "android.speech.extra.ORIGIN";
239 
240     /**
241      * Optional limit on the maximum number of results to return. If omitted the recognizer
242      * will choose how many results to return. Must be an integer.
243      */
244     public static final String EXTRA_MAX_RESULTS = "android.speech.extra.MAX_RESULTS";
245 
246     /**
247      * Optional boolean, to be used with {@link #ACTION_WEB_SEARCH}, to indicate whether to
248      * only fire web searches in response to a user's speech. The default is false, meaning
249      * that other types of actions can be taken based on the user's speech.
250      */
251     public static final String EXTRA_WEB_SEARCH_ONLY = "android.speech.extra.WEB_SEARCH_ONLY";
252 
253     /**
254      * Optional boolean to indicate whether partial results should be returned by the recognizer
255      * as the user speaks (default is false).  The server may ignore a request for partial
256      * results in some or all cases.
257      */
258     public static final String EXTRA_PARTIAL_RESULTS = "android.speech.extra.PARTIAL_RESULTS";
259 
260     /**
261      * When the intent is {@link #ACTION_RECOGNIZE_SPEECH}, the speech input activity will
262      * return results to you via the activity results mechanism.  Alternatively, if you use this
263      * extra to supply a PendingIntent, the results will be added to its bundle and the
264      * PendingIntent will be sent to its target.
265      */
266     public static final String EXTRA_RESULTS_PENDINGINTENT =
267             "android.speech.extra.RESULTS_PENDINGINTENT";
268 
269     /**
270      * If you use {@link #EXTRA_RESULTS_PENDINGINTENT} to supply a forwarding intent, you can
271      * also use this extra to supply additional extras for the final intent.  The search results
272      * will be added to this bundle, and the combined bundle will be sent to the target.
273      */
274     public static final String EXTRA_RESULTS_PENDINGINTENT_BUNDLE =
275             "android.speech.extra.RESULTS_PENDINGINTENT_BUNDLE";
276 
277     /** Result code returned when no matches are found for the given speech */
278     public static final int RESULT_NO_MATCH = Activity.RESULT_FIRST_USER;
279     /** Result code returned when there is a generic client error */
280     public static final int RESULT_CLIENT_ERROR = Activity.RESULT_FIRST_USER + 1;
281     /** Result code returned when the recognition server returns an error */
282     public static final int RESULT_SERVER_ERROR = Activity.RESULT_FIRST_USER + 2;
283     /** Result code returned when a network error was encountered */
284     public static final int RESULT_NETWORK_ERROR = Activity.RESULT_FIRST_USER + 3;
285     /** Result code returned when an audio error was encountered */
286     public static final int RESULT_AUDIO_ERROR = Activity.RESULT_FIRST_USER + 4;
287 
288     /**
289      * An ArrayList&lt;String&gt; of the recognition results when performing
290      * {@link #ACTION_RECOGNIZE_SPEECH}. Generally this list should be ordered in
291      * descending order of speech recognizer confidence. (See {@link #EXTRA_CONFIDENCE_SCORES}).
292      * Returned in the results; not to be specified in the recognition request. Only present
293      * when {@link Activity#RESULT_OK} is returned in an activity result. In a PendingIntent,
294      * the lack of this extra indicates failure.
295      */
296     public static final String EXTRA_RESULTS = "android.speech.extra.RESULTS";
297 
298     /**
299      * A float array of confidence scores of the recognition results when performing
300      * {@link #ACTION_RECOGNIZE_SPEECH}. The array should be the same size as the ArrayList
301      * returned in {@link #EXTRA_RESULTS}, and should contain values ranging from 0.0 to 1.0,
302      * or -1 to represent an unavailable confidence score.
303      * <p>
304      * Confidence values close to 1.0 indicate high confidence (the speech recognizer is
305      * confident that the recognition result is correct), while values close to 0.0 indicate
306      * low confidence.
307      * <p>
308      * Returned in the results; not to be specified in the recognition request. This extra is
309      * optional and might not be provided. Only present when {@link Activity#RESULT_OK} is
310      * returned in an activity result.
311      */
312     public static final String EXTRA_CONFIDENCE_SCORES = "android.speech.extra.CONFIDENCE_SCORES";
313 
314     /**
315      * Returns the broadcast intent to fire with
316      * {@link Context#sendOrderedBroadcast(Intent, String, BroadcastReceiver, android.os.Handler, int, String, Bundle)}
317      * to receive details from the package that implements voice search.
318      * <p>
319      * This is based on the value specified by the voice search {@link Activity} in
320      * {@link #DETAILS_META_DATA}, and if this is not specified, will return null. Also if there
321      * is no chosen default to resolve for {@link #ACTION_WEB_SEARCH}, this will return null.
322      * <p>
323      * If an intent is returned and is fired, a {@link Bundle} of extras will be returned to the
324      * provided result receiver, and should ideally contain values for
325      * {@link #EXTRA_LANGUAGE_PREFERENCE} and {@link #EXTRA_SUPPORTED_LANGUAGES}.
326      * <p>
327      * (Whether these are actually provided is up to the particular implementation. It is
328      * recommended that {@link Activity}s implementing {@link #ACTION_WEB_SEARCH} provide this
329      * information, but it is not required.)
330      *
331      * @param context a context object
332      * @return the broadcast intent to fire or null if not available
333      */
getVoiceDetailsIntent(Context context)334     public static final Intent getVoiceDetailsIntent(Context context) {
335         Intent voiceSearchIntent = new Intent(ACTION_WEB_SEARCH);
336         ResolveInfo ri = context.getPackageManager().resolveActivity(
337                 voiceSearchIntent, PackageManager.GET_META_DATA);
338         if (ri == null || ri.activityInfo == null || ri.activityInfo.metaData == null) return null;
339 
340         String className = ri.activityInfo.metaData.getString(DETAILS_META_DATA);
341         if (className == null) return null;
342 
343         Intent detailsIntent = new Intent(ACTION_GET_LANGUAGE_DETAILS);
344         detailsIntent.setComponent(new ComponentName(ri.activityInfo.packageName, className));
345         return detailsIntent;
346     }
347 
348     /**
349      * Meta-data name under which an {@link Activity} implementing {@link #ACTION_WEB_SEARCH} can
350      * use to expose the class name of a {@link BroadcastReceiver} which can respond to request for
351      * more information, from any of the broadcast intents specified in this class.
352      * <p>
353      * Broadcast intents can be directed to the class name specified in the meta-data by creating
354      * an {@link Intent}, setting the component with
355      * {@link Intent#setComponent(android.content.ComponentName)}, and using
356      * {@link Context#sendOrderedBroadcast(Intent, String, BroadcastReceiver, android.os.Handler, int, String, android.os.Bundle)}
357      * with another {@link BroadcastReceiver} which can receive the results.
358      * <p>
359      * The {@link #getVoiceDetailsIntent(Context)} method is provided as a convenience to create
360      * a broadcast intent based on the value of this meta-data, if available.
361      * <p>
362      * This is optional and not all {@link Activity}s which implement {@link #ACTION_WEB_SEARCH}
363      * are required to implement this. Thus retrieving this meta-data may be null.
364      */
365     public static final String DETAILS_META_DATA = "android.speech.DETAILS";
366 
367     /**
368      * A broadcast intent which can be fired to the {@link BroadcastReceiver} component specified
369      * in the meta-data defined in the {@link #DETAILS_META_DATA} meta-data of an
370      * {@link Activity} satisfying {@link #ACTION_WEB_SEARCH}.
371      * <p>
372      * When fired with
373      * {@link Context#sendOrderedBroadcast(Intent, String, BroadcastReceiver, android.os.Handler, int, String, android.os.Bundle)},
374      * a {@link Bundle} of extras will be returned to the provided result receiver, and should
375      * ideally contain values for {@link #EXTRA_LANGUAGE_PREFERENCE} and
376      * {@link #EXTRA_SUPPORTED_LANGUAGES}.
377      * <p>
378      * (Whether these are actually provided is up to the particular implementation. It is
379      * recommended that {@link Activity}s implementing {@link #ACTION_WEB_SEARCH} provide this
380      * information, but it is not required.)
381      */
382     public static final String ACTION_GET_LANGUAGE_DETAILS =
383             "android.speech.action.GET_LANGUAGE_DETAILS";
384 
385     /**
386      * Specify this boolean extra in a broadcast of {@link #ACTION_GET_LANGUAGE_DETAILS} to
387      * indicate that only the current language preference is needed in the response. This
388      * avoids any additional computation if all you need is {@link #EXTRA_LANGUAGE_PREFERENCE}
389      * in the response.
390      */
391     public static final String EXTRA_ONLY_RETURN_LANGUAGE_PREFERENCE =
392             "android.speech.extra.ONLY_RETURN_LANGUAGE_PREFERENCE";
393 
394     /**
395      * The key to the extra in the {@link Bundle} returned by {@link #ACTION_GET_LANGUAGE_DETAILS}
396      * which is a {@link String} that represents the current language preference this user has
397      * specified - a locale string like "en-US".
398      */
399     public static final String EXTRA_LANGUAGE_PREFERENCE =
400             "android.speech.extra.LANGUAGE_PREFERENCE";
401 
402     /**
403      * The key to the extra in the {@link Bundle} returned by {@link #ACTION_GET_LANGUAGE_DETAILS}
404      * which is an {@link ArrayList} of {@link String}s that represents the languages supported by
405      * this implementation of voice recognition - a list of strings like "en-US", "cmn-Hans-CN",
406      * etc.
407      */
408     public static final String EXTRA_SUPPORTED_LANGUAGES =
409             "android.speech.extra.SUPPORTED_LANGUAGES";
410 
411     /**
412      * Optional boolean, to be used with {@link #ACTION_RECOGNIZE_SPEECH},
413      * {@link #ACTION_VOICE_SEARCH_HANDS_FREE}, {@link #ACTION_WEB_SEARCH} to indicate whether to
414      * only use an offline speech recognition engine. The default is false, meaning that either
415      * network or offline recognition engines may be used.
416      */
417     public static final String EXTRA_PREFER_OFFLINE = "android.speech.extra.PREFER_OFFLINE";
418 }
419