US20080195396A1 - System, method and computer program product for adding voice activation and voice control to a media player - Google Patents

System, method and computer program product for adding voice activation and voice control to a media player Download PDF

Info

Publication number
US20080195396A1
US20080195396A1 US12/104,195 US10419508A US2008195396A1 US 20080195396 A1 US20080195396 A1 US 20080195396A1 US 10419508 A US10419508 A US 10419508A US 2008195396 A1 US2008195396 A1 US 2008195396A1
Authority
US
United States
Prior art keywords
utterance
command
sub
logic
assembly
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/104,195
Inventor
Mark Greene
Michael Hegarty
Dermot Cantwell
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=37637638&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=US20080195396(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by Individual filed Critical Individual
Priority to US12/104,195 priority Critical patent/US20080195396A1/en
Assigned to STRAGENT, LLC reassignment STRAGENT, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: VOICEDEMAND, INC.
Publication of US20080195396A1 publication Critical patent/US20080195396A1/en
Assigned to SEESAW FOUNDATION reassignment SEESAW FOUNDATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: STRAGENT, LLC
Assigned to TAG FOUNDATION reassignment TAG FOUNDATION CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: SEESAW FOUNDATION
Assigned to SEESAW FOUNDATION reassignment SEESAW FOUNDATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: STRAGENT, LLC
Assigned to STRAGENT, LLC reassignment STRAGENT, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TAG FOUNDATION
Assigned to STRAGENT, LLC reassignment STRAGENT, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: STRAGENT, LLC
Assigned to VOICEDEMAND, INC. reassignment VOICEDEMAND, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CANTWELL, DERMOT, GREENE, MARK, HEGARTY, MICHAEL
Assigned to KILPAUK GARDEN, SERIES 64 OF ALLIED SECURITY TRUST I reassignment KILPAUK GARDEN, SERIES 64 OF ALLIED SECURITY TRUST I ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: STRAGENT, LLC
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KILPAUK GARDEN, SERIES 64 OF ALLIED SECURITY TRUST I
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H03ELECTRONIC CIRCUITRY
    • H03JTUNING RESONANT CIRCUITS; SELECTING RESONANT CIRCUITS
    • H03J9/00Remote-control of tuned circuits; Combined remote-control of tuning and other functions, e.g. brightness, amplification
    • H03J9/02Remote-control of tuned circuits; Combined remote-control of tuning and other functions, e.g. brightness, amplification using radio transmission; using near-field transmission
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B19/00Driving, starting, stopping record carriers not specifically of filamentary or web form, or of supports therefor; Control thereof; Control of operating function ; Driving both disc and head
    • G11B19/02Control of operating function, e.g. switching from recording to reproducing

Definitions

  • a media player system, method and computer program product are provided.
  • an utterance is received.
  • a command for a media player is then generated based on the utterance. Such command is utilized for providing wireless control of the media player.
  • FIG. 1 illustrates a network architecture, in accordance with one embodiment.
  • FIG. 2 shows a representative hardware environment that may be associated with the devices of FIG. 1 , in accordance with one embodiment.
  • FIG. 3 shows a method for providing wireless control of a media player, in accordance with one embodiment.
  • FIG. 4 shows a method for providing wireless control of a media player, in accordance with another embodiment.
  • FIG. 5 shows a media player in connection with an assembly for receiving utterances, in accordance with another embodiment.
  • FIG. 6 shows a media player in connection with an assembly for receiving utterances, in accordance with still yet another embodiment.
  • FIG. 7 shows a method for providing wireless control of a media player when a library is loaded on the media player, in accordance with one embodiment.
  • FIG. 8 shows a method for providing wireless control of a media player when a library is not loaded on the media player, in accordance with another embodiment.
  • FIG. 1 illustrates a network architecture 100 , in accordance with one embodiment.
  • a network 102 is provided.
  • the network 102 may take any form including, but not limited to a telecommunications network, a local area network (LAN), a wireless network, a wide area network (WAN) such as the Internet, etc. While only one network is shown, it should be understood that two or more similar or different networks 102 may be provided.
  • LAN local area network
  • WAN wide area network
  • Coupled to the network 102 is a plurality of devices.
  • a server device 104 and an end user computer 106 may be coupled to the network 102 for communication purposes.
  • Such end user computer 106 may include a desktop computer, lap-top computer, and/or any other type of logic.
  • various other devices may be coupled to the network 102 including a media player 108 , a mobile phone 110 , etc.
  • any of the foregoing devices in the present network architecture 100 may be equipped with voice control of an associated media player. More exemplary information regarding such architecture and associated functionality will be set forth hereinafter in greater detail.
  • FIG. 2 illustrates an exemplary system 200 , in accordance with one embodiment.
  • the system 200 may be implemented in the context of any of the devices of the network architecture 100 of FIG. 1 .
  • the system 200 may be implemented in any desired environment.
  • a system 200 including at least one central processor 201 which is connected to a communication bus 202 .
  • the system 200 also includes main memory 204 [e.g. random access memory (RAM), etc.].
  • main memory 204 e.g. random access memory (RAM), etc.
  • graphics processor 206 e.g. graphics processing unit (GPU), etc.
  • the system 200 may also include a secondary storage 210 .
  • the secondary storage 210 includes, for example, a hard disk drive and/or a removable storage drive, representing a floppy disk drive, a magnetic tape drive, a compact disk drive, etc.
  • the removable storage drive reads from and/or writes to a removable storage unit in a well known manner.
  • Computer programs, or computer control logic algorithms may be stored in the main memory 204 and/or the secondary storage 210 . Such computer programs, when executed, enable the system 200 to perform various functions.
  • the memory 204 , storage 210 and/or any other storage are possible examples of computer-readable media.
  • FIG. 3 shows a method 300 for providing wireless control of a media player, in accordance with one embodiment.
  • the present method 300 may be implemented in the context of the architecture and environment of FIGS. 1 and/or 2 . Of course, however, the method 300 may be carried out in any desired environment.
  • an utterance is received, as shown in operation 302 .
  • the utterance may include any audible word, number, and/or sound capable of being received. Still yet, the utterance may be received by a microphone and/or any other desired input device capable of receiving the utterance. As an option, the utterance may be received utilizing an input device including an integrated microphone in a set of headphones which enables voice control in an active environment. As another option, the utterance may be received utilizing an input device including a wireless device which can be positioned for optimum voice control in an automobile, an indoor environment, and/or an outdoor environment.
  • a command for a media player is then generated based on the utterance.
  • a media player may include an iPod(R), a portable satellite radio player, and/or any portable software and/or hardware capable of outputting any sort of media [e.g. audible media (e.g. music, news, non-fiction information, fictional stories, etc.), visual media (e.g. pictures, video in the form of movies, news, programming, etc.), etc.].
  • audible media e.g. music, news, non-fiction information, fictional stories, etc.
  • visual media e.g. pictures, video in the form of movies, news, programming, etc.
  • the media player may be used in conjunction (e.g. built-in, retrofitted with, coupled to, etc.) any desired device including, but not limited to a cellular phone, personal digital assistant, etc. (e.g. see, for example, any of the devices of FIG. 1 , etc.).
  • any desired device including, but not limited to a cellular phone, personal digital assistant, etc. (e.g. see, for example, any of the devices of FIG. 1 , etc.).
  • the media player may further be a stand alone product.
  • the commands may include, for example, commands that operate the media player, commands that change states of the media player, etc.
  • commands may include play, pause, fast forward, rewind, on, off, shuffle, repeat, search, volume up, volume down, playlist, next playlist, etc.
  • the search command may provide a user with the ability to command the media player to search for a particular song or artist and the play command may provide the user with the ability to command a particular song to be played.
  • the commands may be programmable and/or registered by the media player.
  • the command may include any signal, instruction, code, data, etc. that is capable of being utilized for providing wireless control of the media player.
  • the command may be an utterance that is translated into a hex code capable of being recognized by the media player.
  • Table 1 illustrates examples of such hex codes capable of being recognized by the media player, in an embodiment where the media player includes the aforementioned iPod(R).
  • the operations of the method 300 of FIG. 3 may be carried out by the media player itself, and/or by way of a separate assembly that is either built-in the media player or capable of being retrofitted on the media player.
  • a separate assembly will be described in more detail with respect to FIGS. 4-6 .
  • FIG. 4 shows a method for providing wireless control of a media player, in accordance with another embodiment.
  • the present method 600 may be implemented in the context of the architecture and environment of FIGS. 1-3 . Of course, however, the method 600 may be carried out in any desired environment.
  • program variables within an assembly attached to a media player are initiated, as shown in operation 602 .
  • the program variables initialized may include initializing trigger and command words along with any input or output utilized.
  • An exemplary assembly will be described in more detail with respect to FIGS. 5 and 6 .
  • the assembly listens for a trigger word, as in operation 604 .
  • the trigger word may be any word capable of being recognized and understood by the assembly. It may also optionally be any word capable of initiating a voice recognition capability of the assembly. For example, a user may be required to first say “start” in order to trigger the assembly.
  • a command word is awaited, such as in operation 608 .
  • the command word may include any of the utterances described with respect to FIG. 3 .
  • the command word may include any word capable of being recognized by the assembly and capable of being translated into computer code that the media player can act upon.
  • a time limit e.g. 120 seconds, etc.
  • the method may return to operation 604 .
  • the command is converted into computer code that can be read by the media player, as shown in operation 612 .
  • the computer code may include computer code that can be read by the personal media player.
  • the determination of operation 610 may be false (or be more scrutinized, etc.) and the method may return to operation 604 .
  • the method may also return to operation 608 .
  • the computer code is then sent from the assembly to the media player, as shown in operation 614 , and a send button releases the command to the media player, as in operation 616 .
  • Such send button may indicate to the media player that the computer code is complete.
  • the computer code may be sent as a bit stream to the media player.
  • the bit stream may consist of seven codes of ten bits each.
  • each bit may be sent approximately every 55 uSecs with a code repetition rate of 66 (i.e. each code is sent about every 15 milliseconds).
  • command is a sleep command, as shown in operation 619 .
  • a sleep command may be created by a user uttering “sleep” and such utterance then being translated into computer code that tells the media player to power down. If the determination in operation 619 is positive, the media player is powered down and all program variables are released, as depicted in operation 620 . Powering down may, as an option, include setting the media player to a low power mode. Alternatively, if the command is not for the media player to sleep, the method 600 returns to operation 602 where it continues to listen for a trigger word. Again, method 600 is set forth to illustrate just one example of a method for wireless controlling a media player, and should not be construed as limiting in any manner.
  • FIG. 5 shows a media player in connection with an assembly for receiving utterances, in accordance with another embodiment.
  • the media player/assembly may be implemented in the context of the architecture and environment of FIGS. 1-4 . Of course, however, the media player/assembly may be implemented in any desired environment.
  • a media player 401 is shown connected to an assembly 403 .
  • media player 401 is shown with the assembly 403 holding the media player 401 .
  • assembly 403 could be optionally mounted on the personal media player 401 or connected in any other manner.
  • media player 401 is configured to connect to the assembly 403 .
  • the assembly 403 includes electrical connection (not shown), voice activation software (not shown), voice activation hardware (not shown), a memory integrated circuit (not shown), an FM transmitter 409 , and a power unit 402 .
  • the voice activation software may be capable of detecting an utterance, translating the utterance into computer code, and transmitting the code to the media player 401 .
  • the power unit 402 may be capable of charging the media player 401 and may include a plug that connects to an automobile cigarette lighter device or AC/DC converter device to provide a required voltage.
  • the connection may be located on the end of a flexible metal rod that supports the assembly.
  • the FM transmitter 409 may further include up and down arrows on the front of the assembly 403 as shown in FIG. 5 for sweeping across 88 MHz to 108 MHz such that a frequency that has little outside traffic may be locked in for transmitting a stereo signal at least six feet.
  • FM transmitter 409 may also include left and right channel programming capabilities.
  • a power LED 406 and charging LED 407 may be utilized for displaying whether the media player 401 is charging or listening for an utterance. Further included may be a verification speaker 404 for verifying utterances received, a directional microphone 405 for receiving and transferring utterances to a processing circuit (not shown) that translates the utterances into computer code capable of being read by the media player 401 .
  • the processing circuit may also include a voice engine, onboard memory, and a plurality of circuit peripherals.
  • the utterance may be verified by replaying the utterance for the user who provided the utterance and allowing the user to either accept or reject the utterance. For example, in playing the utterance back for the user, the user may be prompted to either state “yes” or “no.” If the user rejects the utterance, the user may then be prompted to give the utterance again. In this way, the verifying may allow for the adjusting of the received utterance if it is not verified. Of course, any type of verification process may optionally be utilized for verifying the utterance.
  • a connector (not shown) for sending and receiving data between the assembly 403 and the media player 401 as well as providing power to the media player 401 .
  • a mounting brace 408 may be provided.
  • a front face 410 may be provided.
  • an FM frequency display 411 may be provided.
  • output from the media player 401 may be sent through the assembly 403 .
  • Such output may be FM modulated with the FM transmitter 409 of the assembly 403 for reception by nearby FM receivers.
  • FIG. 6 shows a media player in connection with an assembly for receiving utterances, in accordance with still yet another embodiment.
  • the present media player/assembly may be implemented in the context of the architecture and environment of FIGS. 1-5 . Of course, however, the media player/assembly may be implemented in any desired environment.
  • the assembly 507 is shown connected to a media player 501 by being mounted on top of the media player 501 .
  • Such assembly 507 may optionally include a signal line out 502 , a directional or omni-directional microphone 503 , a power LED 504 and listening LED 505 (which indicates whether the assembly 507 is awaiting a command—see operation 608 of FIG. 4 , for example), an FM frequency adjustment 506 , and/or an FM frequency display 508 .
  • Such features may include the same functionalities described with respect to FIG. 5 .
  • FIG. 7 shows a method 700 for providing wireless control of a media player when a library is loaded on the media player, in accordance with one embodiment.
  • the present method 700 may be implemented in the context of the architecture and environment of FIGS. 1-6 . Of course, however, the method 700 may be implemented in any desired environment.
  • program variables within an assembly attached to a media player are initiated, as shown in operation 702 .
  • the program variables initialized may include initializing trigger and/or command words along with any input or output utilized.
  • the assembly may, for example, include the embodiments described in FIGS. 5 and 6 above.
  • the trigger word may be any word capable of being recognized and understood. It may also optionally be any word capable of initiating a voice recognition capability. For example, a user may be required to first say “start” in order to trigger the present embodiment.
  • a command word is awaited, such as in operation 708 .
  • the command word may include any of the utterances described with respect to FIG. 3 .
  • the command word may include any word capable of being recognized and capable of being translated into computer code that the media player can act upon.
  • the method 700 continues to listen for a next utterance in operation 712 .
  • the next utterance may be, for example, an artist name or any other word.
  • a time limit (e.g. 120 seconds, etc.) may be utilized such that if no utterance is detected during operation 712 within the time limit, the method 700 may terminate, as shown in decision 716 . Otherwise, if an utterance is detected during operation 712 within the time limit, the method 700 may navigate to a first item in a library of words associated with the utterance received in operation 712 . In the present embodiment, and shown just by way of example, the method 700 may navigate to a first item located in an artist library, as shown in operation 714 . Of course, it should be noted that any type of data capable of being located within a library may be utilized. An artist name and voice tag associated with the first item in the library may then be read, such as in operation 718 , and it may be determined whether the voice tag matches the utterance received in operation 712 (see decision 720 ).
  • the method 700 may navigate to a next item located in the artist library, as shown in operation 724 . If it is determined in decision 722 that there is not a next item in the artist library to navigate to (e.g. the method 700 has reached the end of the library), the method 700 may terminate. Otherwise, the method 700 may return to operation 718 where the artist name and voice tag associated with the next item are read.
  • the method 700 continues until it is determined in decision 720 that the voice tag matches the utterance received in operation 712 , in which case a command is issued to the media player, such as, for example, a command to play a first among a set of songs associated with the artist name received in operation 712 , as shown in operation 726 .
  • a command is issued to the media player, such as, for example, a command to play a first among a set of songs associated with the artist name received in operation 712 , as shown in operation 726 .
  • FIG. 8 shows a method 800 for providing wireless control of a media player when a library is not loaded on the media player, in accordance with another embodiment.
  • the present method 800 may be implemented in the context of the architecture and environment of FIGS. 1-6 . Of course, however, the method 800 may be implemented in any desired environment.
  • program variables within an assembly attached to a media player are initiated, as shown in operation 802 .
  • the program variables initialized may include initializing trigger and command words along with any input or output utilized.
  • the assembly listens for a trigger word, as in operation 804 .
  • the trigger word may be any word capable of being recognized and understood. It may also optionally be any word capable of initiating a voice recognition capability. For example, a user may be required to first say “start” in order to trigger the present embodiment.
  • a command word is awaited, such as in operation 808 .
  • the command word may include any of the utterances described with respect to FIG. 3 .
  • the command word may include any word capable of being recognized and capable of being translated into computer code that the media player can act upon.
  • the method 800 continues to listen for a next utterance in operation 812 .
  • the next utterance may be, for example, an artist name or any other word associated with a file located in a library loaded on the media player.
  • the file may be a song located in a library of songs loaded on the media player.
  • any type of data capable of being located within a library may be utilized.
  • a time limit (e.g. 120 seconds, etc.) may be utilized such that if no utterance is detected during operation 812 within the time limit, the method 800 may terminate, as shown in decision 814 . Otherwise, if an utterance is detected during operation 812 within the time limit, the method 800 converts the detected utterance into an application specific voice tag, as shown in operation 816 .
  • the method 800 may navigate to a first item in a library associated with the utterance received in operation 812 .
  • the method 800 may navigate to a first item located in an artist library as shown in operation 818 .
  • an artist name associated with the first item may be read, such as in operation 820 , and the artist name may be converted to an application specific voice tag (see operation 822 ).
  • the method 800 may terminate. Otherwise, the method 800 may return to operation 820 where the artist name and voice tag associated with the next item in the artist library are read.
  • the method 800 continues until it is determined in decision 824 that the voice tags match, in which case a command is issued to the media player, such as, for example, a command to play a song (or set of songs) associated with the artist name received in operation 812 (as shown in operation 830 ).
  • a command is issued to the media player, such as, for example, a command to play a song (or set of songs) associated with the artist name received in operation 812 (as shown in operation 830 ).

Abstract

A media player system, method and computer program product are provided. In use, an utterance is received. A command for a media player is then generated based on the utterance. Such command is utilized for providing wireless control of the media player.

Description

    SUMMARY
  • A media player system, method and computer program product are provided. In use, an utterance is received. A command for a media player is then generated based on the utterance. Such command is utilized for providing wireless control of the media player.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 illustrates a network architecture, in accordance with one embodiment.
  • FIG. 2 shows a representative hardware environment that may be associated with the devices of FIG. 1, in accordance with one embodiment.
  • FIG. 3 shows a method for providing wireless control of a media player, in accordance with one embodiment.
  • FIG. 4 shows a method for providing wireless control of a media player, in accordance with another embodiment.
  • FIG. 5 shows a media player in connection with an assembly for receiving utterances, in accordance with another embodiment.
  • FIG. 6 shows a media player in connection with an assembly for receiving utterances, in accordance with still yet another embodiment.
  • FIG. 7 shows a method for providing wireless control of a media player when a library is loaded on the media player, in accordance with one embodiment.
  • FIG. 8 shows a method for providing wireless control of a media player when a library is not loaded on the media player, in accordance with another embodiment.
  • DETAILED DESCRIPTION
  • FIG. 1 illustrates a network architecture 100, in accordance with one embodiment. As shown, a network 102 is provided. In the context of the present network architecture 100, the network 102 may take any form including, but not limited to a telecommunications network, a local area network (LAN), a wireless network, a wide area network (WAN) such as the Internet, etc. While only one network is shown, it should be understood that two or more similar or different networks 102 may be provided.
  • Coupled to the network 102 is a plurality of devices. For example, a server device 104 and an end user computer 106 may be coupled to the network 102 for communication purposes. Such end user computer 106 may include a desktop computer, lap-top computer, and/or any other type of logic. Still yet, various other devices may be coupled to the network 102 including a media player 108, a mobile phone 110, etc.
  • It should be noted that any of the foregoing devices in the present network architecture 100, as well as any other unillustrated hardware and/or software, may be equipped with voice control of an associated media player. More exemplary information regarding such architecture and associated functionality will be set forth hereinafter in greater detail.
  • FIG. 2 illustrates an exemplary system 200, in accordance with one embodiment. As an option, the system 200 may be implemented in the context of any of the devices of the network architecture 100 of FIG. 1. Of course, the system 200 may be implemented in any desired environment.
  • As shown, a system 200 is provided including at least one central processor 201 which is connected to a communication bus 202. The system 200 also includes main memory 204 [e.g. random access memory (RAM), etc.]. The system 200 also includes an optional graphics processor 206 and a display 208.
  • The system 200 may also include a secondary storage 210. The secondary storage 210 includes, for example, a hard disk drive and/or a removable storage drive, representing a floppy disk drive, a magnetic tape drive, a compact disk drive, etc. The removable storage drive reads from and/or writes to a removable storage unit in a well known manner.
  • Computer programs, or computer control logic algorithms, may be stored in the main memory 204 and/or the secondary storage 210. Such computer programs, when executed, enable the system 200 to perform various functions. The memory 204, storage 210 and/or any other storage are possible examples of computer-readable media.
  • FIG. 3 shows a method 300 for providing wireless control of a media player, in accordance with one embodiment. As an option, the present method 300 may be implemented in the context of the architecture and environment of FIGS. 1 and/or 2. Of course, however, the method 300 may be carried out in any desired environment.
  • Initially, an utterance is received, as shown in operation 302. The utterance may include any audible word, number, and/or sound capable of being received. Still yet, the utterance may be received by a microphone and/or any other desired input device capable of receiving the utterance. As an option, the utterance may be received utilizing an input device including an integrated microphone in a set of headphones which enables voice control in an active environment. As another option, the utterance may be received utilizing an input device including a wireless device which can be positioned for optimum voice control in an automobile, an indoor environment, and/or an outdoor environment.
  • Next, in operation 304, a command for a media player is then generated based on the utterance. In the context of the present description, a media player may include an iPod(R), a portable satellite radio player, and/or any portable software and/or hardware capable of outputting any sort of media [e.g. audible media (e.g. music, news, non-fiction information, fictional stories, etc.), visual media (e.g. pictures, video in the form of movies, news, programming, etc.), etc.].
  • Still yet, as an option, the media player may be used in conjunction (e.g. built-in, retrofitted with, coupled to, etc.) any desired device including, but not limited to a cellular phone, personal digital assistant, etc. (e.g. see, for example, any of the devices of FIG. 1, etc.). Of course, it is contemplated that the media player may further be a stand alone product.
  • Even still, the commands may include, for example, commands that operate the media player, commands that change states of the media player, etc. Specifically, for media players that play music, for example, such commands may include play, pause, fast forward, rewind, on, off, shuffle, repeat, search, volume up, volume down, playlist, next playlist, etc. As an example, the search command may provide a user with the ability to command the media player to search for a particular song or artist and the play command may provide the user with the ability to command a particular song to be played. In various embodiments, the commands may be programmable and/or registered by the media player. Of course, the command may include any signal, instruction, code, data, etc. that is capable of being utilized for providing wireless control of the media player. For example, the command may be an utterance that is translated into a hex code capable of being recognized by the media player. Table 1 illustrates examples of such hex codes capable of being recognized by the media player, in an embodiment where the media player includes the aforementioned iPod(R).
  • TABLE 1
    Shuffle = FF 55 04 02 00 00 80 7A,
    Pause = FF 55 03 02 00 01 FA
    Playlist = FF 55 04 02 00 00 40 BA
  • It should be noted that the operations of the method 300 of FIG. 3 may be carried out by the media player itself, and/or by way of a separate assembly that is either built-in the media player or capable of being retrofitted on the media player. One exemplary assembly will be described in more detail with respect to FIGS. 4-6.
  • More illustrative information will now be set forth regarding various optional architectures and features with which the foregoing method 300 may or may not be implemented, per the desires of the user. It should be strongly noted that the following information is set forth for illustrative purposes and should not be construed as limiting in any manner. Any of the following features may be optionally incorporated with or without the exclusion of other features described.
  • FIG. 4 shows a method for providing wireless control of a media player, in accordance with another embodiment. As an option, the present method 600 may be implemented in the context of the architecture and environment of FIGS. 1-3. Of course, however, the method 600 may be carried out in any desired environment.
  • First, program variables within an assembly attached to a media player are initiated, as shown in operation 602. The program variables initialized may include initializing trigger and command words along with any input or output utilized. An exemplary assembly will be described in more detail with respect to FIGS. 5 and 6. Next, the assembly listens for a trigger word, as in operation 604. The trigger word may be any word capable of being recognized and understood by the assembly. It may also optionally be any word capable of initiating a voice recognition capability of the assembly. For example, a user may be required to first say “start” in order to trigger the assembly.
  • Once it is determined that a trigger word is detected, as in operation 606, a command word is awaited, such as in operation 608. As an option, the command word may include any of the utterances described with respect to FIG. 3. Of course the command word may include any word capable of being recognized by the assembly and capable of being translated into computer code that the media player can act upon. As another option, a time limit (e.g. 120 seconds, etc.) may be utilized such that if no trigger word is detected during operation 606 within the time limit, the method may return to operation 604.
  • Next, upon determination that a command word has been detected, as in operation 610, the command is converted into computer code that can be read by the media player, as shown in operation 612. As an example, in one embodiment, the computer code may include computer code that can be read by the personal media player. However, if a specified period of time expires in operation 608, the determination of operation 610 may be false (or be more scrutinized, etc.) and the method may return to operation 604. In addition, if a word is detected but it is not a registered or programmed command word, the method may also return to operation 608.
  • The computer code is then sent from the assembly to the media player, as shown in operation 614, and a send button releases the command to the media player, as in operation 616. Such send button may indicate to the media player that the computer code is complete. Further, the computer code may be sent as a bit stream to the media player. For example, the bit stream may consist of seven codes of ten bits each. Still yet, each bit may be sent approximately every 55 uSecs with a code repetition rate of 66 (i.e. each code is sent about every 15 milliseconds).
  • It is next determined whether the command is a sleep command, as shown in operation 619. For example, such command may be created by a user uttering “sleep” and such utterance then being translated into computer code that tells the media player to power down. If the determination in operation 619 is positive, the media player is powered down and all program variables are released, as depicted in operation 620. Powering down may, as an option, include setting the media player to a low power mode. Alternatively, if the command is not for the media player to sleep, the method 600 returns to operation 602 where it continues to listen for a trigger word. Again, method 600 is set forth to illustrate just one example of a method for wireless controlling a media player, and should not be construed as limiting in any manner.
  • FIG. 5 shows a media player in connection with an assembly for receiving utterances, in accordance with another embodiment. As an option, the media player/assembly may be implemented in the context of the architecture and environment of FIGS. 1-4. Of course, however, the media player/assembly may be implemented in any desired environment.
  • A media player 401 is shown connected to an assembly 403. In the present embodiment, media player 401 is shown with the assembly 403 holding the media player 401. Of course such assembly 403 could be optionally mounted on the personal media player 401 or connected in any other manner. In any way, media player 401 is configured to connect to the assembly 403.
  • The assembly 403 includes electrical connection (not shown), voice activation software (not shown), voice activation hardware (not shown), a memory integrated circuit (not shown), an FM transmitter 409, and a power unit 402. The voice activation software may be capable of detecting an utterance, translating the utterance into computer code, and transmitting the code to the media player 401. In addition, the power unit 402 may be capable of charging the media player 401 and may include a plug that connects to an automobile cigarette lighter device or AC/DC converter device to provide a required voltage. Optionally, the connection may be located on the end of a flexible metal rod that supports the assembly.
  • The FM transmitter 409 may further include up and down arrows on the front of the assembly 403 as shown in FIG. 5 for sweeping across 88 MHz to 108 MHz such that a frequency that has little outside traffic may be locked in for transmitting a stereo signal at least six feet. FM transmitter 409 may also include left and right channel programming capabilities.
  • Additionally, a power LED 406 and charging LED 407 may be utilized for displaying whether the media player 401 is charging or listening for an utterance. Further included may be a verification speaker 404 for verifying utterances received, a directional microphone 405 for receiving and transferring utterances to a processing circuit (not shown) that translates the utterances into computer code capable of being read by the media player 401. The processing circuit may also include a voice engine, onboard memory, and a plurality of circuit peripherals.
  • In use, the utterance may be verified by replaying the utterance for the user who provided the utterance and allowing the user to either accept or reject the utterance. For example, in playing the utterance back for the user, the user may be prompted to either state “yes” or “no.” If the user rejects the utterance, the user may then be prompted to give the utterance again. In this way, the verifying may allow for the adjusting of the received utterance if it is not verified. Of course, any type of verification process may optionally be utilized for verifying the utterance.
  • Also included may be a connector (not shown) for sending and receiving data between the assembly 403 and the media player 401 as well as providing power to the media player 401. Still yet, a mounting brace 408, a front face 410, and an FM frequency display 411 may be provided.
  • As an option, output from the media player 401 may be sent through the assembly 403. Such output may be FM modulated with the FM transmitter 409 of the assembly 403 for reception by nearby FM receivers.
  • FIG. 6 shows a media player in connection with an assembly for receiving utterances, in accordance with still yet another embodiment. As an option, the present media player/assembly may be implemented in the context of the architecture and environment of FIGS. 1-5. Of course, however, the media player/assembly may be implemented in any desired environment.
  • In the present embodiment, the assembly 507 is shown connected to a media player 501 by being mounted on top of the media player 501. Such assembly 507 may optionally include a signal line out 502, a directional or omni-directional microphone 503, a power LED 504 and listening LED 505 (which indicates whether the assembly 507 is awaiting a command—see operation 608 of FIG. 4, for example), an FM frequency adjustment 506, and/or an FM frequency display 508. Such features may include the same functionalities described with respect to FIG. 5.
  • FIG. 7 shows a method 700 for providing wireless control of a media player when a library is loaded on the media player, in accordance with one embodiment. As an option, the present method 700 may be implemented in the context of the architecture and environment of FIGS. 1-6. Of course, however, the method 700 may be implemented in any desired environment.
  • First, program variables within an assembly attached to a media player (or within the media player itself) are initiated, as shown in operation 702. The program variables initialized may include initializing trigger and/or command words along with any input or output utilized. In an embodiment where a separate assembly is utilized, the assembly may, for example, include the embodiments described in FIGS. 5 and 6 above.
  • Next, the assembly listens for a trigger word, as in operation 704. The trigger word may be any word capable of being recognized and understood. It may also optionally be any word capable of initiating a voice recognition capability. For example, a user may be required to first say “start” in order to trigger the present embodiment.
  • Once it is determined that a trigger word is detected, as in decision 706, a command word is awaited, such as in operation 708. As an option, the command word may include any of the utterances described with respect to FIG. 3. Of course, the command word may include any word capable of being recognized and capable of being translated into computer code that the media player can act upon. In the present embodiment, if the command word received is “Search”, as shown in decision 710, the method 700 continues to listen for a next utterance in operation 712. The next utterance may be, for example, an artist name or any other word.
  • As another option, a time limit (e.g. 120 seconds, etc.) may be utilized such that if no utterance is detected during operation 712 within the time limit, the method 700 may terminate, as shown in decision 716. Otherwise, if an utterance is detected during operation 712 within the time limit, the method 700 may navigate to a first item in a library of words associated with the utterance received in operation 712. In the present embodiment, and shown just by way of example, the method 700 may navigate to a first item located in an artist library, as shown in operation 714. Of course, it should be noted that any type of data capable of being located within a library may be utilized. An artist name and voice tag associated with the first item in the library may then be read, such as in operation 718, and it may be determined whether the voice tag matches the utterance received in operation 712 (see decision 720).
  • If the voice tag associated with the first item in the library does not match the utterance received in operation 712, the method 700 may navigate to a next item located in the artist library, as shown in operation 724. If it is determined in decision 722 that there is not a next item in the artist library to navigate to (e.g. the method 700 has reached the end of the library), the method 700 may terminate. Otherwise, the method 700 may return to operation 718 where the artist name and voice tag associated with the next item are read.
  • The method 700 continues until it is determined in decision 720 that the voice tag matches the utterance received in operation 712, in which case a command is issued to the media player, such as, for example, a command to play a first among a set of songs associated with the artist name received in operation 712, as shown in operation 726.
  • FIG. 8 shows a method 800 for providing wireless control of a media player when a library is not loaded on the media player, in accordance with another embodiment. As an option, the present method 800 may be implemented in the context of the architecture and environment of FIGS. 1-6. Of course, however, the method 800 may be implemented in any desired environment.
  • First, program variables within an assembly attached to a media player (or within the media player itself) are initiated, as shown in operation 802. The program variables initialized may include initializing trigger and command words along with any input or output utilized. Next, the assembly listens for a trigger word, as in operation 804. The trigger word may be any word capable of being recognized and understood. It may also optionally be any word capable of initiating a voice recognition capability. For example, a user may be required to first say “start” in order to trigger the present embodiment.
  • Once it is determined that a trigger word is detected, as in decision 806, a command word is awaited, such as in operation 808. As an option, the command word may include any of the utterances described with respect to FIG. 3. Of course, the command word may include any word capable of being recognized and capable of being translated into computer code that the media player can act upon. In the present embodiment, if the command word received is “Search”, as shown in decision 810, the method 800 continues to listen for a next utterance in operation 812. The next utterance may be, for example, an artist name or any other word associated with a file located in a library loaded on the media player. Just by way of example, the file may be a song located in a library of songs loaded on the media player. Of course, it should be noted that any type of data capable of being located within a library may be utilized.
  • As another option, a time limit (e.g. 120 seconds, etc.) may be utilized such that if no utterance is detected during operation 812 within the time limit, the method 800 may terminate, as shown in decision 814. Otherwise, if an utterance is detected during operation 812 within the time limit, the method 800 converts the detected utterance into an application specific voice tag, as shown in operation 816.
  • Next, the method 800 may navigate to a first item in a library associated with the utterance received in operation 812. In the present embodiment, and shown just by way of example, the method 800 may navigate to a first item located in an artist library as shown in operation 818. For example, an artist name associated with the first item may be read, such as in operation 820, and the artist name may be converted to an application specific voice tag (see operation 822).
  • It is next determined in decision 824 whether the application specific voice tag generated from the first item in the library (see operation 822) matches the application specific voice tag generated from the utterance received in operation 812 (see operation 816). If the application specific voice tag associated with the first item in the library from operation 822 does not match the application specific voice tag associated with the utterance from operation 816, the method 800 may navigate to a next item located in the library, as shown in operation 828.
  • If it is determined in decision 826 that there is not a next item in the library to navigate to (e.g. the method 800 has reached the end of the library), the method 800 may terminate. Otherwise, the method 800 may return to operation 820 where the artist name and voice tag associated with the next item in the artist library are read.
  • The method 800 continues until it is determined in decision 824 that the voice tags match, in which case a command is issued to the media player, such as, for example, a command to play a song (or set of songs) associated with the artist name received in operation 812 (as shown in operation 830).
  • While various embodiments have been described above, it should be understood that they have been presented by way of example only, and not limitation. For example, any of the network elements may employ any of the desired functionality set forth hereinabove. Thus, the breadth and scope of a preferred embodiment should not be limited by any of the above-described exemplary embodiments, but should be defined only in accordance with the following claims and their equivalents.

Claims (20)

1. A sub-system, comprising:
logic for controlling an assembly coupled to an automobile for receiving power therefrom and further connectable to a satellite radio player adapted for playing music, news, and non-fiction information, the assembly including a speaker and a microphone;
logic for receiving a trigger signal;
logic for, after the receipt of the trigger signal, receiving an utterance utilizing the microphone of the assembly;
logic for verifying the utterance utilizing the speaker of the assembly;
logic for, after the verification of the utterance, generating a corresponding command for the satellite radio player based on the utterance, the corresponding command selected from a command set including a play command, a search command, an artist command, a volume up command, and a volume down command;
logic for channeling output of the satellite radio player;
wherein the corresponding command provides wireless control of the satellite radio player.
2. The sub-system of claim 1, wherein the trigger signal is audible.
3. The sub-system of claim 1, wherein the trigger signal includes an audible trigger word.
4. The sub-system of claim 1, wherein it is determined whether a timeout has occurred before the utterance is received.
5. The sub-system of claim 4, wherein if the timeout has occurred, the receiving of the trigger signal is repeated.
6. The sub-system of claim 1, wherein it is determined whether the utterance is verified.
7. The sub-system of claim 6, wherein it is determined whether the utterance is verified, based on whether the utterance is programmed.
8. The sub-system of claim 6, wherein it is determined whether the utterance is verified, based on whether the utterance is registered.
9. The sub-system of claim 6, wherein if the utterance is not verified, the corresponding command is not generated.
10. The sub-system of claim 6, wherein if the utterance is not verified, the utterance is outputted via the speaker.
11. The sub-system of claim 10, wherein a user is allowed to accept or reject the utterance outputted via the speaker.
12. The sub-system of claim 1, wherein the logic includes hardware logic.
13. The sub-system of claim 1, wherein the logic includes software logic.
14. The sub-system of claim 1, wherein the command set further includes a search command.
15. The sub-system of claim 1, wherein the utterance is compared against a library of words.
16. The sub-system of claim 15, wherein the corresponding command is generated in response to identifying a match between the utterance and the library of words.
17. The sub-system of claim 16, wherein the library of words includes an artist name and a voice tag.
18. The sub-system of claim 16, wherein the library of words includes an artist name and an application specific voice tag.
19. A system, comprising:
an automobile;
an assembly coupled to the automobile for receiving power therefrom and further connected to a satellite radio player adapted for playing music, news, and non-fiction information, the assembly including a power source for providing power to the satellite radio player, a speaker, and a microphone;
logic for receiving a trigger signal;
logic for, after the receipt of the trigger signal, receiving an utterance utilizing the microphone of the assembly;
logic for verifying the utterance utilizing the speaker of the assembly;
logic for, after the verification of the utterance, generating a corresponding command for the satellite radio player based on the utterance, the corresponding command selected from a command set including a play command, a search command, an artist command, a volume up command, and a volume down command;
logic for channeling output of the satellite radio player;
wherein the corresponding command provides wireless control of the satellite radio player.
20. A computer program product embodied on a computer readable medium, comprising:
computer code for controlling an assembly coupled to an automobile for receiving power therefrom and further connectable to a satellite radio player adapted for playing music, news, and non-fiction information, the assembly including a speaker and a microphone;
computer code for receiving of a trigger signal;
computer code for, after the receipt of the trigger signal, receiving an utterance utilizing the microphone of the assembly;
computer code for verifying the utterance utilizing the speaker of the assembly;
computer code for, after the verification of the utterance, generating a corresponding command for the satellite radio player based on the utterance, the corresponding command selected from a command set including a play command, a search command, an artist command, a volume up command, and a volume down command;
computer code for channeling output of the satellite radio player;
wherein the corresponding command provides wireless control of the satellite radio player.
US12/104,195 2005-07-11 2008-04-16 System, method and computer program product for adding voice activation and voice control to a media player Abandoned US20080195396A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US12/104,195 US20080195396A1 (en) 2005-07-11 2008-04-16 System, method and computer program product for adding voice activation and voice control to a media player

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US69859505P 2005-07-11 2005-07-11
US11/281,964 US7424431B2 (en) 2005-07-11 2005-11-16 System, method and computer program product for adding voice activation and voice control to a media player
US12/104,195 US20080195396A1 (en) 2005-07-11 2008-04-16 System, method and computer program product for adding voice activation and voice control to a media player

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/281,964 Division US7424431B2 (en) 2005-07-11 2005-11-16 System, method and computer program product for adding voice activation and voice control to a media player

Publications (1)

Publication Number Publication Date
US20080195396A1 true US20080195396A1 (en) 2008-08-14

Family

ID=37637638

Family Applications (5)

Application Number Title Priority Date Filing Date
US11/281,964 Expired - Fee Related US7424431B2 (en) 2005-07-11 2005-11-16 System, method and computer program product for adding voice activation and voice control to a media player
US12/104,207 Expired - Fee Related US7953599B2 (en) 2005-07-11 2008-04-16 System, method and computer program product for adding voice activation and voice control to a media player
US12/104,195 Abandoned US20080195396A1 (en) 2005-07-11 2008-04-16 System, method and computer program product for adding voice activation and voice control to a media player
US12/123,379 Expired - Fee Related US7567907B2 (en) 2005-07-11 2008-05-19 System, method and computer program product for adding voice activation and voice control to a media player
US13/090,225 Abandoned US20110196683A1 (en) 2005-07-11 2011-04-19 System, Method And Computer Program Product For Adding Voice Activation And Voice Control To A Media Player

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US11/281,964 Expired - Fee Related US7424431B2 (en) 2005-07-11 2005-11-16 System, method and computer program product for adding voice activation and voice control to a media player
US12/104,207 Expired - Fee Related US7953599B2 (en) 2005-07-11 2008-04-16 System, method and computer program product for adding voice activation and voice control to a media player

Family Applications After (2)

Application Number Title Priority Date Filing Date
US12/123,379 Expired - Fee Related US7567907B2 (en) 2005-07-11 2008-05-19 System, method and computer program product for adding voice activation and voice control to a media player
US13/090,225 Abandoned US20110196683A1 (en) 2005-07-11 2011-04-19 System, Method And Computer Program Product For Adding Voice Activation And Voice Control To A Media Player

Country Status (2)

Country Link
US (5) US7424431B2 (en)
WO (1) WO2007008248A2 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080215337A1 (en) * 2005-07-11 2008-09-04 Mark Greene System, method and computer program product for adding voice activation and voice control to a media player
US8073590B1 (en) 2008-08-22 2011-12-06 Boadin Technology, LLC System, method, and computer program product for utilizing a communication channel of a mobile device by a vehicular assembly
US8078397B1 (en) 2008-08-22 2011-12-13 Boadin Technology, LLC System, method, and computer program product for social networking utilizing a vehicular assembly
US8131458B1 (en) 2008-08-22 2012-03-06 Boadin Technology, LLC System, method, and computer program product for instant messaging utilizing a vehicular assembly
US8265862B1 (en) 2008-08-22 2012-09-11 Boadin Technology, LLC System, method, and computer program product for communicating location-related information
US9633659B1 (en) * 2016-01-20 2017-04-25 Motorola Mobility Llc Method and apparatus for voice enrolling an electronic computing device
US9916830B1 (en) * 2012-09-26 2018-03-13 Amazon Technologies, Inc. Altering audio to improve automatic speech recognition

Families Citing this family (58)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7398209B2 (en) 2002-06-03 2008-07-08 Voicebox Technologies, Inc. Systems and methods for responding to natural language speech utterance
US7693720B2 (en) 2002-07-15 2010-04-06 Voicebox Technologies, Inc. Mobile systems and methods for responding to natural language speech utterance
US7640160B2 (en) 2005-08-05 2009-12-29 Voicebox Technologies, Inc. Systems and methods for responding to natural language speech utterance
US7620549B2 (en) 2005-08-10 2009-11-17 Voicebox Technologies, Inc. System and method of supporting adaptive misrecognition in conversational speech
US7949529B2 (en) 2005-08-29 2011-05-24 Voicebox Technologies, Inc. Mobile systems and methods of supporting natural language human-machine interactions
WO2007027989A2 (en) 2005-08-31 2007-03-08 Voicebox Technologies, Inc. Dynamic speech sharpening
US7697827B2 (en) 2005-10-17 2010-04-13 Konicek Jeffrey C User-friendlier interfaces for a camera
US20070185601A1 (en) * 2006-02-07 2007-08-09 Apple Computer, Inc. Presentation of audible media in accommodation with external sound
KR100762636B1 (en) * 2006-02-14 2007-10-01 삼성전자주식회사 System and nethod for controlling voice detection of network terminal
AU2007297597B2 (en) * 2006-09-21 2013-02-21 Kyorin Pharmaceuticals Co., Ltd. Serine hydrolase inhibitors
US8073681B2 (en) 2006-10-16 2011-12-06 Voicebox Technologies, Inc. System and method for a cooperative conversational voice user interface
US20080091643A1 (en) * 2006-10-17 2008-04-17 Bellsouth Intellectual Property Corporation Audio Tagging, Browsing and Searching Stored Content Files
US7818176B2 (en) 2007-02-06 2010-10-19 Voicebox Technologies, Inc. System and method for selecting and presenting advertisements based on natural language processing of voice-based input
US7801729B2 (en) * 2007-03-13 2010-09-21 Sensory, Inc. Using multiple attributes to create a voice search playlist
US8140335B2 (en) 2007-12-11 2012-03-20 Voicebox Technologies, Inc. System and method for providing a natural language voice user interface in an integrated voice navigation services environment
US7933974B2 (en) * 2008-05-23 2011-04-26 Microsoft Corporation Media content for a mobile media device
US9177604B2 (en) * 2008-05-23 2015-11-03 Microsoft Technology Licensing, Llc Media content for a mobile media device
US8589161B2 (en) 2008-05-27 2013-11-19 Voicebox Technologies, Inc. System and method for an integrated, multi-modal, multi-device natural language voice services environment
US9305548B2 (en) 2008-05-27 2016-04-05 Voicebox Technologies Corporation System and method for an integrated, multi-modal, multi-device natural language voice services environment
US8832122B2 (en) * 2008-09-30 2014-09-09 Apple Inc. Media list management
US8326637B2 (en) 2009-02-20 2012-12-04 Voicebox Technologies, Inc. System and method for processing multi-modal device interactions in a natural language voice services environment
FR2948831B1 (en) * 2009-07-31 2022-01-28 Jerome Gilbert UNIVERSAL SYSTEM TO RECHARGE AT LEAST ONE PORTABLE DEVICE
US9171541B2 (en) 2009-11-10 2015-10-27 Voicebox Technologies Corporation System and method for hybrid processing in a natural language voice services environment
US9502025B2 (en) 2009-11-10 2016-11-22 Voicebox Technologies Corporation System and method for providing a natural language content dedication service
US8626511B2 (en) * 2010-01-22 2014-01-07 Google Inc. Multi-dimensional disambiguation of voice commands
US8996381B2 (en) * 2011-09-27 2015-03-31 Sensory, Incorporated Background speech recognition assistant
US8768707B2 (en) 2011-09-27 2014-07-01 Sensory Incorporated Background speech recognition assistant using speaker verification
US9031847B2 (en) * 2011-11-15 2015-05-12 Microsoft Technology Licensing, Llc Voice-controlled camera operations
US8818810B2 (en) 2011-12-29 2014-08-26 Robert Bosch Gmbh Speaker verification in a health monitoring system
US9317605B1 (en) 2012-03-21 2016-04-19 Google Inc. Presenting forked auto-completions
AU2013202129A1 (en) * 2012-04-04 2013-10-24 John Forrester Systems and methods for monitoring media interactions
US9436300B2 (en) * 2012-07-10 2016-09-06 Nokia Technologies Oy Method and apparatus for providing a multimodal user interface track
US9355174B2 (en) 2012-09-07 2016-05-31 Iheartmedia Management Services, Inc. Multi-input playlist selection
DE102013222930B4 (en) * 2012-11-12 2017-01-26 Nvidia Corporation Voice control on a mobile information device
DE102013001219B4 (en) 2013-01-25 2019-08-29 Inodyn Newmedia Gmbh Method and system for voice activation of a software agent from a standby mode
US9646606B2 (en) 2013-07-03 2017-05-09 Google Inc. Speech recognition using domain knowledge
US10003840B2 (en) 2014-04-07 2018-06-19 Spotify Ab System and method for providing watch-now functionality in a media content environment
US10134059B2 (en) 2014-05-05 2018-11-20 Spotify Ab System and method for delivering media content with music-styled advertisements, including use of tempo, genre, or mood
CN105334997A (en) * 2014-08-12 2016-02-17 扬智科技股份有限公司 Intelligent input apparatus as well as setting method and control method therefor
US9898459B2 (en) 2014-09-16 2018-02-20 Voicebox Technologies Corporation Integration of domain information into state transitions of a finite state transducer for natural language processing
US9626703B2 (en) 2014-09-16 2017-04-18 Voicebox Technologies Corporation Voice commerce
EP3207467A4 (en) 2014-10-15 2018-05-23 VoiceBox Technologies Corporation System and method for providing follow-up responses to prior natural language inputs of a user
US10614799B2 (en) 2014-11-26 2020-04-07 Voicebox Technologies Corporation System and method of providing intent predictions for an utterance prior to a system detection of an end of the utterance
US10431214B2 (en) 2014-11-26 2019-10-01 Voicebox Technologies Corporation System and method of determining a domain and/or an action related to a natural language input
US10956936B2 (en) 2014-12-30 2021-03-23 Spotify Ab System and method for providing enhanced user-sponsor interaction in a media environment, including support for shake action
US20160189249A1 (en) * 2014-12-30 2016-06-30 Spotify Ab System and method for delivering media content and advertisements across connected platforms, including use of companion advertisements
US9633661B1 (en) 2015-02-02 2017-04-25 Amazon Technologies, Inc. Speech-responsive portable speaker
WO2016145235A1 (en) * 2015-03-12 2016-09-15 Startimes Communication Network Technology Co. Ltd. Location based services audio system
US10235367B2 (en) * 2016-01-11 2019-03-19 Microsoft Technology Licensing, Llc Organization, retrieval, annotation and presentation of media data files using signals captured from a viewing environment
CN107197431A (en) * 2016-03-15 2017-09-22 西安中兴新软件有限责任公司 A kind of multi-medium play method and device
US10331784B2 (en) 2016-07-29 2019-06-25 Voicebox Technologies Corporation System and method of disambiguating natural language processing requests
US10304463B2 (en) 2016-10-03 2019-05-28 Google Llc Multi-user personalization at a voice interface device
US11127405B1 (en) * 2018-03-14 2021-09-21 Amazon Technologies, Inc. Selective requests for authentication for voice-based launching of applications
US10877637B1 (en) 2018-03-14 2020-12-29 Amazon Technologies, Inc. Voice-based device operation mode management
US10885910B1 (en) 2018-03-14 2021-01-05 Amazon Technologies, Inc. Voice-forward graphical user interface mode management
KR102612835B1 (en) * 2018-04-20 2023-12-13 삼성전자주식회사 Electronic device and method for executing function of electronic device
US11393478B2 (en) * 2018-12-12 2022-07-19 Sonos, Inc. User specific context switching
CN111081283A (en) * 2019-12-25 2020-04-28 惠州Tcl移动通信有限公司 Music playing method and device, storage medium and terminal equipment

Citations (99)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5404443A (en) * 1989-07-25 1995-04-04 Nissan Motor Company, Limited Display control system with touch switch panel for controlling on-board display for vehicle
US5644737A (en) * 1995-06-06 1997-07-01 Microsoft Corporation Method and system for stacking toolbars in a computer display
US5895464A (en) * 1997-04-30 1999-04-20 Eastman Kodak Company Computer program product and a method for using natural language for the description, search and retrieval of multi-media objects
US6038441A (en) * 1997-08-05 2000-03-14 Lear Corporation Method and system for creating records of voice messages in an automotive vehicle
US6073143A (en) * 1995-10-20 2000-06-06 Sanyo Electric Co., Ltd. Document conversion system including data monitoring means that adds tag information to hyperlink information and translates a document when such tag information is included in a document retrieval request
US6173266B1 (en) * 1997-05-06 2001-01-09 Speechworks International, Inc. System and method for developing interactive speech applications
US6192340B1 (en) * 1999-10-19 2001-02-20 Max Abecassis Integration of music from a personal library with real-time information
US6212498B1 (en) * 1997-03-28 2001-04-03 Dragon Systems, Inc. Enrollment in speech recognition
US6349257B1 (en) * 1999-09-15 2002-02-19 International Business Machines Corporation System for personalized mobile navigation information
US6397086B1 (en) * 1999-06-22 2002-05-28 E-Lead Electronic Co., Ltd. Hand-free operator capable of infrared controlling a vehicle's audio stereo system
US6405367B1 (en) * 1998-06-05 2002-06-11 Hewlett-Packard Company Apparatus and method for increasing the performance of Java programs running on a server
US6408272B1 (en) * 1999-04-12 2002-06-18 General Magic, Inc. Distributed voice user interface
US6407467B1 (en) * 1999-04-30 2002-06-18 Mannesmann Vdo Ag Built-in appliance intended for a motor vehicle
US20020095425A1 (en) * 2001-01-16 2002-07-18 Loay Abu-Husein Apparatus and method for updating applications to embedded devices and peripherals within a network environment
US20020151998A1 (en) * 2001-03-30 2002-10-17 Yrjo Kemppi Method and system for creating and presenting an individual audio information program
US6542812B1 (en) * 1999-10-19 2003-04-01 American Calcar Inc. Technique for effective navigation based on user preferences
US20030084404A1 (en) * 2001-10-30 2003-05-01 Dweck Jay S. Systems and methods for facilitating access to documents via a set of content selection tags
US20030126136A1 (en) * 2001-06-22 2003-07-03 Nosa Omoigui System and method for knowledge retrieval, management, delivery and presentation
US20030167174A1 (en) * 2002-03-01 2003-09-04 Koninlijke Philips Electronics N.V. Automatic audio recorder-player and operating method therefor
US20030229498A1 (en) * 2002-06-06 2003-12-11 International Business Machines Corporation Categorization and recall methodology for physical media in large carousel systems
US6697730B2 (en) * 2000-04-04 2004-02-24 Georgia Tech Research Corp. Communications and computing based urban transit system
US6711474B1 (en) * 2000-01-24 2004-03-23 G. Victor Treyz Automobile personal computer systems
US20040097272A1 (en) * 2001-01-28 2004-05-20 Guido Schuffert Hands-free device for operating mobile telephones in motor vehicles
US20040095260A1 (en) * 2002-11-20 2004-05-20 Nec Corporation Transport vehicle service guiding system, transport vehichle service guiding method, and transport vehicle service guiding program
US20040128141A1 (en) * 2002-11-12 2004-07-01 Fumihiko Murase System and program for reproducing information
US20040148362A1 (en) * 2001-11-02 2004-07-29 Lee Friedman Systems and methods for managing and aggregating media formats
US20040158607A1 (en) * 2003-02-06 2004-08-12 Coppinger Clifford L. System and method for associating an email attachment file with a storage location
US20040158746A1 (en) * 2003-02-07 2004-08-12 Limin Hu Automatic log-in processing and password management system for multiple target web sites
US6782240B1 (en) * 2000-04-27 2004-08-24 Joseph A Tabe Megatel communication information system
US6785656B2 (en) * 2001-06-05 2004-08-31 Xm Satellite Radio, Inc. Method and apparatus for digital audio playback using local stored content
US20050005242A1 (en) * 1998-07-17 2005-01-06 B.E. Technology, Llc Computer interface method and apparatus with portable network organization system and targeted advertising
US20050008167A1 (en) * 2003-04-30 2005-01-13 Achim Gleissner Device for picking up/reproducing audio signals
US20050027539A1 (en) * 2003-07-30 2005-02-03 Weber Dean C. Media center controller system and method
US20050043067A1 (en) * 2003-08-21 2005-02-24 Odell Thomas W. Voice recognition in a vehicle radio system
US20050065925A1 (en) * 2003-09-23 2005-03-24 Salesforce.Com, Inc. Query optimization in a multi-tenant database system
US20050065909A1 (en) * 2003-08-05 2005-03-24 Musgrove Timothy A. Product placement engine and method
US20050080772A1 (en) * 2003-10-09 2005-04-14 Jeremy Bem Using match confidence to adjust a performance threshold
US6885874B2 (en) * 2001-11-27 2005-04-26 Motorola, Inc. Group location and route sharing system for communication units in a trunked communication system
US20050131677A1 (en) * 2003-12-12 2005-06-16 Assadollahi Ramin O. Dialog driven personal information manager
US20050143134A1 (en) * 2003-12-30 2005-06-30 Lear Corporation Vehicular, hands-free telephone system
US20050165609A1 (en) * 1998-11-12 2005-07-28 Microsoft Corporation Speech recognition user interface
US20050179540A1 (en) * 2001-10-01 2005-08-18 Rubenstein Jeffrey D. Apparatus for communicating with a vehicle during remote vehicle operations, program product, and associated methods
US20050240588A1 (en) * 2004-04-26 2005-10-27 Siegel Hilliard B Method and system for managing access to media files
US6983203B1 (en) * 2000-07-28 2006-01-03 Alpine Electronics, Inc. POI icon display method and navigation system
US20060061458A1 (en) * 2004-09-21 2006-03-23 Gregory Simon Wireless vehicle control system and method
US20060075429A1 (en) * 2004-04-30 2006-04-06 Vulcan Inc. Voice control of television-related information
US7031724B2 (en) * 2003-03-12 2006-04-18 General Motors Corporation Location-based services for a telematics service subscriber
US20060085735A1 (en) * 2003-09-18 2006-04-20 Fujitsu Limited Annotation management system, annotation managing method, document transformation server, document transformation program, and electronic document attachment program
US20060095860A1 (en) * 2004-11-02 2006-05-04 Alan Wada Method and system of providing dynamic dialogs
US20060111893A1 (en) * 2004-11-19 2006-05-25 International Business Machines Corporation Display of results of cross language search
US20060155429A1 (en) * 2004-06-18 2006-07-13 Applied Digital, Inc. Vehicle entertainment and accessory control system
US20060167861A1 (en) * 2004-06-25 2006-07-27 Yan Arrouye Methods and systems for managing data
US20060184516A1 (en) * 2003-07-17 2006-08-17 Gerard Ellis Search engine
US20060195605A1 (en) * 2004-12-30 2006-08-31 Prabakar Sundarrajan Systems and methods for providing client-side accelerated access to remote applications via TCP buffering
US20060193450A1 (en) * 2005-02-25 2006-08-31 Microsoft Corporation Communication conversion between text and audio
US7158878B2 (en) * 2004-03-23 2007-01-02 Google Inc. Digital mapping system
US7162215B2 (en) * 2002-04-12 2007-01-09 General Motors Corporation Method and system for setting user preference satellite radio music selections in a mobile vehicle
US20070054704A1 (en) * 2005-09-06 2007-03-08 Takao Satoh Information addition system and mobile communication terminal
US20070061335A1 (en) * 2005-09-14 2007-03-15 Jorey Ramer Multimodal search query processing
US20070078950A1 (en) * 2002-06-13 2007-04-05 Salesforce.Com, Inc. Offline web services api to mirror online web services api
US7203721B1 (en) * 1999-10-08 2007-04-10 At Road, Inc. Portable browser device with voice recognition and feedback capability
US20070088741A1 (en) * 2005-09-09 2007-04-19 Salesforce.Com, Inc. Systems and methods for exporting, publishing, browsing and installing on-demand applications in a multi-tenant database environment
US7209929B2 (en) * 2003-04-17 2007-04-24 Salesforce.Com, Inc. Java object cache server for databases
US20070124276A1 (en) * 2003-09-23 2007-05-31 Salesforce.Com, Inc. Method of improving a query to a database system
US20070130130A1 (en) * 2005-12-02 2007-06-07 Salesforce.Com, Inc. Systems and methods for securing customer data in a multi-tenant environment
US20070130137A1 (en) * 2005-12-02 2007-06-07 Salesforce.Com, Inc. Methods and systems for optimizing text searches over structured data in a multi-tenant environment
US20070179800A1 (en) * 2005-12-31 2007-08-02 General Motors Corporation Email-based command interface for a telematics-equipped vehicle
US20070185843A1 (en) * 2006-01-23 2007-08-09 Chacha Search, Inc. Automated tool for human assisted mining and capturing of precise results
US20080010243A1 (en) * 2006-06-02 2008-01-10 Salesforce.Com, Inc. Method and system for pushing data to a plurality of devices in an on-demand service environment
US20080026793A1 (en) * 2006-07-27 2008-01-31 Microsoft Corporation Providing input and output for a mobile device
US20080027643A1 (en) * 2006-07-28 2008-01-31 Basir Otman A Vehicle communication system with navigation
US20080033714A1 (en) * 2002-08-07 2008-02-07 Itt Manufacturing Enterprises, Inc. Acronym Extraction System and Method of Identifying Acronyms and Extracting Corresponding Expansions from Text
US20080032721A1 (en) * 2006-08-04 2008-02-07 Gm Global Technology Operations, Inc. Method and system for communicating information to a user of a mobile platform via broadcast services
US20080036586A1 (en) * 2006-08-11 2008-02-14 Eric Shigeru Ohki Method and system for receiving and sending navigational data via a wireless messaging service on a navigation system
US20080046845A1 (en) * 2006-06-23 2008-02-21 Rohit Chandra Method and Apparatus for Controlling the Functionality of a Highlighting Service
US20080059447A1 (en) * 2006-08-24 2008-03-06 Spock Networks, Inc. System, method and computer program product for ranking profiles
US20080122695A1 (en) * 2004-06-16 2008-05-29 Jackson Kit Wang Systems and Methods For Geographical Positioning Using Radio Spectrum Signatures
US7392132B2 (en) * 2002-06-14 2008-06-24 Matsushita Electric Industrial Co., Ltd. Position notifying device
US7398209B2 (en) * 2002-06-03 2008-07-08 Voicebox Technologies, Inc. Systems and methods for responding to natural language speech utterance
US7493257B2 (en) * 2003-08-06 2009-02-17 Samsung Electronics Co., Ltd. Method and apparatus handling speech recognition errors in spoken dialogue systems
US7516191B2 (en) * 2001-03-26 2009-04-07 Salesforce.Com, Inc. System and method for invocation of services
US7567907B2 (en) * 2005-07-11 2009-07-28 Stragent, Llc System, method and computer program product for adding voice activation and voice control to a media player
US7685200B2 (en) * 2007-03-01 2010-03-23 Microsoft Corp Ranking and suggesting candidate objects
US7689711B2 (en) * 2001-03-26 2010-03-30 Salesforce.Com, Inc. System and method for routing messages between applications
US7706967B2 (en) * 1997-08-19 2010-04-27 Continental Automotive Systems Us, Inc. Vehicle information system
US7721328B2 (en) * 2004-10-01 2010-05-18 Salesforce.Com Inc. Application identity design
US7725605B2 (en) * 2004-08-06 2010-05-25 Salesforce.Com, Inc. Providing on-demand access to services in a wide area network
US7730478B2 (en) * 2006-10-04 2010-06-01 Salesforce.Com, Inc. Method and system for allowing access to developed applications via a multi-tenant on-demand database service
US7734608B2 (en) * 2006-09-22 2010-06-08 Salesforce.Com, Inc. System, method and computer program product for querying data relationships over a network
US7739351B2 (en) * 2004-03-23 2010-06-15 Salesforce.Com, Inc. Synchronous interface to asynchronous processes
US7761871B2 (en) * 2004-03-10 2010-07-20 Handmark, Inc. Data access architecture
US7765243B2 (en) * 2004-07-26 2010-07-27 Sandisk Il Ltd. Unified local-remote logical volume
US7865303B2 (en) * 2006-11-09 2011-01-04 General Motors Llc Method of providing a navigational route for a vehicle navigation system
US7869941B2 (en) * 2006-12-29 2011-01-11 Aol Inc. Meeting notification and modification service
US7880602B2 (en) * 2007-03-01 2011-02-01 Fujitsu Ten Limited Image display control apparatus
US7904882B2 (en) * 2003-10-16 2011-03-08 Salesforce.Com, Inc. Managing virtual business instances within a computer network
US7912629B2 (en) * 2007-11-30 2011-03-22 Nokia Corporation Methods, apparatuses, and computer program products for traffic data aggregation using virtual trip lines and a combination of location and time based measurement triggers in GPS-enabled mobile handsets
US7912628B2 (en) * 2006-03-03 2011-03-22 Inrix, Inc. Determining road traffic conditions using data from multiple data sources
US7921013B1 (en) * 2000-11-03 2011-04-05 At&T Intellectual Property Ii, L.P. System and method for sending multi-media messages using emoticons

Family Cites Families (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5774859A (en) * 1995-01-03 1998-06-30 Scientific-Atlanta, Inc. Information system having a speech interface
ES2146778T3 (en) * 1995-01-04 2000-08-16 Koninkl Philips Electronics Nv A CONTROL SYSTEM, VOICE OPERATED, FOR USE WITH A CONSUMER PRODUCT.
US5721783A (en) * 1995-06-07 1998-02-24 Anderson; James C. Hearing aid with wireless remote processor
DE19533541C1 (en) * 1995-09-11 1997-03-27 Daimler Benz Aerospace Ag Method for the automatic control of one or more devices by voice commands or by voice dialog in real time and device for executing the method
US5926789A (en) * 1996-12-19 1999-07-20 Bell Communications Research, Inc. Audio-based wide area information system
US5930751A (en) * 1997-05-30 1999-07-27 Lucent Technologies Inc. Method of implicit confirmation for automatic speech recognition
US6038534A (en) * 1997-09-11 2000-03-14 Cowboy Software, Inc. Mimicking voice commands as keyboard signals
DE69712485T2 (en) * 1997-10-23 2002-12-12 Sony Int Europe Gmbh Voice interface for a home network
US6298324B1 (en) * 1998-01-05 2001-10-02 Microsoft Corporation Speech recognition system with changing grammars and grammar help command
US6505159B1 (en) * 1998-03-03 2003-01-07 Microsoft Corporation Apparatus and method for providing speech input to a speech recognition system
US6067278A (en) 1998-04-06 2000-05-23 Recoton Corporation Digital recorder for car radio
EP0953486B1 (en) * 1998-04-30 2006-04-26 CLARION Co., Ltd. Automotive information system and method of controlling the same, recording medium storing control program, disk playback apparatus, and semiconductor integrated circuit
US6882286B1 (en) * 1999-04-20 2005-04-19 Funai Electric Co., Ltd. Remote controller and electrical apparatus controlled by the same
US6584439B1 (en) * 1999-05-21 2003-06-24 Winbond Electronics Corporation Method and apparatus for controlling voice controlled devices
US7548851B1 (en) * 1999-10-12 2009-06-16 Jack Lau Digital multimedia jukebox
US9076448B2 (en) * 1999-11-12 2015-07-07 Nuance Communications, Inc. Distributed real time speech recognition system
US6615172B1 (en) * 1999-11-12 2003-09-02 Phoenix Solutions, Inc. Intelligent query engine for processing voice based queries
US6598018B1 (en) * 1999-12-15 2003-07-22 Matsushita Electric Industrial Co., Ltd. Method for natural dialog interface to car devices
US6539358B1 (en) * 2000-05-24 2003-03-25 Delphi Technologies, Inc. Voice-interactive docking station for a portable computing device
US6513010B1 (en) * 2000-05-30 2003-01-28 Voxi Ab Method and apparatus for separating processing for language-understanding from an application and its functionality
JP3760755B2 (en) * 2000-10-11 2006-03-29 日産自動車株式会社 Voice input device
US6570495B1 (en) * 2000-10-27 2003-05-27 Audiovox Corporation Voice confirmation system and method for a vehicle
US20020095294A1 (en) * 2001-01-12 2002-07-18 Rick Korfin Voice user interface for controlling a consumer media data storage and playback device
US20030069734A1 (en) * 2001-10-05 2003-04-10 Everhart Charles Allen Technique for active voice recognition grammar adaptation for dynamic multimedia application
US7031477B1 (en) * 2002-01-25 2006-04-18 Matthew Rodger Mella Voice-controlled system for providing digital audio content in an automobile
US20030171834A1 (en) * 2002-03-07 2003-09-11 Silvester Kelan C. Method and apparatus for connecting a portable media player wirelessly to an automobile entertainment system
US20030187657A1 (en) * 2002-03-26 2003-10-02 Erhart George W. Voice control of streaming audio
DE10216117A1 (en) * 2002-04-12 2003-10-23 Philips Intellectual Property Symbol sequence voice recognition involves determining sub-symbol sequence to be corrected by comparing second sequence with sub-sequences longer or shorter than second sequence by number of symbols
US6950638B2 (en) * 2002-04-30 2005-09-27 General Motors Corporation Method and system for scheduling user preference satellite radio station selections in a mobile vehicle
US8713140B2 (en) * 2002-04-30 2014-04-29 General Motors Llc Method and system for modifying satellite radio program subscriptions in a mobile vehicle
JP4132962B2 (en) * 2002-05-16 2008-08-13 パイオニア株式会社 Interactive information providing apparatus, interactive information providing program, and storage medium storing the same
US20040225519A1 (en) * 2002-06-25 2004-11-11 Martin Keith D. Intelligent music track selection
US20050014536A1 (en) * 2003-07-08 2005-01-20 Jeff Grady Modular adaptor assembly for personal digital appliance
US6591085B1 (en) * 2002-07-17 2003-07-08 Netalog, Inc. FM transmitter and power supply/charging assembly for MP3 player
US7072686B1 (en) * 2002-08-09 2006-07-04 Avon Associates, Inc. Voice controlled multimedia and communications device
US6907397B2 (en) * 2002-09-16 2005-06-14 Matsushita Electric Industrial Co., Ltd. System and method of media file access and retrieval using speech recognition
US7395208B2 (en) * 2002-09-27 2008-07-01 Microsoft Corporation Integrating external voices
US6993615B2 (en) * 2002-11-15 2006-01-31 Microsoft Corporation Portable computing device-integrated appliance
US20040151327A1 (en) * 2002-12-11 2004-08-05 Ira Marlow Audio device integration system
CN1729276A (en) * 2002-12-19 2006-02-01 皇家飞利浦电子股份有限公司 Method and system for network downloading of music files
US20040143440A1 (en) * 2003-01-03 2004-07-22 Venkatesh Prasad Vehicle speech recognition system
US20040176959A1 (en) * 2003-03-05 2004-09-09 Wilhelm Andrew L. System and method for voice-enabling audio compact disc players via descriptive voice commands
US20040171346A1 (en) * 2003-04-14 2004-09-02 Ten Ventures Inc. Method of Adding Wireless Remote Control to Audio Playback Devices Equipped for Wired Remote Control
US7757173B2 (en) * 2003-07-18 2010-07-13 Apple Inc. Voice menu system
US7008346B2 (en) * 2003-08-25 2006-03-07 General Motors Corporation Seven-speed transmission
TWM241766U (en) 2003-09-03 2004-08-21 Formosa Teletek Corp Portable wireless communication device for media playing
TWI235358B (en) * 2003-11-21 2005-07-01 Acer Inc Interactive speech method and system thereof
CN1890708B (en) * 2003-12-05 2011-12-07 株式会社建伍 Audio device control device,audio device control method, and program
US20050138069A1 (en) 2003-12-19 2005-06-23 General Motors Corporation Providing a playlist package of digitized entertainment files for storage and playback
WO2005064592A1 (en) * 2003-12-26 2005-07-14 Kabushikikaisha Kenwood Device control device, speech recognition device, agent device, on-vehicle device control device, navigation device, audio device, device control method, speech recognition method, agent processing method, on-vehicle device control method, navigation method, and audio device control method, and program
US7346370B2 (en) * 2004-04-29 2008-03-18 Cellport Systems, Inc. Enabling interoperability between distributed devices using different communication link technologies
US20060041926A1 (en) * 2004-04-30 2006-02-23 Vulcan Inc. Voice control of multimedia content
US20060039263A1 (en) * 2004-08-20 2006-02-23 Remy Trotabas Audio system with removable digital media player
US20060089754A1 (en) * 2004-10-27 2006-04-27 Andrew Mortenson An installed Vehicle Personal Computing (VPC) system with touch interaction, voice interaction or sensor interaction(s) that provides access to multiple information sources and software applications such as internet connected data applications, dynamic traffic-aware navigational routing, vehicle tracking, emergency accident dispatching, business applications, office applications, music and video player(s), personal info portal, vehicle monitoring, alarm and camera security and recording.
US8942985B2 (en) * 2004-11-16 2015-01-27 Microsoft Corporation Centralized method and system for clarifying voice commands
US7668576B2 (en) * 2004-12-16 2010-02-23 Dashjack, Inc. Incorporating a portable digital music player into a vehicle audio system
US20060176660A1 (en) 2005-02-07 2006-08-10 Ahmad Amiri Ultra mobile communicating computer
US20060206339A1 (en) * 2005-03-11 2006-09-14 Silvera Marja M System and method for voice-enabled media content selection on mobile devices
US7826945B2 (en) * 2005-07-01 2010-11-02 You Zhang Automobile speech-recognition interface
US20090222270A2 (en) * 2006-02-14 2009-09-03 Ivc Inc. Voice command interface device

Patent Citations (103)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5404443A (en) * 1989-07-25 1995-04-04 Nissan Motor Company, Limited Display control system with touch switch panel for controlling on-board display for vehicle
US5644737A (en) * 1995-06-06 1997-07-01 Microsoft Corporation Method and system for stacking toolbars in a computer display
US6073143A (en) * 1995-10-20 2000-06-06 Sanyo Electric Co., Ltd. Document conversion system including data monitoring means that adds tag information to hyperlink information and translates a document when such tag information is included in a document retrieval request
US6212498B1 (en) * 1997-03-28 2001-04-03 Dragon Systems, Inc. Enrollment in speech recognition
US5895464A (en) * 1997-04-30 1999-04-20 Eastman Kodak Company Computer program product and a method for using natural language for the description, search and retrieval of multi-media objects
US6173266B1 (en) * 1997-05-06 2001-01-09 Speechworks International, Inc. System and method for developing interactive speech applications
US6038441A (en) * 1997-08-05 2000-03-14 Lear Corporation Method and system for creating records of voice messages in an automotive vehicle
US7706967B2 (en) * 1997-08-19 2010-04-27 Continental Automotive Systems Us, Inc. Vehicle information system
US6405367B1 (en) * 1998-06-05 2002-06-11 Hewlett-Packard Company Apparatus and method for increasing the performance of Java programs running on a server
US20050005242A1 (en) * 1998-07-17 2005-01-06 B.E. Technology, Llc Computer interface method and apparatus with portable network organization system and targeted advertising
US20050165609A1 (en) * 1998-11-12 2005-07-28 Microsoft Corporation Speech recognition user interface
US6965863B1 (en) * 1998-11-12 2005-11-15 Microsoft Corporation Speech recognition user interface
US6408272B1 (en) * 1999-04-12 2002-06-18 General Magic, Inc. Distributed voice user interface
US6407467B1 (en) * 1999-04-30 2002-06-18 Mannesmann Vdo Ag Built-in appliance intended for a motor vehicle
US6397086B1 (en) * 1999-06-22 2002-05-28 E-Lead Electronic Co., Ltd. Hand-free operator capable of infrared controlling a vehicle's audio stereo system
US6349257B1 (en) * 1999-09-15 2002-02-19 International Business Machines Corporation System for personalized mobile navigation information
US7203721B1 (en) * 1999-10-08 2007-04-10 At Road, Inc. Portable browser device with voice recognition and feedback capability
US6542812B1 (en) * 1999-10-19 2003-04-01 American Calcar Inc. Technique for effective navigation based on user preferences
US6192340B1 (en) * 1999-10-19 2001-02-20 Max Abecassis Integration of music from a personal library with real-time information
US6711474B1 (en) * 2000-01-24 2004-03-23 G. Victor Treyz Automobile personal computer systems
US6697730B2 (en) * 2000-04-04 2004-02-24 Georgia Tech Research Corp. Communications and computing based urban transit system
US6782240B1 (en) * 2000-04-27 2004-08-24 Joseph A Tabe Megatel communication information system
US6983203B1 (en) * 2000-07-28 2006-01-03 Alpine Electronics, Inc. POI icon display method and navigation system
US7921013B1 (en) * 2000-11-03 2011-04-05 At&T Intellectual Property Ii, L.P. System and method for sending multi-media messages using emoticons
US20020095425A1 (en) * 2001-01-16 2002-07-18 Loay Abu-Husein Apparatus and method for updating applications to embedded devices and peripherals within a network environment
US20040097272A1 (en) * 2001-01-28 2004-05-20 Guido Schuffert Hands-free device for operating mobile telephones in motor vehicles
US7516191B2 (en) * 2001-03-26 2009-04-07 Salesforce.Com, Inc. System and method for invocation of services
US7689711B2 (en) * 2001-03-26 2010-03-30 Salesforce.Com, Inc. System and method for routing messages between applications
US20020151998A1 (en) * 2001-03-30 2002-10-17 Yrjo Kemppi Method and system for creating and presenting an individual audio information program
US6785656B2 (en) * 2001-06-05 2004-08-31 Xm Satellite Radio, Inc. Method and apparatus for digital audio playback using local stored content
US20030126136A1 (en) * 2001-06-22 2003-07-03 Nosa Omoigui System and method for knowledge retrieval, management, delivery and presentation
US20050179540A1 (en) * 2001-10-01 2005-08-18 Rubenstein Jeffrey D. Apparatus for communicating with a vehicle during remote vehicle operations, program product, and associated methods
US20030084404A1 (en) * 2001-10-30 2003-05-01 Dweck Jay S. Systems and methods for facilitating access to documents via a set of content selection tags
US20040148362A1 (en) * 2001-11-02 2004-07-29 Lee Friedman Systems and methods for managing and aggregating media formats
US6885874B2 (en) * 2001-11-27 2005-04-26 Motorola, Inc. Group location and route sharing system for communication units in a trunked communication system
US20030167174A1 (en) * 2002-03-01 2003-09-04 Koninlijke Philips Electronics N.V. Automatic audio recorder-player and operating method therefor
US7162215B2 (en) * 2002-04-12 2007-01-09 General Motors Corporation Method and system for setting user preference satellite radio music selections in a mobile vehicle
US7398209B2 (en) * 2002-06-03 2008-07-08 Voicebox Technologies, Inc. Systems and methods for responding to natural language speech utterance
US20030229498A1 (en) * 2002-06-06 2003-12-11 International Business Machines Corporation Categorization and recall methodology for physical media in large carousel systems
US20070078950A1 (en) * 2002-06-13 2007-04-05 Salesforce.Com, Inc. Offline web services api to mirror online web services api
US7392132B2 (en) * 2002-06-14 2008-06-24 Matsushita Electric Industrial Co., Ltd. Position notifying device
US20080033714A1 (en) * 2002-08-07 2008-02-07 Itt Manufacturing Enterprises, Inc. Acronym Extraction System and Method of Identifying Acronyms and Extracting Corresponding Expansions from Text
US20040128141A1 (en) * 2002-11-12 2004-07-01 Fumihiko Murase System and program for reproducing information
US20040095260A1 (en) * 2002-11-20 2004-05-20 Nec Corporation Transport vehicle service guiding system, transport vehichle service guiding method, and transport vehicle service guiding program
US20040158607A1 (en) * 2003-02-06 2004-08-12 Coppinger Clifford L. System and method for associating an email attachment file with a storage location
US20040158746A1 (en) * 2003-02-07 2004-08-12 Limin Hu Automatic log-in processing and password management system for multiple target web sites
US7031724B2 (en) * 2003-03-12 2006-04-18 General Motors Corporation Location-based services for a telematics service subscriber
US7209929B2 (en) * 2003-04-17 2007-04-24 Salesforce.Com, Inc. Java object cache server for databases
US20050008167A1 (en) * 2003-04-30 2005-01-13 Achim Gleissner Device for picking up/reproducing audio signals
US20060184516A1 (en) * 2003-07-17 2006-08-17 Gerard Ellis Search engine
US20050027539A1 (en) * 2003-07-30 2005-02-03 Weber Dean C. Media center controller system and method
US20050065909A1 (en) * 2003-08-05 2005-03-24 Musgrove Timothy A. Product placement engine and method
US7493257B2 (en) * 2003-08-06 2009-02-17 Samsung Electronics Co., Ltd. Method and apparatus handling speech recognition errors in spoken dialogue systems
US20050043067A1 (en) * 2003-08-21 2005-02-24 Odell Thomas W. Voice recognition in a vehicle radio system
US20060085735A1 (en) * 2003-09-18 2006-04-20 Fujitsu Limited Annotation management system, annotation managing method, document transformation server, document transformation program, and electronic document attachment program
US20070124276A1 (en) * 2003-09-23 2007-05-31 Salesforce.Com, Inc. Method of improving a query to a database system
US20050065925A1 (en) * 2003-09-23 2005-03-24 Salesforce.Com, Inc. Query optimization in a multi-tenant database system
US7529728B2 (en) * 2003-09-23 2009-05-05 Salesforce.Com, Inc. Query optimization in a multi-tenant database system
US20050080772A1 (en) * 2003-10-09 2005-04-14 Jeremy Bem Using match confidence to adjust a performance threshold
US7904882B2 (en) * 2003-10-16 2011-03-08 Salesforce.Com, Inc. Managing virtual business instances within a computer network
US20050131677A1 (en) * 2003-12-12 2005-06-16 Assadollahi Ramin O. Dialog driven personal information manager
US20050143134A1 (en) * 2003-12-30 2005-06-30 Lear Corporation Vehicular, hands-free telephone system
US7761871B2 (en) * 2004-03-10 2010-07-20 Handmark, Inc. Data access architecture
US7158878B2 (en) * 2004-03-23 2007-01-02 Google Inc. Digital mapping system
US7739351B2 (en) * 2004-03-23 2010-06-15 Salesforce.Com, Inc. Synchronous interface to asynchronous processes
US20050240588A1 (en) * 2004-04-26 2005-10-27 Siegel Hilliard B Method and system for managing access to media files
US20060075429A1 (en) * 2004-04-30 2006-04-06 Vulcan Inc. Voice control of television-related information
US20080122695A1 (en) * 2004-06-16 2008-05-29 Jackson Kit Wang Systems and Methods For Geographical Positioning Using Radio Spectrum Signatures
US20060155429A1 (en) * 2004-06-18 2006-07-13 Applied Digital, Inc. Vehicle entertainment and accessory control system
US20060167861A1 (en) * 2004-06-25 2006-07-27 Yan Arrouye Methods and systems for managing data
US7765243B2 (en) * 2004-07-26 2010-07-27 Sandisk Il Ltd. Unified local-remote logical volume
US7725605B2 (en) * 2004-08-06 2010-05-25 Salesforce.Com, Inc. Providing on-demand access to services in a wide area network
US20060061458A1 (en) * 2004-09-21 2006-03-23 Gregory Simon Wireless vehicle control system and method
US7721328B2 (en) * 2004-10-01 2010-05-18 Salesforce.Com Inc. Application identity design
US20060095860A1 (en) * 2004-11-02 2006-05-04 Alan Wada Method and system of providing dynamic dialogs
US20060111893A1 (en) * 2004-11-19 2006-05-25 International Business Machines Corporation Display of results of cross language search
US20060195605A1 (en) * 2004-12-30 2006-08-31 Prabakar Sundarrajan Systems and methods for providing client-side accelerated access to remote applications via TCP buffering
US20060193450A1 (en) * 2005-02-25 2006-08-31 Microsoft Corporation Communication conversion between text and audio
US7953599B2 (en) * 2005-07-11 2011-05-31 Stragent, Llc System, method and computer program product for adding voice activation and voice control to a media player
US7567907B2 (en) * 2005-07-11 2009-07-28 Stragent, Llc System, method and computer program product for adding voice activation and voice control to a media player
US20070054704A1 (en) * 2005-09-06 2007-03-08 Takao Satoh Information addition system and mobile communication terminal
US20070088741A1 (en) * 2005-09-09 2007-04-19 Salesforce.Com, Inc. Systems and methods for exporting, publishing, browsing and installing on-demand applications in a multi-tenant database environment
US20070061335A1 (en) * 2005-09-14 2007-03-15 Jorey Ramer Multimodal search query processing
US20070130130A1 (en) * 2005-12-02 2007-06-07 Salesforce.Com, Inc. Systems and methods for securing customer data in a multi-tenant environment
US20070130137A1 (en) * 2005-12-02 2007-06-07 Salesforce.Com, Inc. Methods and systems for optimizing text searches over structured data in a multi-tenant environment
US20070179800A1 (en) * 2005-12-31 2007-08-02 General Motors Corporation Email-based command interface for a telematics-equipped vehicle
US20070185843A1 (en) * 2006-01-23 2007-08-09 Chacha Search, Inc. Automated tool for human assisted mining and capturing of precise results
US7912628B2 (en) * 2006-03-03 2011-03-22 Inrix, Inc. Determining road traffic conditions using data from multiple data sources
US20080010243A1 (en) * 2006-06-02 2008-01-10 Salesforce.Com, Inc. Method and system for pushing data to a plurality of devices in an on-demand service environment
US20080046845A1 (en) * 2006-06-23 2008-02-21 Rohit Chandra Method and Apparatus for Controlling the Functionality of a Highlighting Service
US20080026793A1 (en) * 2006-07-27 2008-01-31 Microsoft Corporation Providing input and output for a mobile device
US20080027643A1 (en) * 2006-07-28 2008-01-31 Basir Otman A Vehicle communication system with navigation
US20080032721A1 (en) * 2006-08-04 2008-02-07 Gm Global Technology Operations, Inc. Method and system for communicating information to a user of a mobile platform via broadcast services
US7646296B2 (en) * 2006-08-11 2010-01-12 Honda Motor Co., Ltd. Method and system for receiving and sending navigational data via a wireless messaging service on a navigation system
US20080036586A1 (en) * 2006-08-11 2008-02-14 Eric Shigeru Ohki Method and system for receiving and sending navigational data via a wireless messaging service on a navigation system
US20080059447A1 (en) * 2006-08-24 2008-03-06 Spock Networks, Inc. System, method and computer program product for ranking profiles
US7734608B2 (en) * 2006-09-22 2010-06-08 Salesforce.Com, Inc. System, method and computer program product for querying data relationships over a network
US7730478B2 (en) * 2006-10-04 2010-06-01 Salesforce.Com, Inc. Method and system for allowing access to developed applications via a multi-tenant on-demand database service
US7865303B2 (en) * 2006-11-09 2011-01-04 General Motors Llc Method of providing a navigational route for a vehicle navigation system
US7869941B2 (en) * 2006-12-29 2011-01-11 Aol Inc. Meeting notification and modification service
US7880602B2 (en) * 2007-03-01 2011-02-01 Fujitsu Ten Limited Image display control apparatus
US7685200B2 (en) * 2007-03-01 2010-03-23 Microsoft Corp Ranking and suggesting candidate objects
US7912629B2 (en) * 2007-11-30 2011-03-22 Nokia Corporation Methods, apparatuses, and computer program products for traffic data aggregation using virtual trip lines and a combination of location and time based measurement triggers in GPS-enabled mobile handsets

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Microsoft Plus! Voice Command for Windows Media Player, 25 October 2001, 26 Pages. *

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080215337A1 (en) * 2005-07-11 2008-09-04 Mark Greene System, method and computer program product for adding voice activation and voice control to a media player
US7953599B2 (en) * 2005-07-11 2011-05-31 Stragent, Llc System, method and computer program product for adding voice activation and voice control to a media player
US20110196683A1 (en) * 2005-07-11 2011-08-11 Stragent, Llc System, Method And Computer Program Product For Adding Voice Activation And Voice Control To A Media Player
US8073590B1 (en) 2008-08-22 2011-12-06 Boadin Technology, LLC System, method, and computer program product for utilizing a communication channel of a mobile device by a vehicular assembly
US8078397B1 (en) 2008-08-22 2011-12-13 Boadin Technology, LLC System, method, and computer program product for social networking utilizing a vehicular assembly
US8131458B1 (en) 2008-08-22 2012-03-06 Boadin Technology, LLC System, method, and computer program product for instant messaging utilizing a vehicular assembly
US8265862B1 (en) 2008-08-22 2012-09-11 Boadin Technology, LLC System, method, and computer program product for communicating location-related information
US9916830B1 (en) * 2012-09-26 2018-03-13 Amazon Technologies, Inc. Altering audio to improve automatic speech recognition
US10354649B2 (en) * 2012-09-26 2019-07-16 Amazon Technologies, Inc. Altering audio to improve automatic speech recognition
US11488591B1 (en) 2012-09-26 2022-11-01 Amazon Technologies, Inc. Altering audio to improve automatic speech recognition
US9633659B1 (en) * 2016-01-20 2017-04-25 Motorola Mobility Llc Method and apparatus for voice enrolling an electronic computing device

Also Published As

Publication number Publication date
US20110196683A1 (en) 2011-08-11
US7953599B2 (en) 2011-05-31
WO2007008248A2 (en) 2007-01-18
US7567907B2 (en) 2009-07-28
US20080215337A1 (en) 2008-09-04
US20070011007A1 (en) 2007-01-11
US20080221888A1 (en) 2008-09-11
WO2007008248A3 (en) 2007-12-13
US7424431B2 (en) 2008-09-09

Similar Documents

Publication Publication Date Title
US7567907B2 (en) System, method and computer program product for adding voice activation and voice control to a media player
US11211076B2 (en) Key phrase detection with audio watermarking
US6061306A (en) Portable digital player compatible with a cassette player
EP2267695B1 (en) Controlling music players using environment audio analysis
US5255326A (en) Interactive audio control system
US20070121981A1 (en) Portable speaker of portable multimedia device
JP2002186079A (en) Non-connected personal on-demand audio entertainment system permitting wireless downloading of contents
CN100410907C (en) Music play control apparatus using at least one earphone function key and method thereof
CN101411072A (en) Method and apparatus for wirelessly streaming multi-channel content
CN109473095A (en) A kind of intelligent home control system and control method
KR20070043986A (en) Method for data transmission utilizing a portable multimedia device
US20030182003A1 (en) Playback apparatus, headphone, and playback method
US20070060195A1 (en) Communication apparatus for playing sound signals
CN112789868B (en) Bluetooth speaker configured to produce sound and to function as both a receiver and a source
CN104796816A (en) Automatic-switching WIFI intelligent sound box
US7561698B1 (en) Auto shutoff for portable audio devices
CN105529029A (en) Audio control device, earphone and portable equipment
US20040254661A1 (en) Wireless sound apparatus having a MP3 encoder/decoder/recorder IC unit
US11450347B2 (en) Data group playback device and system and method for same
WO2018058997A1 (en) Networked speakers and system thereof
KR100649985B1 (en) Apparatus and method for using voice recognition remote controller in mobile communication terminal
US9635449B2 (en) Active earphone authentication method
CN1988733A (en) Wireless earphone capable of remote control portable information processor
CN104811850A (en) WIFI based intelligent sound box
US20020141092A1 (en) Data reproducing apparatus and data reproducing system

Legal Events

Date Code Title Description
AS Assignment

Owner name: STRAGENT, LLC, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:VOICEDEMAND, INC.;REEL/FRAME:020815/0571

Effective date: 20071029

AS Assignment

Owner name: SEESAW FOUNDATION, TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:STRAGENT, LLC;REEL/FRAME:026365/0075

Effective date: 20110524

AS Assignment

Owner name: STRAGENT, LLC, TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:STRAGENT, LLC;REEL/FRAME:033967/0138

Effective date: 20080829

Owner name: SEESAW FOUNDATION, TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:STRAGENT, LLC;REEL/FRAME:033967/0150

Effective date: 20110524

Owner name: VOICEDEMAND, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GREENE, MARK;HEGARTY, MICHAEL;CANTWELL, DERMOT;REEL/FRAME:033967/0130

Effective date: 20051115

Owner name: STRAGENT, LLC, TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TAG FOUNDATION;REEL/FRAME:033967/0159

Effective date: 20141001

Owner name: TAG FOUNDATION, TEXAS

Free format text: CHANGE OF NAME;ASSIGNOR:SEESAW FOUNDATION;REEL/FRAME:034012/0764

Effective date: 20111012

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: KILPAUK GARDEN, SERIES 64 OF ALLIED SECURITY TRUST

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:STRAGENT, LLC;REEL/FRAME:034411/0025

Effective date: 20141111

AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KILPAUK GARDEN, SERIES 64 OF ALLIED SECURITY TRUST I;REEL/FRAME:036565/0060

Effective date: 20150803