US20170188437A1 - Voice-Controlled Light Switches - Google Patents
Voice-Controlled Light Switches Download PDFInfo
- Publication number
- US20170188437A1 US20170188437A1 US14/980,392 US201514980392A US2017188437A1 US 20170188437 A1 US20170188437 A1 US 20170188437A1 US 201514980392 A US201514980392 A US 201514980392A US 2017188437 A1 US2017188437 A1 US 2017188437A1
- Authority
- US
- United States
- Prior art keywords
- microphone
- audio signal
- switch
- faceplate
- recited
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000005236 sound signal Effects 0.000 claims description 188
- 238000012545 processing Methods 0.000 claims description 61
- 238000000034 method Methods 0.000 claims description 53
- 230000004044 response Effects 0.000 claims description 13
- 230000008878 coupling Effects 0.000 claims description 4
- 238000010168 coupling process Methods 0.000 claims description 4
- 238000005859 coupling reaction Methods 0.000 claims description 4
- 238000001228 spectrum Methods 0.000 abstract description 2
- 238000005406 washing Methods 0.000 abstract description 2
- 230000008569 process Effects 0.000 description 33
- 230000000875 corresponding effect Effects 0.000 description 24
- 230000015654 memory Effects 0.000 description 14
- 238000010586 diagram Methods 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 3
- 230000002085 persistent effect Effects 0.000 description 3
- 230000001413 cellular effect Effects 0.000 description 2
- 239000000835 fiber Substances 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- NOESYZHRGYRDHS-UHFFFAOYSA-N insulin Chemical compound N1C(=O)C(NC(=O)C(CCC(N)=O)NC(=O)C(CCC(O)=O)NC(=O)C(C(C)C)NC(=O)C(NC(=O)CN)C(C)CC)CSSCC(C(NC(CO)C(=O)NC(CC(C)C)C(=O)NC(CC=2C=CC(O)=CC=2)C(=O)NC(CCC(N)=O)C(=O)NC(CC(C)C)C(=O)NC(CCC(O)=O)C(=O)NC(CC(N)=O)C(=O)NC(CC=2C=CC(O)=CC=2)C(=O)NC(CSSCC(NC(=O)C(C(C)C)NC(=O)C(CC(C)C)NC(=O)C(CC=2C=CC(O)=CC=2)NC(=O)C(CC(C)C)NC(=O)C(C)NC(=O)C(CCC(O)=O)NC(=O)C(C(C)C)NC(=O)C(CC(C)C)NC(=O)C(CC=2NC=NC=2)NC(=O)C(CO)NC(=O)CNC2=O)C(=O)NCC(=O)NC(CCC(O)=O)C(=O)NC(CCCNC(N)=N)C(=O)NCC(=O)NC(CC=3C=CC=CC=3)C(=O)NC(CC=3C=CC=CC=3)C(=O)NC(CC=3C=CC(O)=CC=3)C(=O)NC(C(C)O)C(=O)N3C(CCC3)C(=O)NC(CCCCN)C(=O)NC(C)C(O)=O)C(=O)NC(CC(N)=O)C(O)=O)=O)NC(=O)C(C(C)CC)NC(=O)C(CO)NC(=O)C(C(C)O)NC(=O)C1CSSCC2NC(=O)C(CC(C)C)NC(=O)C(NC(=O)C(CCC(N)=O)NC(=O)C(CC(N)=O)NC(=O)C(NC(=O)C(N)CC=1C=CC=CC=1)C(C)C)CC1=CN=CN1 NOESYZHRGYRDHS-UHFFFAOYSA-N 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 230000007704 transition Effects 0.000 description 2
- 241000282326 Felis catus Species 0.000 description 1
- 102000004877 Insulin Human genes 0.000 description 1
- 108090001061 Insulin Proteins 0.000 description 1
- 241000699670 Mus sp. Species 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 238000003490 calendering Methods 0.000 description 1
- 239000003990 capacitor Substances 0.000 description 1
- 238000004140 cleaning Methods 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 229940125396 insulin Drugs 0.000 description 1
- 230000005291 magnetic effect Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000035755 proliferation Effects 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 239000000779 smoke Substances 0.000 description 1
- 230000001755 vocal effect Effects 0.000 description 1
Images
Classifications
-
- H05B37/0236—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
- H04L12/2803—Home automation networks
- H04L12/2816—Controlling appliance services of a home automation network by calling their functionalities
- H04L12/282—Controlling appliance services of a home automation network by calling their functionalities based on user interaction within the home
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/30—Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
- H04L12/2803—Home automation networks
- H04L12/2823—Reporting information sensed by appliance or service execution status of appliance services in a home automation network
- H04L12/2825—Reporting to a device located outside the home and the home network
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/02—Casings; Cabinets ; Supports therefor; Mountings therein
- H04R1/028—Casings; Cabinets ; Supports therefor; Mountings therein associated with devices performing functions other than acoustics, e.g. electric candles
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/20—Arrangements for obtaining desired frequency or directional characteristics
- H04R1/32—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only
- H04R1/40—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
- H04R1/406—Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers microphones
-
- H05B37/0272—
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05B—ELECTRIC HEATING; ELECTRIC LIGHT SOURCES NOT OTHERWISE PROVIDED FOR; CIRCUIT ARRANGEMENTS FOR ELECTRIC LIGHT SOURCES, IN GENERAL
- H05B47/00—Circuit arrangements for operating light sources in general, i.e. where the type of light source is not relevant
- H05B47/10—Controlling the light source
- H05B47/105—Controlling the light source in response to determined parameters
- H05B47/115—Controlling the light source in response to determined parameters by determining the presence or movement of objects or living beings
- H05B47/12—Controlling the light source in response to determined parameters by determining the presence or movement of objects or living beings by detecting audible sound
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05B—ELECTRIC HEATING; ELECTRIC LIGHT SOURCES NOT OTHERWISE PROVIDED FOR; CIRCUIT ARRANGEMENTS FOR ELECTRIC LIGHT SOURCES, IN GENERAL
- H05B47/00—Circuit arrangements for operating light sources in general, i.e. where the type of light source is not relevant
- H05B47/10—Controlling the light source
- H05B47/175—Controlling the light source by remote control
- H05B47/19—Controlling the light source by remote control via wireless transmission
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05B—ELECTRIC HEATING; ELECTRIC LIGHT SOURCES NOT OTHERWISE PROVIDED FOR; CIRCUIT ARRANGEMENTS FOR ELECTRIC LIGHT SOURCES, IN GENERAL
- H05B47/00—Circuit arrangements for operating light sources in general, i.e. where the type of light source is not relevant
- H05B47/10—Controlling the light source
- H05B47/175—Controlling the light source by remote control
- H05B47/196—Controlling the light source by remote control characterised by user interface arrangements
- H05B47/197—Sound control or voice control
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2201/00—Details of transducers, loudspeakers or microphones covered by H04R1/00 but not provided for in any of its subgroups
- H04R2201/02—Details casings, cabinets or mounting therein for transducers covered by H04R1/02 but not provided for in any of its subgroups
- H04R2201/021—Transducers or their casings adapted for mounting in or to a wall or ceiling
Definitions
- Homes are becoming more wired and connected with the proliferation of computing devices such as desktops, tablets, entertainment systems, and portable communication devices.
- computing devices such as desktops, tablets, entertainment systems, and portable communication devices.
- many different ways have been introduced to allow users to interact with these devices, such as through mechanical means (e.g., keyboards, mice, etc.), touch screens, motion, and gesture.
- Another way to interact with computing devices is through speech.
- FIG. 1 shows an illustrative voice interaction computing architecture set in a home environment.
- the architecture includes at least one voice-controlled light switch and a voice-controlled device physically situated in the home, along with a user.
- the light switch includes one or more microphones for generating an audio signal that may include a command of the user, such as a command to turn on the lights in the illustrated room.
- FIG. 2 illustrates an example voice-controlled light switch that includes a microphone on a distal end of a switch (or “toggle”) of the light switch, and microphones on the faceplate of the light switch.
- the light switch may include functionality to determine whether the switch in a first position (e.g., the up position) or a second position (e.g., the down position) and may select audio signals to process based on this position. For instance, the if the switch is in the first position, the light switch may select, for processing, the audio signals generated by the microphone on the switch and one or more of the microphones on the top section of the faceplate, which may allow for more accurate speech processing as described below.
- FIG. 3 illustrates another example configuration of a voice-controlled light switch.
- the light switch is in the form of a rocker switch, and includes two microphones on the switch (or “rocker”) and microphones on the top and bottom of the faceplate.
- FIG. 4 illustrates yet another example configuration of a voice-controlled light switch.
- the light switch is in the form of a sliding dimmer switch, and includes two microphones on the switch (or “sliding knob”) that is configured to move to different positions of the dimmer switch, as well as microphones on the left and ride sides of the faceplate.
- FIG. 5 illustrates yet another example configuration of a voice-controlled light switch.
- the light switch is in the form of a rotatable-knob dimmer switch, and includes a microphone on the switch (or “rotatable knob”) that is configured to move radially to different positions, as well as microphones distributed across the faceplate of the dimmer switch.
- FIG. 6 illustrates a flow diagram of an example process for selecting which audio signals to process based on a position of a switch in a voice-controlled light switch.
- Processing the signals may include sending these signals to a remote service for further processing, such as performing beamforming on the signals to generate a processed audio signal and performing speech recognition on the processed audio signal. Additionally or alternatively, processing these signals may include performing beamforming, noise-cancelation, or the like locally at the switch or at another device in the environment, and/or performing speech recognition locally at the switch or at another device in the environment.
- FIG. 7 illustrates a flow diagram of an example process for selecting audio signals for processing based on a position of a switch, as well as based on a position of an audio source (e.g., a speaking user) within the environment.
- an audio source e.g., a speaking user
- FIG. 8 shows a block diagram of selected functional components implemented in the voice-controlled light switch of FIG. 1 .
- This disclosure describes, in part, voice-controlled light switches that act as voice-controlled endpoints at which users may provide voice commands.
- These light switches may physically couple to a power source, as well as to one or more power loads, such as appliances (e.g., lights, outlets, home appliances, etc.).
- the light switches may include physical switches to selectively provide power to the power loads, either in a binary fashion (on or off) or along of spectrum of positions.
- the power loads coupled to the light switches may include lights, electrical outlets, home appliances (e.g., dishwashers, washing machines, smoke detectors, security alarms, automated door locks, automated window shades, thermostats, etc.), factory machinery, healthcare devices (e.g., insulin pumps, CPAP machines, etc.), cleaning devices (e.g., autonomous robotic vacuum cleaners, etc.), or any other device that receives electrical power.
- home appliances e.g., dishwashers, washing machines, smoke detectors, security alarms, automated door locks, automated window shades, thermostats, etc.
- healthcare devices e.g., insulin pumps, CPAP machines, etc.
- cleaning devices e.g., autonomous robotic vacuum cleaners, etc.
- the voice-controlled light switches described herein include one or more microphones for generating audio signals, from which one or more voice commands may be identified.
- a light switch may include a microphone on the physical switch, on the faceplate, and/or the like, with this microphone generating audio signals based on sound detected within the environment of the light switch.
- the light switch may perform speech recognition on the generated audio signal to identify a voice command, or the light switch may include a network interface for sending the audio signal to another device or to a remote service, which may perform the speech processing.
- the power source of the light switch may comprise a persistent power source, such as the primary alternating current (AC) power supply connected to a home or other building.
- AC primary alternating current
- this power source may be used to power the microphones and other components of the light switch, in addition to the power load to which the light switch couples.
- the light switch may couple to the AC power supply, in other instances the light switch may couple to any other type of power source, such as batteries, capacitors, solar panels, or the like.
- the voice-controlled light switch includes a speech-recognition engine configured to identify a predefined utterance (e.g., a predefined sound or a predefined word or phrase) from a generated audio signal and, upon identifying this utterance, may begin providing generated audio signals to a remote service for performing speech recognition. Therefore, a user may determine when the audio signals are sent and when voice commands will be identified by speaking the predefined utterance (e.g., “wake up”).
- the voice commands may be for an array of activities, such as turning on the light in the user's environment, a request for information (e.g., “what is the temperature outside?”), a request to perform a task (“remind me to pick up Johnny at 4 pm”), or any other type of request.
- the voice-controlled light switches include multiple microphones, each configured to generate a respective audio signal. Together, these multiple audio signals may be processed to generate a processed audio signal, having a signal-to-noise (SNR) ratio that is greater than the SNR ratio of any singular one of the audio signals.
- the light switch, or a remote service may utilize the multiple audio signals to generate a processed audio signal using beamforming techniques, noise-cancellation techniques, or the like.
- the voice-controlled light switch, the remote service, or another device may then perform speech recognition on the processed audio signal having the increased SNR, thus increasing accuracy and efficacy of the speech recognition.
- the voice-controlled light switch includes at least a first microphone on a switch of the light switch. That is, the first microphone may reside on the physical control used to cause the power load coupled to the light switch to couple to the power source and, hence, receive power, and may be used to electrically decouple the power load from the power source such that the power load does not receive power from the power source.
- the voice-controlled light switch may comprise a toggle light switch, a rocker light switch, a sliding-dimmer light switch, a rotatable-knob light switch, or any other type of switch.
- the light switch may include one or more microphones located on the faceplate of the light switch or any other location on or adjacent to the light switch.
- the light switch when the light switch comprises a toggle light switch configured to reside in an up position and a down position, the light switch may include a second microphone on a top portion of the faceplate, as well as a third microphone located on a bottom portion of the faceplate.
- the second and third microphones are substantially in-line with the first microphone in a horizontal direction.
- the light switch may include operating logic configured to determine when the switch resides in the up position and when the switch resides in the down position. In response to determining that the switch resides in the up position, the light switch may select for processing a first audio signal generated by the first microphone and a second audio signal generated by the second microphone (on the top portion of the faceplate). This processing may include performing beamforming or other processing locally on the signals or sending the signals to the remote service for performing beamforming or other processing techniques thereon. Given the locations of the first and second microphones, when the switch is in the up position these microphones may be substantially in-line with one another in the vertical direction.
- these microphones may also be substantially vertically in-line with a mouth of a speaking user.
- Using audio signals generated from microphones in this configuration is particularly helpful when performing beamforming on these audio signals to generate a processed audio signal having a relatively high SNR.
- selecting the first and second audio signals for processing when the switch resides in the up position may result in particularly accurate speech recognition.
- the first and second microphones are in the same substantial horizontal plane (i.e., are substantially the same vertical height off of the floor), when the user speaks in the room towards the wall/light switch, the audio of the user may reach the first microphone and then the second microphone which is substantially vertically in-line with the first microphone. This configuration may result in effective beamforming or other signal processing.
- the operating logic may select for processing the first audio signal (generated by the first microphone on the switch) and a third audio signal generated by the third microphone on the bottom portion of the faceplate.
- these microphones are substantially in-line with one another in the vertical direction when the switch is in the down position (as well as substantially vertically in-line with a mouth of a speaking user)
- these audio signals represent excellent candidates for performing beamforming thereon to generate a processed audio signal having a relatively high SNR.
- the voice-controlled light switch may comprise one or more additional microphones on the faceplate aligned horizontally with the second microphone and one or more additional microphones on the faceplate aligned horizontally with the third microphone.
- the light switch may include a fourth microphone to the left of the second microphone and a fifth microphone to the right of the second microphone.
- the operating logic of the light switch may determine whether to select one or more of the following pairs of audio signals: (1) the first audio signal and a fourth audio signal generated by the fourth microphone; (2) the first audio signal and the second audio signal, and/or (3) a fifth audio signal generated by the fifth microphone.
- the operating logic may determine a location of an audio source in the environment (e.g., the speaking user) to determine which one or more pairs of audio signals to select for processing. For instance, the operating logic may attempt to determine which pair of these pairs of microphones is most in-line with the speaking user. In other instances, the operating logic may select a blend of these pairs.
- the operating logic may define a first line based on the position of the fourth microphone and the position of the first microphone, a second line based on position of the second microphone and the position of the first microphone, and a third line based on the position of the fifth microphone and the first microphone.
- the operating logic may then determine which line points closet to the estimated direction of the user.
- the operating logic utilizes time-difference-of-arrival (TDOA) techniques to estimate the location of the user within the environment.
- TDOA time-difference-of-arrival
- the techniques described herein may encompass an apparatus that does not couple a power source to a power load. Instead, the apparatus may simply comprise a physical switch (e.g., a rotatable knob) that includes multiple microphones that may be selected for processing based on a position of the switch.
- a physical switch e.g., a rotatable knob
- voice-controlled light switch While the examples that follow describe several example configurations of a voice-controlled light switch, other implementations may utilize any other number of microphones located at an array of positions.
- these voice-controlled light switches may include an array of other components, such as output devices (e.g., speakers, displays, etc.), different types of network interfaces (e.g., wired network interfaces, BLE interfaces, Zigbee interfaces, WiFi interfaces, etc.), and the like.
- FIG. 1 shows an illustrative voice interaction computing architecture 100 set in a home environment 102 that includes a user 104 , a voice-controlled light switch 108 , and a user device 110 .
- the user issues a voice command 106 (“Wake up . . . please turn on the lights.”)
- the voice-controlled light switch 108 may include one or more microphones 112 that may generate one or more audio signals for identifying the voice command and performing a corresponding action, such as turning on a light 114 within the environment 102 .
- the voice-controlled light switch 108 may identity the predefined word or phrase (“wake up”) and may thereafter begin streaming one or more audio signals to another device for processing, such as to a remote service 116 comprising one or more remote devices (or “computing resources”).
- a remote service 116 comprising one or more remote devices (or “computing resources”).
- These remote computing resources may form a portion of a network-accessible computing platform implemented as a computing infrastructure of processors, storage, software, data access, and so forth that is maintained and accessible via a network such as the Internet.
- the remote computing resources do not require end-user knowledge of the physical location and configuration of the system that delivers the services.
- remote computing devices include “on-demand computing”, “software as a service (SaaS)”, “platform computing”, “network-accessible platform”, “cloud services”, “data centers”, and so forth.
- SaaS software as a service
- FIG. 1 illustrates the remote service as being accessible over a network, in other instances the remote service/device may comprise a local hub within an environment of the voice-controlled light switch 108 .
- the remote service 116 may perform processing on these audio signals (e.g., beamforming, acoustic echo cancelation, etc.) as well as speech recognition to identify one or more voice commands, such as the command 106 to turn on the light. After identifying the voice command, the remote service 116 may send an instruction to the voice-controlled light switch 108 or to another device (in the environment 102 or otherwise) to perform a corresponding action. In this example, the remote service 116 may send an instruction to the light switch 108 to turn on the light 114 . In response to receiving this instruction, the voice-controlled light switch 108 may execute the instruction and turn on the light 114 .
- these audio signals e.g., beamforming, acoustic echo cancelation, etc.
- speech recognition e.g., voice commands, etc.
- the remote service 116 may send an instruction to the voice-controlled light switch 108 or to another device (in the environment 102 or otherwise) to perform a corresponding action.
- the remote service 116 may send an instruction to the
- a switch internal to the light switch may cause circuitry of the light switch to provide power to the lights connected to the switch, without the toggle switch moving positions (as in the case of a user manually flipping the toggle switch).
- the internal switch or other functionality may couple or decouple one or wires together with one another to cause power to flow to the connected lights.
- the internal switch may block or throttle current to effectively decouple these wires such that current is not provided to the connected lights or other power load.
- the voice-controlled light switch 108 may communicate with the remote service 116 over one or more networks 118 , which may comprise wired technologies (e.g., wires, USB, fiber optic cable, etc.), wireless technologies (e.g., WiFi, RF, cellular, satellite, Bluetooth, etc.), or other connection technologies.
- the network 118 is representative of any type of communication network, including data and/or voice network, and may be implemented using wired infrastructure (e.g., cable, CATS, fiber optic cable, etc.), a wireless infrastructure (e.g., RF, cellular, microwave, satellite, Bluetooth, etc.), and/or other connection technologies.
- the voice-controlled light switch 108 may also communicate with user devices via short-range wireless communication protocols (e.g., Bluetooth, Zigbee, etc.), such as the user device 110 .
- the voice-controlled light switch 108 may provide audio signals to the user device 110 or other user devices, which may in turn send the audio signals to the remote service 116 or may itself perform the processing (e.g., beamforming) and/or speech recognition.
- the voice-controlled light switch 108 may communicate with local user devices to send instructions to these devices. For instance, if the user 104 states a command to “turn on my kitchen television”, the voice-controlled light switch 108 may send an instruction to do so to the kitchen television using the appropriate protocol used by the television.
- the voice-controlled light switch 108 may include the microphones 112 , circuitry 120 for coupling the light switch to a power source and to the light 114 , a physical switch 122 (e.g., a toggle switch, a rocker switch, a slider, a knob, etc.), one or more network interfaces 124 , a processor 126 , one or more speakers 128 , and memory 130 .
- the memory 130 may store a speech-recognition engine 132 and a microphone-selection module 134 .
- the speech-recognition engine 132 may function to perform speech recognition on audio signals generated by the microphones 112 to identify a predefined word or phrase, such as “wake up”.
- the light switch 108 may begin sending one or more of the audio signals to the remote service 116 using one of the network interfaces 124 .
- the microphone-selection module 134 may determine which microphones and corresponding audio signals to select based on the current position of the switch and/or the current position of an audio source, such as a user speaking with the environment.
- the voice-controlled light switch 108 may include a sensor to determine the position of the physical switch 122 .
- the microphone-selection module 134 may determine to send, to the remote service 116 , an audio signal generated by a microphone on the switch and an audio signal generated by a microphone on a top portion of the faceplate. Further, if the faceplate includes multiple microphone the microphone-selection module 134 may select an audio signal corresponding to a microphone most in-line with the user 104 and the microphone on switch, relative to the other microphones on the top portion of the faceplate. To do so, the microphone-selection module 134 may first determine the position of the user 104 in the environment using TDOA techniques, using an image sensor to identify the user within the environment, or in any other manner.
- the microphone-selection module 134 may select an audio signal corresponding to a microphone on the bottom portion of the faceplate. Again, when the bottom portion includes multiple microphones, the module 134 may select the microphone most in-line with the position of the user 104 and the microphone on the switch.
- the remote service 116 includes a processor 136 and memory 138 , which have access to a signal-processing engine 140 , a speech-recognition engine 142 , and a command-processing engine 144 .
- the voice-controlled light switch 108 or other client devices within the environment 102 may upload audio signals to the remote service 116 for processing, given that the remote service 116 may have a computational capacity that far exceeds the computational capacity of the voice-controlled light switch 108 . Therefore, upon receiving multiple audio signals from the voice-controlled light switch 108 , as selected by the microphone-selection module 134 , the signal-processing engine 140 may perform one or more relatively complex processing techniques on the signals to generate one or more processed audio signals.
- the signal-processing engine 140 may perform beamforming, acoustic-echo cancelation, background noise reduction, or other techniques to generate one or more processed audio signals having a higher SNR than the pre-processed audio signals.
- the voice-controlled light switch 108 may additionally or alternatively include the signal-processing engine 140 for performing the beamforming, acoustic-echo cancelation, background noise reduction, or other techniques prior to uploading the processed audio signals to the remote service 116 .
- the speech-recognition engine 142 may perform speech-recognition on the processed audio signal(s) to identify one or more voice commands represented therein, such as the example voice command 106 .
- the command-processing engine 132 may cause performance of one or more action in response to identifying the voice command. In the illustrated example, for instance, the command-processing engine 132 may issue an instruction to the voice-controlled light switch 108 to turn on the light 114 .
- the voice-controlled light switch 108 transitions from an “asleep” state, in which the local speech-recognition engine 132 performs ASR to identify a predefined utterance, to an “awake” state, in which the voice-controlled light switch 108 provides audio signals to the remote service 116 for processing and performing ASR thereon.
- the voice-controlled light switch 108 may include a selectable physical button or soft button that causes the voice-controlled light switch 108 to transition from the “asleep” state to the “awake” state in response to a user selecting the button.
- the voice-controlled light switch 108 may receive vocal input from the user 104 and the voice-controlled light switch 108 and/or the remote service 116 may perform speech recognition to interpret a user's operational request or command.
- the requests may be for essentially any type of operation, such as turning on or off appliances, adjusting appliance settings, authentication, database inquires, requesting and consuming entertainment (e.g., gaming, finding and playing music, movies or other content, etc.), personal information management (e.g., calendaring, note taking, etc.), online shopping, financial transactions, and so forth.
- FIG. 2 illustrates an example configuration of the voice-controlled light switch 108 .
- the voice-controlled light switch 108 includes a first microphone 112 ( 1 ) on a distal end of the switch 122 (or “toggle”) of the light switch 108 . Because the microphone 112 ( 1 ) resides on the end of the switch 122 , the first microphone 112 ( 1 ) may reside nearer the top of a faceplate 202 of the light switch 108 (when the switch 122 is in the up position) or nearer the bottom of the faceplate 202 (when the switch 122 is in the down position).
- the voice-controlled light switch 108 may include a second microphone 112 ( 2 ) located nearer the top edge of the faceplate 202 , as well as a third microphone 112 ( 3 ) located nearer the bottom edge of the faceplate 202 .
- the microphone 112 ( 1 ) may be nearer the microphone 112 ( 2 ) when the switch 122 is in the up position and may be nearer the microphone 112 ( 3 ) when the switch 122 is in the down position. That is, a distance between the microphone 112 ( 1 ) and the microphone 112 ( 2 ) may be shorter than a distance between microphone 112 ( 1 ) and the microphone 112 ( 3 ) when the switch 122 is in up position.
- a distance between the microphone 112 ( 1 ) and the microphone 112 ( 3 ) may be shorter than a distance between microphone 112 ( 1 ) and the microphone 112 ( 2 ) when the switch 122 is in down position.
- the microphone 112 ( 1 ) and the microphone 112 ( 2 ) may be in substantially the same horizontal plane (i.e., may have the same substantial vertical height from the floor) when the switch 122 is in the up position.
- the microphone 112 ( 1 ) and the microphone 112 ( 3 ) may be in substantially the same horizontal plane (i.e., may have the same substantial vertical height from the floor) when the switch 122 is in the down position.
- the voice-controlled light switch 108 may include operating logic to determine when the switch is in the up position and when the switch is in the down position.
- the microphone-selection engine may determine whether the switch 122 is in the up position or in the down position, and may select one or more audio signals to send to the remote service 116 based on this determination. For instance, in response to determining that the switch is the up position, the microphone-selection module 134 may send, to the remote service, audio signals generated by the first microphone 112 ( 1 ) and by the second microphone 112 ( 2 ).
- the microphone-selection module 134 may select, for processing, the audio signals generated by the first microphone 112 ( 1 ) and by the second microphone 112 ( 2 ). For instance, the light switch 108 may perform beamforming on these audio signals to generate one or more processed audio signals and may then send these one or more processed audio signals to the remote service 116 . In some instances, when the switch is in the up position, the light switch 108 may use an audio signal generated by the third microphone 112 ( 3 ) for identifying and background noise and, thereafter, removing at least some of the background noise from the audio signals generated by the first send second microphones.
- the microphone-selection module 134 may select, for processing, the audio signals generated by the first microphone 112 ( 1 ) and by the third microphone 112 ( 3 ). For instance, the light switch 108 may perform beamforming on these audio signals to generate one or more processed audio signals and may then send these one or more processed audio signals to the remote service 116 . In some instances, when the switch is in the down position, the light switch 108 may use an audio signal generated by the second microphone 112 ( 2 ) for identifying and background noise and, thereafter, removing at least some of the background noise from the audio signals generated by the first send third microphones.
- these microphones may be substantially vertically in-line with one another, these signals make good candidates for performing beamforming thereon to generate a processed audio signal having a relatively high SNR as compared to the signals by themselves. Further, in instances where the microphone-selection module 134 determines that the switch is the down position, the microphone-selection module 134 may send, to the remote service, audio signals generated by the first microphone 112 ( 1 ) and by the third microphone 112 ( 3 ).
- the voice-controlled light switch 108 may include a fourth microphone 112 ( 4 ) that is substantially horizontally in-line with, but to the left of, the second microphone 112 ( 2 ), as well as a fifth microphone 112 ( 5 ) that is substantially horizontally in-line with, but to the right of, the second microphone 112 ( 2 ).
- the voice-controlled light switch 108 may include a sixth microphone 112 ( 6 ) that is substantially horizontally in-line with, but to the left of, the third microphone 112 ( 3 ), as well as a seventh microphone 112 ( 7 ) that is substantially horizontally in-line with, but to the right of, the third microphone 112 ( 3 ).
- the microphone-selection module 134 may also take into account a location of the audio source, such as the speaking user, within the environment. To do so, the microphone-selection module may utilize TDOA techniques, images captured by a camera, or the like. For instance, given that each of the microphones 112 ( 1 )-( 7 ) may detect a command sound at different times, the microphone-selection module 134 may compare these different times to estimate the direction of the user relative to the voice-controlled light switch 108 .
- the microphone-selection module 134 may determine which microphone on the faceplate 202 (or which combination of microphones) to send to the remote service 116 .
- the microphone-selection engine defines a respective line between each microphone on the appropriate half of the faceplate 202 (e.g., each of the microphones on the top if the switch 122 is in the up position) and the first microphone 112 ( 1 ) on the switch 122 .
- the microphone-selection module 134 may then determine which line or lines most closely corresponds with the direction of the user.
- the microphone-selection module 134 selects this microphone pair and sends, to the remote service 116 , the audio signals generated by this pair of microphones.
- the microphone-selection module 134 may select the second microphone 112 ( 2 ) along with the microphone 112 ( 1 ). If, however, the user were to the right side of the voice-controlled light switch 108 when looking at the light switch 108 , the voice-controlled module 134 may select the fourth microphone 112 ( 4 ).
- the microphone-selection module 134 may select the fifth microphone 112 ( 5 ) along with the first microphone 112 ( 1 ). In other instances, meanwhile, the microphone-selection engine may send more than two audio signals to the remote service 116 for increased accuracy. For instance, if the user were standing center-left relative to the voice-controlled light switch 108 , the microphone-selection module 134 may send audio signals corresponding to the first microphone 112 ( 1 ), the second microphone 112 ( 2 ) and the fifth microphone 112 ( 5 ). Further, it is noted that while FIG. 2 and subsequent figures illustrate example arrays of microphones, it is to be appreciated that other implementations may utilize more microphones, fewer microphones, and/or different configurations of microphones.
- FIG. 3 illustrates another example configuration of the voice-controlled light switch 108 .
- the light switch 108 is in the form of a rocker switch, and includes two microphones on the switch 122 (or “rocker”) and microphones on the top and bottom of the faceplate. That is, the switch 122 includes the first microphones 112 ( 1 ) on the top portion of the rocker switch 122 , as well as an eighth microphone 112 ( 8 ) on the bottom portion of the switch 122 .
- the top portion of the faceplate 202 of the voice-controlled light switch 108 again includes the three microphones 112 ( 4 ), 112 ( 2 ), and 112 ( 5 ), while the bottom portion of the faceplate 202 includes the three microphones 112 ( 6 ), 112 ( 3 ), and 112 ( 7 ).
- the microphone-selection module 134 may determine whether the switch 122 is in the up or down position and may select, for processing, audio signals generated by microphones corresponding to the up or down position. In this case, for instance, if the switch 122 is in the up position, the microphone-selection module 134 may select, for processing, audio signals generated by the eight microphone 112 ( 8 ), as well as one or more of the three microphones 112 ( 6 ), 112 ( 3 ), and 112 ( 7 ).
- the microphone-selection engine may select the bottom microphones when the switch 122 is in the up position.
- the microphone-selection module 134 may select the microphones corresponding to the top of the light switch 108 , namely the first microphone 112 ( 1 ) along with one or more of the three microphones 112 ( 4 ), 112 ( 2 ), and 112 ( 5 ).
- FIG. 4 illustrates yet another example configuration of the voice-controlled light switch 108 of FIG. 1 .
- the light switch is in the form of a sliding dimmer light switch that includes a sliding switch 122 , which itself includes a first microphone 112 ( 1 ) on the left side of the switch 122 and a second microphone 112 ( 2 ) on the right side of the switch 122 .
- the example light switch 108 includes three microphones on the left side of the faceplate plate 202 as well as three microphones on the right side of the faceplate 202 .
- the microphone-selection module 134 may select one or more pairs of microphones for use based on the location of the sliding switch 122 , potentially along with the location of the user 104 .
- the microphone-selection signal may select, for processing, audio signals corresponding to the microphone 112 ( 5 ) and 112 ( 1 ), or audio signals corresponding to the microphones 112 ( 7 ) and 112 ( 2 ).
- the microphone selection module 134 may select, for processing, audio signals corresponding to the microphone 112 ( 3 ) and 112 ( 1 ), or audio signals corresponding to the microphones 112 ( 4 ) and 112 ( 2 ).
- the microphone selection module 134 may select, for processing, audio signals corresponding to the microphone 112 ( 6 ) and 112 ( 1 ), or audio signals corresponding to the microphones 112 ( 8 ) and 112 ( 2 ).
- FIG. 5 illustrates yet another example configuration of the voice-controlled light switch 108 of FIG. 1 .
- the light switch 108 is in the form of a rotatable-knob dimmer switch, which comprises a rotatable switch 122 that a user is able to rotate radially to adjust an amount of power applied to a power load, similar to the dimmer light switch discussed above with reference to FIG. 4 .
- the rotatable switch 122 may comprise the first microphone 112 ( 1 ).
- the faceplate 202 of the light switch 108 may include one or more microphones, potentially equally distributed radially about the rotatable switch 122 .
- the light switch 108 includes microphones 112 ( 2 ), 112 ( 3 ), . . . , 112 ( 9 ).
- the microphone-selection module 134 may select audio signals corresponding to one or more pairs of microphones. For instance, the microphone-selection module 134 may determine which of the microphones 112 ( 2 )-( 9 ) is closest to the current location of the microphone 112 ( 1 ) and may send, to the remote service 116 , the audio signal corresponding to this microphone and to the first microphone 112 ( 1 ).
- the microphone-selection module 134 may select the audio signal corresponding to the microphone 112 ( 4 ) and may therefore send, to the remote service 116 , the audio signals corresponding to the first microphone 112 ( 1 ) and the fourth microphone 112 ( 4 ).
- FIG. 5 illustrates the first microphone 112 ( 1 ) along with the additional microphones on the faceplate
- the rotatable knob 122 may include the first microphone on a face of the knob and a second microphone located nearer the face plate and vertically in-line with the first microphone. As such, no matter the position of the knob, the audio signals generated by these microphones may be used to increase the SNR of the signals using beamforming or other processing techniques.
- FIG. describes a light switch that couples a power source to a power load
- an apparatus may have the same form factor as the apparatus of FIG. 5 , while not coupling to a power source.
- This apparatus may instead simply comprise a rotatable know or other type switch that includes the first microphone 112 ( 1 ) and one or more additional microphones, on a faceplate and/or on a second position on the knob closer to the faceplate. While this apparatus might not couple to a power source, this apparatus may include the same functionality for determining a position of the knob and selecting one or more audio signals for processing based on this position.
- FIG. 6 illustrates a flow diagram of an example process 600 for selecting which audio signals to process based on a position of a switch in a voice-controlled light switch.
- Processing the signals may include sending these signals to a remote service for further processing, such as performing beamforming on the signals to generate a processed audio signal and performing speech recognition on the processed audio signal. Additionally or alternatively, processing these signals may include performing beamforming, noise-cancelation, or the like locally at the switch or at another device in the environment, and/or performing speech recognition locally at the switch or at another device in the environment.
- the process 600 (and each process described herein) is illustrated as a logical flow graph, each operation of which represents a sequence of operations that can be implemented in hardware, software, or a combination thereof. In some instances, some or all of the operations are executed by the components of a voice-controlled light switch, such as the light switch 108 described above.
- the operations represent computer-executable instructions stored on one or more computer-readable media that, when executed by one or more processors, perform the recited operations.
- computer-executable instructions include routines, programs, objects, components, data structures, and the like that perform particular functions or implement particular abstract data types.
- the computer-readable media may include non-transitory computer-readable storage media, which may include hard drives, floppy diskettes, optical disks, CD-ROMs, DVDs, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, flash memory, magnetic or optical cards, solid-state memory devices, or other types of storage media suitable for storing electronic instructions.
- the computer-readable media may include a transitory computer-readable signal (in compressed or uncompressed form). Examples of computer-readable signals, whether modulated using a carrier or not, include, but are not limited to, signals that a computer system hosting or running a computer program can be configured to access, including signals downloaded through the Internet or other networks.
- the process 600 generates one or more audio signals. For instance, if a voice-controlled light switch comprises multiple microphones, one or more of these microphones may generate a corresponding audio signal based on sound detected within an environment of the light switch.
- the process 600 performs speech recognition on one or more of these generated audio signals. For instance, a speech-recognition module 134 operating on the light switch may perform speech recognition on one or more of the signals to identify a predefined word or phrase.
- the process 600 queries whether it has identified the predefined word or phrase based on the performing of the speech recognition. If not, then the process 600 returns to the operation 604 to continue monitoring the audio signals for the predefined word or phrase.
- the process 600 determines, at 608 , whether the switch is in a first position or a second position. While the example process 600 describes determine whether the switch is in one of two positions, in other implementations any other number of positions may be utilized. In any event, if it is determined that the switch is in the first position, then at 610 the process 610 ( 1 ) selects, for processing, audio signals corresponding to the first position for processing. As described above, these signals may also be selected based on other criteria, such as a location of the user within the environment of the light switch.
- processing these audio signals may include, but is not limited to, sending the audio signals to a remote service for further processing, such as performing beamforming and speech recognition thereon (operation 610 ( 1 )), performing speech recognition locally at the light switch (operation 610 ( 2 )), and/or performing beamforming, noise cancellation, acoustic echo cancelation or the like on the audio signals locally at the light switch (operation 610 ( 3 )).
- a remote service such as performing beamforming and speech recognition thereon (operation 610 ( 1 )), performing speech recognition locally at the light switch (operation 610 ( 2 )), and/or performing beamforming, noise cancellation, acoustic echo cancelation or the like on the audio signals locally at the light switch (operation 610 ( 3 )).
- processing of these signals may include sending the signals to a remote service (operation 612 ( 1 )), performing speech recognition on the signals locally (operation 612 ( 2 )), and/or performing beamforming or the like on the signals locally (operation 612 ( 3 )).
- FIG. 7 illustrates a flow diagram of an example process 700 for selecting audio signals for processing based on a position of a switch, as well as based on a position of an audio source (e.g., a speaking user) within the environment.
- the process 700 determines that the switch is in the n th position (representative of any available position).
- the process 700 determines a first line defined by a microphone on a switch of a light switch and a first microphone on a faceplate that corresponds to the n th position.
- the process 700 determines a second line defined by the microphone on the switch and a second microphone on a faceplate that corresponds to the n th position.
- the process 700 determines a third line defined by the microphone on the switch and a third microphone on a faceplate that corresponds to the n th position.
- the process 700 estimates a direction of an audio source (e.g., a speaking user) within the environment using TDOA techniques, a camera or the like.
- the process 700 queries whether the first, second, or third line is closest to the estimated direction of the audio source. If the process 700 determines that the first line is closest, then at 714 ( 1 ) the process 700 selects, for processing, audio signals corresponding to the microphone on the switch and the first microphone on the faceplate.
- the process 700 determines that the second line is closest, then at 714 ( 2 ) the process 700 selects, for processing, audio signals corresponding to the microphone on the switch and the second microphone on the faceplate. Finally, if the process 700 determines that the third line is closest, then at 714 ( 3 ) the process 700 selects, for processing, audio signals corresponding to the microphone on the switch and the third microphone on the faceplate. Further, while three example faceplate microphones are described, it is to be appreciated that other implementations may utilize any other number of microphones on the faceplate.
- FIG. 8 shows a block diagram of selected functional components implemented in the voice-controlled light switch 108 of FIG. 1 . While this figure illustrates one example light switch 108 , it is to be appreciated that any other type of client or server computing device may implement the techniques described herein. Furthermore, it is to be appreciated that the light switch 108 may include some or all of the components described above with reference to the remote service 116 .
- the voice-controlled light switch 108 includes the processor 126 and memory 130 .
- the memory 130 may include computer-readable storage media (“CRSM”), which may be any available physical media accessible by the processor 126 to execute instructions stored on the memory.
- CRSM may include random access memory (“RAM”) and Flash memory.
- RAM random access memory
- CRSM may include, but is not limited to, read-only memory (“ROM”), electrically erasable programmable read-only memory (“EEPROM”), or any other medium which can be used to store the desired information and which can be accessed by the processor 126 .
- the voice-controlled light switch 108 includes the microphone(s) 112 that to receive audio input, such as user voice input.
- the light switch 108 may also include the speaker(s) 128 to output audio sounds.
- One or more codecs 802 are coupled to the microphones and the speakers to encode and/or decode the audio signals.
- the codec may convert audio data between analog and digital formats.
- a user may interact with the light switch 108 by speaking to it and the microphones 112 capture sound and generate audio signals that includes the user speech.
- the codec 802 encodes the user speech and transfers that audio data to other components.
- the voice-controlled light switch 108 includes the one or more network interfaces 124 , which may couple to one or more antennas 804 to facilitate a wireless connection to a network.
- the network interfaces 124 may implement one or more of various wireless technologies, such as WiFi, Bluetooth, RF, and so on.
- One or more device interfaces 806 may further be provided as part of the light switch 108 to facilitate a wired connection to a network, or a plug-in network device that communicates with other wireless networks.
- One or more power units 808 may be further provided to distribute power to the various components on the light switch 108 .
- the light switch couples to a persistent power source, such as a primary AC power supply.
- modules such as instruction, datastores, and so forth may be stored within the memory 130 and configured to execute on the processor 126 .
- An operating system module 810 is configured to manage hardware and services (e.g., wireless unit, Codec, etc.) within and coupled to the light switch 108 for the benefit of other modules.
- the memory 130 may include the speech-recognition engine 132 , the microphone-selection module 134 , and/or the signal-processing engine 140 described above, amongst other modules. Some or all of these modules, data stores, and components may reside additionally or alternatively at the remote service 116 .
- the light switch 108 may include the circuitry 120 for coupling the light switch to a power source and to a power load, such as a light, electrical outlet, or the like.
- the circuitry may include one or more wires for making this connection to the persistent power source.
- the light switch may include the physical switch 122 that is selectable to provide power to the power load and to cease providing power to the power load.
- the switch 122 may be adjustable to provide varying levels of power to the connected power load(s), while in other instances the switch 122 may include two positions, on and off.
- the circuitry 120 may also include an internal switch that causes the one or more wires to couple together to cause the power source to provide power to the coupled power load. In some instances, this internal switch of the circuitry 120 may couple the wires in this manner in response to a user flipping the physical switch 122 or in response to execution of an instruction received based on a user issuing a voice command.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Acoustics & Sound (AREA)
- Human Computer Interaction (AREA)
- Signal Processing (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Automation & Control Theory (AREA)
- Computer Networks & Wireless Communication (AREA)
- Computational Linguistics (AREA)
- Otolaryngology (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Circuit For Audible Band Transducer (AREA)
- Selective Calling Equipment (AREA)
Abstract
This disclosure describes, in part, voice-controlled light switches that act as voice-controlled endpoints at which users may provide voice commands. These light switches may physically couple to a power source, as well as to one or more appliances. The light switches may include physical switches to selectively provide power to the appliances, either in a binary fashion (on or off) or along of spectrum of positions. In either instance, the appliances coupled to the light switches may include lights, electrical outlets, home appliances (e.g., dishwashers, washing machines, etc.), factory machinery, or any other device that receives electrical power.
Description
- Homes are becoming more wired and connected with the proliferation of computing devices such as desktops, tablets, entertainment systems, and portable communication devices. As computing devices evolve, many different ways have been introduced to allow users to interact with these devices, such as through mechanical means (e.g., keyboards, mice, etc.), touch screens, motion, and gesture. Another way to interact with computing devices is through speech.
- The detailed description is described with reference to the accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The use of the same reference numbers in different figures indicates similar or identical components or features.
-
FIG. 1 shows an illustrative voice interaction computing architecture set in a home environment. The architecture includes at least one voice-controlled light switch and a voice-controlled device physically situated in the home, along with a user. In this example, the light switch includes one or more microphones for generating an audio signal that may include a command of the user, such as a command to turn on the lights in the illustrated room. -
FIG. 2 illustrates an example voice-controlled light switch that includes a microphone on a distal end of a switch (or “toggle”) of the light switch, and microphones on the faceplate of the light switch. In some instances, the light switch may include functionality to determine whether the switch in a first position (e.g., the up position) or a second position (e.g., the down position) and may select audio signals to process based on this position. For instance, the if the switch is in the first position, the light switch may select, for processing, the audio signals generated by the microphone on the switch and one or more of the microphones on the top section of the faceplate, which may allow for more accurate speech processing as described below. -
FIG. 3 illustrates another example configuration of a voice-controlled light switch. Here, the light switch is in the form of a rocker switch, and includes two microphones on the switch (or “rocker”) and microphones on the top and bottom of the faceplate. -
FIG. 4 illustrates yet another example configuration of a voice-controlled light switch. Here, the light switch is in the form of a sliding dimmer switch, and includes two microphones on the switch (or “sliding knob”) that is configured to move to different positions of the dimmer switch, as well as microphones on the left and ride sides of the faceplate. -
FIG. 5 illustrates yet another example configuration of a voice-controlled light switch. Here, the light switch is in the form of a rotatable-knob dimmer switch, and includes a microphone on the switch (or “rotatable knob”) that is configured to move radially to different positions, as well as microphones distributed across the faceplate of the dimmer switch. -
FIG. 6 illustrates a flow diagram of an example process for selecting which audio signals to process based on a position of a switch in a voice-controlled light switch. Processing the signals may include sending these signals to a remote service for further processing, such as performing beamforming on the signals to generate a processed audio signal and performing speech recognition on the processed audio signal. Additionally or alternatively, processing these signals may include performing beamforming, noise-cancelation, or the like locally at the switch or at another device in the environment, and/or performing speech recognition locally at the switch or at another device in the environment. -
FIG. 7 illustrates a flow diagram of an example process for selecting audio signals for processing based on a position of a switch, as well as based on a position of an audio source (e.g., a speaking user) within the environment. -
FIG. 8 shows a block diagram of selected functional components implemented in the voice-controlled light switch ofFIG. 1 . - This disclosure describes, in part, voice-controlled light switches that act as voice-controlled endpoints at which users may provide voice commands. These light switches may physically couple to a power source, as well as to one or more power loads, such as appliances (e.g., lights, outlets, home appliances, etc.). The light switches may include physical switches to selectively provide power to the power loads, either in a binary fashion (on or off) or along of spectrum of positions. In either instance, the power loads coupled to the light switches may include lights, electrical outlets, home appliances (e.g., dishwashers, washing machines, smoke detectors, security alarms, automated door locks, automated window shades, thermostats, etc.), factory machinery, healthcare devices (e.g., insulin pumps, CPAP machines, etc.), cleaning devices (e.g., autonomous robotic vacuum cleaners, etc.), or any other device that receives electrical power.
- In addition to these traditional light-switch components, the voice-controlled light switches described herein include one or more microphones for generating audio signals, from which one or more voice commands may be identified. For instance, a light switch may include a microphone on the physical switch, on the faceplate, and/or the like, with this microphone generating audio signals based on sound detected within the environment of the light switch. The light switch may perform speech recognition on the generated audio signal to identify a voice command, or the light switch may include a network interface for sending the audio signal to another device or to a remote service, which may perform the speech processing. The power source of the light switch may comprise a persistent power source, such as the primary alternating current (AC) power supply connected to a home or other building. As such, this power source may be used to power the microphones and other components of the light switch, in addition to the power load to which the light switch couples. Further, while the light switch may couple to the AC power supply, in other instances the light switch may couple to any other type of power source, such as batteries, capacitors, solar panels, or the like.
- In some instances, the voice-controlled light switch includes a speech-recognition engine configured to identify a predefined utterance (e.g., a predefined sound or a predefined word or phrase) from a generated audio signal and, upon identifying this utterance, may begin providing generated audio signals to a remote service for performing speech recognition. Therefore, a user may determine when the audio signals are sent and when voice commands will be identified by speaking the predefined utterance (e.g., “wake up”). The voice commands may be for an array of activities, such as turning on the light in the user's environment, a request for information (e.g., “what is the temperature outside?”), a request to perform a task (“remind me to pick up Johnny at 4 pm”), or any other type of request.
- In some instances, the voice-controlled light switches include multiple microphones, each configured to generate a respective audio signal. Together, these multiple audio signals may be processed to generate a processed audio signal, having a signal-to-noise (SNR) ratio that is greater than the SNR ratio of any singular one of the audio signals. For instance, the light switch, or a remote service, may utilize the multiple audio signals to generate a processed audio signal using beamforming techniques, noise-cancellation techniques, or the like. The voice-controlled light switch, the remote service, or another device may then perform speech recognition on the processed audio signal having the increased SNR, thus increasing accuracy and efficacy of the speech recognition.
- In some examples, the voice-controlled light switch includes at least a first microphone on a switch of the light switch. That is, the first microphone may reside on the physical control used to cause the power load coupled to the light switch to couple to the power source and, hence, receive power, and may be used to electrically decouple the power load from the power source such that the power load does not receive power from the power source. The voice-controlled light switch may comprise a toggle light switch, a rocker light switch, a sliding-dimmer light switch, a rotatable-knob light switch, or any other type of switch. In addition, to the first microphone residing on the switch, the light switch may include one or more microphones located on the faceplate of the light switch or any other location on or adjacent to the light switch. For instance, when the light switch comprises a toggle light switch configured to reside in an up position and a down position, the light switch may include a second microphone on a top portion of the faceplate, as well as a third microphone located on a bottom portion of the faceplate. In some instances, the second and third microphones are substantially in-line with the first microphone in a horizontal direction. For example, when the switch—and the first microphone residing thereon—resides substantially in horizontal middle of the overall structure of the light switch, the second and third microphones may also reside in the horizontal middle.
- In the above example, the light switch may include operating logic configured to determine when the switch resides in the up position and when the switch resides in the down position. In response to determining that the switch resides in the up position, the light switch may select for processing a first audio signal generated by the first microphone and a second audio signal generated by the second microphone (on the top portion of the faceplate). This processing may include performing beamforming or other processing locally on the signals or sending the signals to the remote service for performing beamforming or other processing techniques thereon. Given the locations of the first and second microphones, when the switch is in the up position these microphones may be substantially in-line with one another in the vertical direction. Further, given the common height of light switches on walls and average height ranges of users, these microphones may also be substantially vertically in-line with a mouth of a speaking user. Using audio signals generated from microphones in this configuration is particularly helpful when performing beamforming on these audio signals to generate a processed audio signal having a relatively high SNR. As such, selecting the first and second audio signals for processing when the switch resides in the up position may result in particularly accurate speech recognition. Stated otherwise, because the first and second microphones are in the same substantial horizontal plane (i.e., are substantially the same vertical height off of the floor), when the user speaks in the room towards the wall/light switch, the audio of the user may reach the first microphone and then the second microphone which is substantially vertically in-line with the first microphone. This configuration may result in effective beamforming or other signal processing.
- Conversely, when the switch resides in the down position, the operating logic may select for processing the first audio signal (generated by the first microphone on the switch) and a third audio signal generated by the third microphone on the bottom portion of the faceplate. Again, given that these microphones are substantially in-line with one another in the vertical direction when the switch is in the down position (as well as substantially vertically in-line with a mouth of a speaking user), these audio signals represent excellent candidates for performing beamforming thereon to generate a processed audio signal having a relatively high SNR.
- In addition, the voice-controlled light switch may comprise one or more additional microphones on the faceplate aligned horizontally with the second microphone and one or more additional microphones on the faceplate aligned horizontally with the third microphone. For instance, the light switch may include a fourth microphone to the left of the second microphone and a fifth microphone to the right of the second microphone. In these instances, when the switch is in the up position the operating logic of the light switch may determine whether to select one or more of the following pairs of audio signals: (1) the first audio signal and a fourth audio signal generated by the fourth microphone; (2) the first audio signal and the second audio signal, and/or (3) a fifth audio signal generated by the fifth microphone. In some instances, the operating logic may determine a location of an audio source in the environment (e.g., the speaking user) to determine which one or more pairs of audio signals to select for processing. For instance, the operating logic may attempt to determine which pair of these pairs of microphones is most in-line with the speaking user. In other instances, the operating logic may select a blend of these pairs.
- For instance, the operating logic may define a first line based on the position of the fourth microphone and the position of the first microphone, a second line based on position of the second microphone and the position of the first microphone, and a third line based on the position of the fifth microphone and the first microphone. The operating logic may then determine which line points closet to the estimated direction of the user. In some instances, the operating logic utilizes time-difference-of-arrival (TDOA) techniques to estimate the location of the user within the environment.
- While the above example describes a light switch that includes multiples microphones and that couples a power source to a power load, in other instances the techniques described herein may encompass an apparatus that does not couple a power source to a power load. Instead, the apparatus may simply comprise a physical switch (e.g., a rotatable knob) that includes multiple microphones that may be selected for processing based on a position of the switch.
- While the examples that follow describe several example configurations of a voice-controlled light switch, other implementations may utilize any other number of microphones located at an array of positions. In addition, these voice-controlled light switches may include an array of other components, such as output devices (e.g., speakers, displays, etc.), different types of network interfaces (e.g., wired network interfaces, BLE interfaces, Zigbee interfaces, WiFi interfaces, etc.), and the like.
-
FIG. 1 shows an illustrative voiceinteraction computing architecture 100 set in ahome environment 102 that includes auser 104, a voice-controlledlight switch 108, and auser device 110. As illustrated, the user issues a voice command 106 (“Wake up . . . please turn on the lights.”) In some instances, the voice-controlledlight switch 108 may include one ormore microphones 112 that may generate one or more audio signals for identifying the voice command and performing a corresponding action, such as turning on a light 114 within theenvironment 102. - In some instances, the voice-controlled
light switch 108 may identity the predefined word or phrase (“wake up”) and may thereafter begin streaming one or more audio signals to another device for processing, such as to aremote service 116 comprising one or more remote devices (or “computing resources”). These remote computing resources may form a portion of a network-accessible computing platform implemented as a computing infrastructure of processors, storage, software, data access, and so forth that is maintained and accessible via a network such as the Internet. The remote computing resources do not require end-user knowledge of the physical location and configuration of the system that delivers the services. Common expressions associated for these remote computing devices include “on-demand computing”, “software as a service (SaaS)”, “platform computing”, “network-accessible platform”, “cloud services”, “data centers”, and so forth. Further, whileFIG. 1 illustrates the remote service as being accessible over a network, in other instances the remote service/device may comprise a local hub within an environment of the voice-controlledlight switch 108. - The
remote service 116 may perform processing on these audio signals (e.g., beamforming, acoustic echo cancelation, etc.) as well as speech recognition to identify one or more voice commands, such as thecommand 106 to turn on the light. After identifying the voice command, theremote service 116 may send an instruction to the voice-controlledlight switch 108 or to another device (in theenvironment 102 or otherwise) to perform a corresponding action. In this example, theremote service 116 may send an instruction to thelight switch 108 to turn on the light 114. In response to receiving this instruction, the voice-controlledlight switch 108 may execute the instruction and turn on the light 114. That is, a switch internal to the light switch may cause circuitry of the light switch to provide power to the lights connected to the switch, without the toggle switch moving positions (as in the case of a user manually flipping the toggle switch). Instead, the internal switch or other functionality may couple or decouple one or wires together with one another to cause power to flow to the connected lights. Conversely, the internal switch may block or throttle current to effectively decouple these wires such that current is not provided to the connected lights or other power load. - In the example described above, the voice-controlled
light switch 108 may communicate with theremote service 116 over one ormore networks 118, which may comprise wired technologies (e.g., wires, USB, fiber optic cable, etc.), wireless technologies (e.g., WiFi, RF, cellular, satellite, Bluetooth, etc.), or other connection technologies. Thenetwork 118 is representative of any type of communication network, including data and/or voice network, and may be implemented using wired infrastructure (e.g., cable, CATS, fiber optic cable, etc.), a wireless infrastructure (e.g., RF, cellular, microwave, satellite, Bluetooth, etc.), and/or other connection technologies. In some instances, the voice-controlledlight switch 108 may also communicate with user devices via short-range wireless communication protocols (e.g., Bluetooth, Zigbee, etc.), such as theuser device 110. For instance, the voice-controlledlight switch 108 may provide audio signals to theuser device 110 or other user devices, which may in turn send the audio signals to theremote service 116 or may itself perform the processing (e.g., beamforming) and/or speech recognition. In still other instances, the voice-controlledlight switch 108 may communicate with local user devices to send instructions to these devices. For instance, if theuser 104 states a command to “turn on my kitchen television”, the voice-controlledlight switch 108 may send an instruction to do so to the kitchen television using the appropriate protocol used by the television. - As illustrated, the voice-controlled
light switch 108 may include themicrophones 112,circuitry 120 for coupling the light switch to a power source and to the light 114, a physical switch 122 (e.g., a toggle switch, a rocker switch, a slider, a knob, etc.), one ormore network interfaces 124, aprocessor 126, one ormore speakers 128, andmemory 130. As illustrated, thememory 130 may store a speech-recognition engine 132 and a microphone-selection module 134. As described above, the speech-recognition engine 132 may function to perform speech recognition on audio signals generated by themicrophones 112 to identify a predefined word or phrase, such as “wake up”. After identifying this word or phrase, thelight switch 108 may begin sending one or more of the audio signals to theremote service 116 using one of the network interfaces 124. Before doing so, however, the microphone-selection module 134 may determine which microphones and corresponding audio signals to select based on the current position of the switch and/or the current position of an audio source, such as a user speaking with the environment. In some instances, the voice-controlledlight switch 108 may include a sensor to determine the position of thephysical switch 122. - For instance, if the voice-controlled
switch 108 is currently in an up position, then the microphone-selection module 134 may determine to send, to theremote service 116, an audio signal generated by a microphone on the switch and an audio signal generated by a microphone on a top portion of the faceplate. Further, if the faceplate includes multiple microphone the microphone-selection module 134 may select an audio signal corresponding to a microphone most in-line with theuser 104 and the microphone on switch, relative to the other microphones on the top portion of the faceplate. To do so, the microphone-selection module 134 may first determine the position of theuser 104 in the environment using TDOA techniques, using an image sensor to identify the user within the environment, or in any other manner. - Further, if the switch is currently in the down position, the microphone-
selection module 134 may select an audio signal corresponding to a microphone on the bottom portion of the faceplate. Again, when the bottom portion includes multiple microphones, themodule 134 may select the microphone most in-line with the position of theuser 104 and the microphone on the switch. - As illustrated, the
remote service 116 includes aprocessor 136 andmemory 138, which have access to a signal-processing engine 140, a speech-recognition engine 142, and a command-processing engine 144. In some instances, as described below, the voice-controlledlight switch 108 or other client devices within theenvironment 102 may upload audio signals to theremote service 116 for processing, given that theremote service 116 may have a computational capacity that far exceeds the computational capacity of the voice-controlledlight switch 108. Therefore, upon receiving multiple audio signals from the voice-controlledlight switch 108, as selected by the microphone-selection module 134, the signal-processing engine 140 may perform one or more relatively complex processing techniques on the signals to generate one or more processed audio signals. For instance, the signal-processing engine 140 may perform beamforming, acoustic-echo cancelation, background noise reduction, or other techniques to generate one or more processed audio signals having a higher SNR than the pre-processed audio signals. Furthermore, as discussed above and as illustrated inFIG. 8 , the voice-controlledlight switch 108 may additionally or alternatively include the signal-processing engine 140 for performing the beamforming, acoustic-echo cancelation, background noise reduction, or other techniques prior to uploading the processed audio signals to theremote service 116. - Thereafter, the speech-
recognition engine 142 may perform speech-recognition on the processed audio signal(s) to identify one or more voice commands represented therein, such as theexample voice command 106. Thereafter, the command-processing engine 132 may cause performance of one or more action in response to identifying the voice command. In the illustrated example, for instance, the command-processing engine 132 may issue an instruction to the voice-controlledlight switch 108 to turn on the light 114. - Given the above, in some instances the voice-controlled
light switch 108 transitions from an “asleep” state, in which the local speech-recognition engine 132 performs ASR to identify a predefined utterance, to an “awake” state, in which the voice-controlledlight switch 108 provides audio signals to theremote service 116 for processing and performing ASR thereon. In other instances, meanwhile, the voice-controlledlight switch 108 may include a selectable physical button or soft button that causes the voice-controlledlight switch 108 to transition from the “asleep” state to the “awake” state in response to a user selecting the button. - Furthermore, regardless of whether the speech recognition occurs locally or remotely from the
environment 102, the voice-controlledlight switch 108 may receive vocal input from theuser 104 and the voice-controlledlight switch 108 and/or theremote service 116 may perform speech recognition to interpret a user's operational request or command. The requests may be for essentially any type of operation, such as turning on or off appliances, adjusting appliance settings, authentication, database inquires, requesting and consuming entertainment (e.g., gaming, finding and playing music, movies or other content, etc.), personal information management (e.g., calendaring, note taking, etc.), online shopping, financial transactions, and so forth. -
FIG. 2 illustrates an example configuration of the voice-controlledlight switch 108. As illustrated, the voice-controlledlight switch 108 includes a first microphone 112(1) on a distal end of the switch 122 (or “toggle”) of thelight switch 108. Because the microphone 112(1) resides on the end of theswitch 122, the first microphone 112(1) may reside nearer the top of afaceplate 202 of the light switch 108 (when theswitch 122 is in the up position) or nearer the bottom of the faceplate 202 (when theswitch 122 is in the down position). - In addition, the voice-controlled
light switch 108 may include a second microphone 112(2) located nearer the top edge of thefaceplate 202, as well as a third microphone 112(3) located nearer the bottom edge of thefaceplate 202. As such, the microphone 112(1) may be nearer the microphone 112(2) when theswitch 122 is in the up position and may be nearer the microphone 112(3) when theswitch 122 is in the down position. That is, a distance between the microphone 112(1) and the microphone 112(2) may be shorter than a distance between microphone 112(1) and the microphone 112(3) when theswitch 122 is in up position. And, a distance between the microphone 112(1) and the microphone 112(3) may be shorter than a distance between microphone 112(1) and the microphone 112(2) when theswitch 122 is in down position. Further, the microphone 112(1) and the microphone 112(2) may be in substantially the same horizontal plane (i.e., may have the same substantial vertical height from the floor) when theswitch 122 is in the up position. Further, the microphone 112(1) and the microphone 112(3) may be in substantially the same horizontal plane (i.e., may have the same substantial vertical height from the floor) when theswitch 122 is in the down position. - As described above, the voice-controlled
light switch 108 may include operating logic to determine when the switch is in the up position and when the switch is in the down position. In some instances, in response to the local speech-recognition engine 132 identifying the predefined word or phrase, the microphone-selection engine may determine whether theswitch 122 is in the up position or in the down position, and may select one or more audio signals to send to theremote service 116 based on this determination. For instance, in response to determining that the switch is the up position, the microphone-selection module 134 may send, to the remote service, audio signals generated by the first microphone 112(1) and by the second microphone 112(2). In another example, in response to determining that the switch is the up position, the microphone-selection module 134 may select, for processing, the audio signals generated by the first microphone 112(1) and by the second microphone 112(2). For instance, thelight switch 108 may perform beamforming on these audio signals to generate one or more processed audio signals and may then send these one or more processed audio signals to theremote service 116. In some instances, when the switch is in the up position, thelight switch 108 may use an audio signal generated by the third microphone 112(3) for identifying and background noise and, thereafter, removing at least some of the background noise from the audio signals generated by the first send second microphones. - Similarly, in response to determining that the switch is the down position, the microphone-
selection module 134 may select, for processing, the audio signals generated by the first microphone 112(1) and by the third microphone 112(3). For instance, thelight switch 108 may perform beamforming on these audio signals to generate one or more processed audio signals and may then send these one or more processed audio signals to theremote service 116. In some instances, when the switch is in the down position, thelight switch 108 may use an audio signal generated by the second microphone 112(2) for identifying and background noise and, thereafter, removing at least some of the background noise from the audio signals generated by the first send third microphones. - As described above, and as shown via the illustration of the side of the
light switch 108, because these microphones may be substantially vertically in-line with one another, these signals make good candidates for performing beamforming thereon to generate a processed audio signal having a relatively high SNR as compared to the signals by themselves. Further, in instances where the microphone-selection module 134 determines that the switch is the down position, the microphone-selection module 134 may send, to the remote service, audio signals generated by the first microphone 112(1) and by the third microphone 112(3). - Furthermore, and as illustrated, in some instances the voice-controlled
light switch 108 may include a fourth microphone 112(4) that is substantially horizontally in-line with, but to the left of, the second microphone 112(2), as well as a fifth microphone 112(5) that is substantially horizontally in-line with, but to the right of, the second microphone 112(2). Similarly, the voice-controlledlight switch 108 may include a sixth microphone 112(6) that is substantially horizontally in-line with, but to the left of, the third microphone 112(3), as well as a seventh microphone 112(7) that is substantially horizontally in-line with, but to the right of, the third microphone 112(3). - In these instances, in addition to determining whether the
switch 122 is in the up or down position prior to selecting audio signals for processing, the microphone-selection module 134 may also take into account a location of the audio source, such as the speaking user, within the environment. To do so, the microphone-selection module may utilize TDOA techniques, images captured by a camera, or the like. For instance, given that each of the microphones 112(1)-(7) may detect a command sound at different times, the microphone-selection module 134 may compare these different times to estimate the direction of the user relative to the voice-controlledlight switch 108. Once the direction of the user (or other audio source) has been estimated, the microphone-selection module 134 may determine which microphone on the faceplate 202 (or which combination of microphones) to send to theremote service 116. In one example, the microphone-selection engine defines a respective line between each microphone on the appropriate half of the faceplate 202 (e.g., each of the microphones on the top if theswitch 122 is in the up position) and the first microphone 112(1) on theswitch 122. The microphone-selection module 134 may then determine which line or lines most closely corresponds with the direction of the user. In some instances, the microphone-selection module 134 then selects this microphone pair and sends, to theremote service 116, the audio signals generated by this pair of microphones. To illustrate, if the user were standing relatively directly in front of the voice-controlledlight switch 108 and the switch were in the up position, the microphone-selection module 134 may select the second microphone 112(2) along with the microphone 112(1). If, however, the user were to the right side of the voice-controlledlight switch 108 when looking at thelight switch 108, the voice-controlledmodule 134 may select the fourth microphone 112(4). If, however, the user were to the left of the voice-controlledlight switch 108 when looking at thelight switch 108, the microphone-selection module 134 may select the fifth microphone 112(5) along with the first microphone 112(1). In other instances, meanwhile, the microphone-selection engine may send more than two audio signals to theremote service 116 for increased accuracy. For instance, if the user were standing center-left relative to the voice-controlledlight switch 108, the microphone-selection module 134 may send audio signals corresponding to the first microphone 112(1), the second microphone 112(2) and the fifth microphone 112(5). Further, it is noted that whileFIG. 2 and subsequent figures illustrate example arrays of microphones, it is to be appreciated that other implementations may utilize more microphones, fewer microphones, and/or different configurations of microphones. -
FIG. 3 illustrates another example configuration of the voice-controlledlight switch 108. Here, thelight switch 108 is in the form of a rocker switch, and includes two microphones on the switch 122 (or “rocker”) and microphones on the top and bottom of the faceplate. That is, theswitch 122 includes the first microphones 112(1) on the top portion of therocker switch 122, as well as an eighth microphone 112(8) on the bottom portion of theswitch 122. In addition, the top portion of thefaceplate 202 of the voice-controlledlight switch 108 again includes the three microphones 112(4), 112(2), and 112(5), while the bottom portion of thefaceplate 202 includes the three microphones 112(6), 112(3), and 112(7). - As with the example of
FIG. 2 , the microphone-selection module 134 may determine whether theswitch 122 is in the up or down position and may select, for processing, audio signals generated by microphones corresponding to the up or down position. In this case, for instance, if theswitch 122 is in the up position, the microphone-selection module 134 may select, for processing, audio signals generated by the eight microphone 112(8), as well as one or more of the three microphones 112(6), 112(3), and 112(7). That is, given that when therocker switch 122 is in the up position the first microphone 112(1) is nearly flush with thefaceplate 202 while the eight microphone 112(8) juts further out away from thefaceplate 202, and given that this separation between microphones allows for better beamforming, the microphone-selection engine may select the bottom microphones when theswitch 122 is in the up position. Similarly, when therocker switch 122 is determined to be in the down position, then the microphone-selection module 134 may select the microphones corresponding to the top of thelight switch 108, namely the first microphone 112(1) along with one or more of the three microphones 112(4), 112(2), and 112(5). -
FIG. 4 illustrates yet another example configuration of the voice-controlledlight switch 108 ofFIG. 1 . Here, the light switch is in the form of a sliding dimmer light switch that includes a slidingswitch 122, which itself includes a first microphone 112(1) on the left side of theswitch 122 and a second microphone 112(2) on the right side of theswitch 122. In addition, the examplelight switch 108 includes three microphones on the left side of thefaceplate plate 202 as well as three microphones on the right side of thefaceplate 202. In some instances, the microphone-selection module 134 may select one or more pairs of microphones for use based on the location of the slidingswitch 122, potentially along with the location of theuser 104. For instance, if the slidingswitch 122 is all the way to the top of its range, then the microphone-selection signal may select, for processing, audio signals corresponding to the microphone 112(5) and 112(1), or audio signals corresponding to the microphones 112(7) and 112(2). Similarly, if the slidingswitch 122 is approximately halfway down its range, as illustrated, then themicrophone selection module 134 may select, for processing, audio signals corresponding to the microphone 112(3) and 112(1), or audio signals corresponding to the microphones 112(4) and 112(2). In addition, if the slidingswitch 122 is located all the way down its range, then themicrophone selection module 134 may select, for processing, audio signals corresponding to the microphone 112(6) and 112(1), or audio signals corresponding to the microphones 112(8) and 112(2). -
FIG. 5 illustrates yet another example configuration of the voice-controlledlight switch 108 ofFIG. 1 . Here, thelight switch 108 is in the form of a rotatable-knob dimmer switch, which comprises arotatable switch 122 that a user is able to rotate radially to adjust an amount of power applied to a power load, similar to the dimmer light switch discussed above with reference toFIG. 4 . In this example, therotatable switch 122 may comprise the first microphone 112(1). In addition, thefaceplate 202 of thelight switch 108 may include one or more microphones, potentially equally distributed radially about therotatable switch 122. For instance, in this example thelight switch 108 includes microphones 112(2), 112(3), . . . , 112(9). When the voice-controlledlight switch 108 receives an instruction to process one or more audio signals, the microphone-selection module 134 may select audio signals corresponding to one or more pairs of microphones. For instance, the microphone-selection module 134 may determine which of the microphones 112(2)-(9) is closest to the current location of the microphone 112(1) and may send, to theremote service 116, the audio signal corresponding to this microphone and to the first microphone 112(1). For instance, as illustrated the microphone-selection module 134 may select the audio signal corresponding to the microphone 112(4) and may therefore send, to theremote service 116, the audio signals corresponding to the first microphone 112(1) and the fourth microphone 112(4). Further, whileFIG. 5 illustrates the first microphone 112(1) along with the additional microphones on the faceplate, in other instances therotatable knob 122 may include the first microphone on a face of the knob and a second microphone located nearer the face plate and vertically in-line with the first microphone. As such, no matter the position of the knob, the audio signals generated by these microphones may be used to increase the SNR of the signals using beamforming or other processing techniques. - In addition, while FIG. describes a light switch that couples a power source to a power load, in some instances an apparatus may have the same form factor as the apparatus of
FIG. 5 , while not coupling to a power source. This apparatus may instead simply comprise a rotatable know or other type switch that includes the first microphone 112(1) and one or more additional microphones, on a faceplate and/or on a second position on the knob closer to the faceplate. While this apparatus might not couple to a power source, this apparatus may include the same functionality for determining a position of the knob and selecting one or more audio signals for processing based on this position. -
FIG. 6 illustrates a flow diagram of anexample process 600 for selecting which audio signals to process based on a position of a switch in a voice-controlled light switch. Processing the signals may include sending these signals to a remote service for further processing, such as performing beamforming on the signals to generate a processed audio signal and performing speech recognition on the processed audio signal. Additionally or alternatively, processing these signals may include performing beamforming, noise-cancelation, or the like locally at the switch or at another device in the environment, and/or performing speech recognition locally at the switch or at another device in the environment. - The process 600 (and each process described herein) is illustrated as a logical flow graph, each operation of which represents a sequence of operations that can be implemented in hardware, software, or a combination thereof. In some instances, some or all of the operations are executed by the components of a voice-controlled light switch, such as the
light switch 108 described above. In the context of software, the operations represent computer-executable instructions stored on one or more computer-readable media that, when executed by one or more processors, perform the recited operations. Generally, computer-executable instructions include routines, programs, objects, components, data structures, and the like that perform particular functions or implement particular abstract data types. - The computer-readable media may include non-transitory computer-readable storage media, which may include hard drives, floppy diskettes, optical disks, CD-ROMs, DVDs, read-only memories (ROMs), random access memories (RAMs), EPROMs, EEPROMs, flash memory, magnetic or optical cards, solid-state memory devices, or other types of storage media suitable for storing electronic instructions. In addition, in some embodiments the computer-readable media may include a transitory computer-readable signal (in compressed or uncompressed form). Examples of computer-readable signals, whether modulated using a carrier or not, include, but are not limited to, signals that a computer system hosting or running a computer program can be configured to access, including signals downloaded through the Internet or other networks. Finally, the order in which the operations are described is not intended to be construed as a limitation, and any number of the described operations can be combined in any order and/or in parallel to implement the process.
- At 602, the
process 600 generates one or more audio signals. For instance, if a voice-controlled light switch comprises multiple microphones, one or more of these microphones may generate a corresponding audio signal based on sound detected within an environment of the light switch. At 604, theprocess 600 performs speech recognition on one or more of these generated audio signals. For instance, a speech-recognition module 134 operating on the light switch may perform speech recognition on one or more of the signals to identify a predefined word or phrase. At 606, theprocess 600 queries whether it has identified the predefined word or phrase based on the performing of the speech recognition. If not, then theprocess 600 returns to theoperation 604 to continue monitoring the audio signals for the predefined word or phrase. - If the
process 600 identifies the predefined word or phrase, however, then theprocess 600 determines, at 608, whether the switch is in a first position or a second position. While theexample process 600 describes determine whether the switch is in one of two positions, in other implementations any other number of positions may be utilized. In any event, if it is determined that the switch is in the first position, then at 610 the process 610(1) selects, for processing, audio signals corresponding to the first position for processing. As described above, these signals may also be selected based on other criteria, such as a location of the user within the environment of the light switch. Furthermore, and as illustrated, “processing” these audio signals may include, but is not limited to, sending the audio signals to a remote service for further processing, such as performing beamforming and speech recognition thereon (operation 610(1)), performing speech recognition locally at the light switch (operation 610(2)), and/or performing beamforming, noise cancellation, acoustic echo cancelation or the like on the audio signals locally at the light switch (operation 610(3)). Similarly, if theprocess 600 determines that the switch is in the second position, then at 612 theprocess 600 may select, for processing, audio signals corresponding to the second position. Again, processing of these signals may include sending the signals to a remote service (operation 612(1)), performing speech recognition on the signals locally (operation 612(2)), and/or performing beamforming or the like on the signals locally (operation 612(3)). -
FIG. 7 illustrates a flow diagram of anexample process 700 for selecting audio signals for processing based on a position of a switch, as well as based on a position of an audio source (e.g., a speaking user) within the environment. At 702, theprocess 700 determines that the switch is in the nth position (representative of any available position). At 704, theprocess 700 determines a first line defined by a microphone on a switch of a light switch and a first microphone on a faceplate that corresponds to the nth position. At 706, theprocess 700 determines a second line defined by the microphone on the switch and a second microphone on a faceplate that corresponds to the nth position. At 708, theprocess 700 determines a third line defined by the microphone on the switch and a third microphone on a faceplate that corresponds to the nth position. At 710, theprocess 700 estimates a direction of an audio source (e.g., a speaking user) within the environment using TDOA techniques, a camera or the like. Then, at 712, theprocess 700 queries whether the first, second, or third line is closest to the estimated direction of the audio source. If theprocess 700 determines that the first line is closest, then at 714(1) theprocess 700 selects, for processing, audio signals corresponding to the microphone on the switch and the first microphone on the faceplate. If, however, theprocess 700 determines that the second line is closest, then at 714(2) theprocess 700 selects, for processing, audio signals corresponding to the microphone on the switch and the second microphone on the faceplate. Finally, if theprocess 700 determines that the third line is closest, then at 714(3) theprocess 700 selects, for processing, audio signals corresponding to the microphone on the switch and the third microphone on the faceplate. Further, while three example faceplate microphones are described, it is to be appreciated that other implementations may utilize any other number of microphones on the faceplate. -
FIG. 8 shows a block diagram of selected functional components implemented in the voice-controlledlight switch 108 ofFIG. 1 . While this figure illustrates one examplelight switch 108, it is to be appreciated that any other type of client or server computing device may implement the techniques described herein. Furthermore, it is to be appreciated that thelight switch 108 may include some or all of the components described above with reference to theremote service 116. - In the illustrated implementation, the voice-controlled
light switch 108 includes theprocessor 126 andmemory 130. Thememory 130 may include computer-readable storage media (“CRSM”), which may be any available physical media accessible by theprocessor 126 to execute instructions stored on the memory. In one basic implementation, CRSM may include random access memory (“RAM”) and Flash memory. In other implementations, CRSM may include, but is not limited to, read-only memory (“ROM”), electrically erasable programmable read-only memory (“EEPROM”), or any other medium which can be used to store the desired information and which can be accessed by theprocessor 126. - The voice-controlled
light switch 108 includes the microphone(s) 112 that to receive audio input, such as user voice input. Thelight switch 108 may also include the speaker(s) 128 to output audio sounds. One ormore codecs 802 are coupled to the microphones and the speakers to encode and/or decode the audio signals. The codec may convert audio data between analog and digital formats. A user may interact with thelight switch 108 by speaking to it and themicrophones 112 capture sound and generate audio signals that includes the user speech. Thecodec 802 encodes the user speech and transfers that audio data to other components. - In the illustrated example, the voice-controlled
light switch 108 includes the one ormore network interfaces 124, which may couple to one ormore antennas 804 to facilitate a wireless connection to a network. The network interfaces 124 may implement one or more of various wireless technologies, such as WiFi, Bluetooth, RF, and so on. - One or more device interfaces 806 (e.g., USB, broadband connection, etc.) may further be provided as part of the
light switch 108 to facilitate a wired connection to a network, or a plug-in network device that communicates with other wireless networks. One ormore power units 808 may be further provided to distribute power to the various components on thelight switch 108. In some instances, meanwhile, the light switch couples to a persistent power source, such as a primary AC power supply. - Several modules such as instruction, datastores, and so forth may be stored within the
memory 130 and configured to execute on theprocessor 126. Anoperating system module 810 is configured to manage hardware and services (e.g., wireless unit, Codec, etc.) within and coupled to thelight switch 108 for the benefit of other modules. In addition, thememory 130 may include the speech-recognition engine 132, the microphone-selection module 134, and/or the signal-processing engine 140 described above, amongst other modules. Some or all of these modules, data stores, and components may reside additionally or alternatively at theremote service 116. - Furthermore, the
light switch 108 may include thecircuitry 120 for coupling the light switch to a power source and to a power load, such as a light, electrical outlet, or the like. At a minimum, the circuitry may include one or more wires for making this connection to the persistent power source. Finally, the light switch may include thephysical switch 122 that is selectable to provide power to the power load and to cease providing power to the power load. In some instances, and as discussed above, theswitch 122 may be adjustable to provide varying levels of power to the connected power load(s), while in other instances theswitch 122 may include two positions, on and off. Further, thecircuitry 120 may also include an internal switch that causes the one or more wires to couple together to cause the power source to provide power to the coupled power load. In some instances, this internal switch of thecircuitry 120 may couple the wires in this manner in response to a user flipping thephysical switch 122 or in response to execution of an instruction received based on a user issuing a voice command. - Although the subject matter has been described in language specific to structural features, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features described. Rather, the specific features are disclosed as illustrative forms of implementing the claims.
Claims (20)
1. An apparatus comprising:
one or more wires to couple the light switch to a power source and to at least one light;
a switch configured to toggle between at least an on position and an off position, the one or more wires providing power from the power source to the at least one light when the switch is in the on position and not providing power from the power source to the at least one light when the switch is in the off position;
a faceplate;
a first microphone located on a distal end of the switch;
a second microphone located at a first position of the faceplate, the first position being near the distal end of the switch when the switch is in the on position;
a third microphone located at a second position of the faceplate, the second position being near the distal end of the switch when the switch is in the off position;
a wireless network interface;
one or more processors;
one or more computer-readable media storing computer-executable instructions that, when executed, cause the one or more processors to perform acts comprising:
receiving a first audio signal generated by at least one of the first microphone, the second microphone, or the third microphone;
performing speech recognition on the first audio signal;
determining that the first audio signal includes data representing at least one predefined word;
determining that the switch is in the off position;
performing beamforming on a second audio signal generated by the first microphone and on a third audio signal generated by the third microphone to generate one or more processed audio signals; and
sending, to a remote device and via the wireless network interface, the one or more processed audio signals.
2. The apparatus as recited in claim 1 , the acts further comprising:
receiving, from the remote device, an instruction to turn on the at least one light; and
coupling the one or more wires to cause the one or more wires to provide power from the power source to the at least one light while the switch remains in the off position.
3. The apparatus as recited in claim 1 , wherein the first position is located approximately at a middle of a horizontal dimension of the faceplate and the second position is located approximately at the middle of the horizontal dimension of the faceplate, and further comprising:
a fourth microphone located substantially horizontally in-line with the second microphone and towards a left side of the faceplate; a fifth microphone located substantially horizontally in-line with the second microphone and towards a right side of the faceplate;
a sixth microphone located substantially horizontally in-line with the third microphone and towards the left side of the faceplate; and
a seventh microphone located substantially horizontally in-line with the third microphone and towards the right side of the faceplate.
4. The apparatus as recited in claim 3 , further comprising determining, prior to the sending, that a first line defined by the third microphone and the first microphone is more in-line with a user that stated the at least one predefined word than a second line defined by the sixth microphone and the first microphone and a third line defined by the seventh microphone and the first microphone.
5. An apparatus comprising:
one or more wires to couple the apparatus to an alternating-current (AC) power source located within a building and to at least one power load;
a switch configured to move from at least a first position to a second position, when the switch is in the first position the AC power source is electrically coupled to the at least one power load and when the switch is in the second position the AC power source is decoupled from the at least one power load;
a faceplate; and
a microphone located on the faceplate configured to generate an second audio signal.
6. The apparatus as recited in claim 5 , wherein the microphone comprises a first microphone, the audio signal comprises a first audio signal, and further comprising:
a second microphone located on the switch configured to generate a second audio signal;
a wireless network interface;
one or more processors;
one or more computer-readable media storing computer-executable instructions that, when executed, cause the one or more processors to perform acts comprising:
performing beamforming on the first audio signal and the second audio signal to generate one or more processed audio signals; and
sending, using the wireless network interface, the one or more processed audio signals to a remote device configured to perform speech recognition on the one or more processed audio signals.
7. The apparatus as recited in claim 5 , further comprising:
one or more processors; and
one or more computer-readable media storing computer-executable instructions that, when executed, cause the one or more processors to perform acts comprising:
performing speech recognition on the first the audio signal to at least one of identity data representing a predefined utterance or identify data representing a voice command related to the at least one power load or another power load.
8. The apparatus as recited in claim 5 , wherein the microphone comprises a first microphone, the audio signal comprises a first audio signal, and further comprising:
a second microphone located on the switch configured to generate a second audio signal;
a wireless network interface;
one or more processors; and
one or more computer-readable media storing computer-executable instructions that, when executed, cause the one or more processors to perform acts comprising:
performing speech recognition on at least one of the first audio signal or the second audio signal;
identifying data representing a predefined utterance from the at least one of the first audio signal or the second audio signal;
performing beamforming on the first audio signal and the second audio signal to generate one or more processed audio signals; and
sending, to a remote device and via the wireless network interface, the one or more processed audio signals.
9. The apparatus as recited in claim 5 , wherein the microphone comprises a first microphone, the audio signal comprises a first audio signal, the first microphone is located at a third position on the faceplate that is nearer the switch when the switch is in the first position, and further comprising:
a second microphone located on the switch configured to generate a second audio signal;
a third microphone located at a fourth position on the faceplate configured to generate a third audio signal, the fourth position being nearer the switch when the switch is in the first second position;
one or more processors; and
one or more computer-readable media storing computer-executable instructions that, when executed, cause the one or more processors to perform acts comprising:
determining that the switch is in the first position;
selecting the first audio signal generated by the first microphone for at least one of processing or sending at least partly in response to determining that the switch is in the first position;
determining that the switch is in the second position; and
selecting the third audio signal generated by the third microphone for at least one of processing or sending at least partly in response to determining that the switch is in the second position.
10. The apparatus as recited in claim 5 , wherein the microphone comprises a first microphone, the audio signal comprises a first audio signal, and further comprising:
a second microphone located on the switch configured to generate a second audio signal;
a third microphone located substantially horizontally in-line with the first microphone and to the left of the first microphone on the faceplate; and
a fourth microphone located substantially horizontally in-line with the first microphone and to the right of the first microphone on the faceplate.
11. The apparatus as recited in claim 5 , further comprising a button that is selectable to cause the microphone to generate the audio signal.
12. The apparatus as recited in claim 5 , wherein the switch comprises a toggle switch, a rocker switch, a dimmer switch, or a time-delay switch.
13. The apparatus as recited in claim 5 , wherein the at least one power load comprises a light, an electrical outlet, or a home appliance.
14. A method comprising:
receiving a first audio signal generated by a first microphone residing on a physical switch;
receiving a second audio signal generated by a second microphone;
receiving a third audio signal generated by a third microphone;
determining that the physical switch resides in a first position; and
at least partly based on the physical switch residing in the first position, selecting at least the first audio signal and the second audio signal for processing.
15. The method as recited in claim 0, wherein the selecting the first audio signal and the second audio signal for processing comprises sending the first audio signal and the second audio signal to a remote device without sending the third audio signal to the remote device.
16. The method as recited in claim 0, wherein the selecting the first audio signal and the second audio signal for processing comprises performing beamforming on the first audio signal and the second audio signal to generate one or more processed audio signals, and further comprising at least one of sending the one or more processed audio signals to a remote device or performing speech recognition on the one or more processed audio signals.
17. The method as recited in claim 0, further comprising:
performing speech recognition on at least one of the first audio signal, the second audio signal, or the third audio signal; and
determining that the at least one of the first audio signal, the second audio signal, or the third audio signal includes data representing a predefined utterance;
and wherein the selecting the first audio signal and the second audio signal for processing is further based at least in part on the determining that the at least one of the first audio signal, the second audio signal, or the third audio signal includes the data representing the predefined utterance.
18. The method as recited in claim 0, further comprising:
determining that the physical switch resides in a second position; and
at least partly in response to determining that the physical switch resides in the second position, selecting at least the first audio signal and the third audio signal for processing.
19. The method as recited in claim 0, further comprising:
receiving a fourth audio signal generated by a fourth microphone, the fourth microphone residing substantially horizontally in-line with the second microphone;
determining that a first line defined by the second microphone and the first microphone points closer to a source of audio than does a second line defined by the fourth microphone and the first microphone;
and wherein the selecting the first audio signal and the second audio signal for processing is further based at least in part on the determining that the first line points closer to the source of audio than does the second line.
20. The method as recited in claim 0, wherein:
the receiving the first audio signal comprises receiving the first audio signal generated by the first microphone residing on a physical switch of a light switch;
the receiving the second audio signal comprises receiving the second audio signal generated by the second microphone residing on a top portion of a faceplate of the light switch; and
the receiving the third audio signal comprises receiving the third audio signal generated by the third microphone residing on a bottom portion of the faceplate of the light switch.
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/980,392 US9826599B2 (en) | 2015-12-28 | 2015-12-28 | Voice-controlled light switches |
CN201680070020.3A CN108605001B (en) | 2015-12-28 | 2016-12-27 | Voice Controlled Light Switch |
PCT/US2016/068758 WO2017117149A1 (en) | 2015-12-28 | 2016-12-27 | Voice-controlled light switches |
EP16836191.3A EP3398298B1 (en) | 2015-12-28 | 2016-12-27 | Voice-controlled light switches |
US15/818,524 US10251244B1 (en) | 2015-12-28 | 2017-11-20 | Voice-controlled light switches |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/980,392 US9826599B2 (en) | 2015-12-28 | 2015-12-28 | Voice-controlled light switches |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/818,524 Continuation US10251244B1 (en) | 2015-12-28 | 2017-11-20 | Voice-controlled light switches |
Publications (2)
Publication Number | Publication Date |
---|---|
US20170188437A1 true US20170188437A1 (en) | 2017-06-29 |
US9826599B2 US9826599B2 (en) | 2017-11-21 |
Family
ID=58018197
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/980,392 Active US9826599B2 (en) | 2015-12-28 | 2015-12-28 | Voice-controlled light switches |
US15/818,524 Active US10251244B1 (en) | 2015-12-28 | 2017-11-20 | Voice-controlled light switches |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/818,524 Active US10251244B1 (en) | 2015-12-28 | 2017-11-20 | Voice-controlled light switches |
Country Status (4)
Country | Link |
---|---|
US (2) | US9826599B2 (en) |
EP (1) | EP3398298B1 (en) |
CN (1) | CN108605001B (en) |
WO (1) | WO2017117149A1 (en) |
Cited By (94)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108347813A (en) * | 2018-01-29 | 2018-07-31 | 安徽建筑大学 | A kind of intelligent building field building lighting control system |
WO2019013835A1 (en) * | 2017-07-12 | 2019-01-17 | Universal Electronics Inc. | Apparatus, system and method for directing voice input in a controlling device |
US10251244B1 (en) | 2015-12-28 | 2019-04-02 | Amazon Technologies, Inc. | Voice-controlled light switches |
US20190138704A1 (en) * | 2012-04-13 | 2019-05-09 | View, Inc. | Controlling optically-switchable devices |
US10410651B2 (en) * | 2016-12-29 | 2019-09-10 | Beijing Xiaoniao Tingting Technology Co., LTD. | De-reverberation control method and device of sound producing equipment |
EP3550939A1 (en) | 2018-04-02 | 2019-10-09 | Signify Holding B.V. | System and methods for augmenting voice commands using connected lighting systems |
US10477649B1 (en) * | 2017-07-31 | 2019-11-12 | Amazon Technologies, Inc. | Computing-device lighting schemas |
CN110536213A (en) * | 2018-05-24 | 2019-12-03 | 英飞凌科技股份有限公司 | System and method for monitoring |
USD883234S1 (en) | 2017-12-27 | 2020-05-05 | Leviton Manufacturing Co., Inc. | Wiring device |
US10711958B2 (en) | 2018-05-09 | 2020-07-14 | Samsung Electronics Co., Ltd. | LED device and LED lamp using the same |
CN111602112A (en) * | 2017-12-20 | 2020-08-28 | 豪倍公司 | Voice response wall-in device |
CN111833874A (en) * | 2020-07-10 | 2020-10-27 | 上海茂声智能科技有限公司 | Man-machine interaction method, system, equipment and storage medium based on identifier |
US10880284B1 (en) * | 2016-08-19 | 2020-12-29 | Amazon Technologies, Inc. | Repurposing limited functionality devices as authentication factors |
US10930276B2 (en) | 2017-07-12 | 2021-02-23 | Universal Electronics Inc. | Apparatus, system and method for directing voice input in a controlling device |
US11031014B2 (en) * | 2018-09-25 | 2021-06-08 | Sonos, Inc. | Voice detection optimization based on selected voice assistant service |
US11073800B2 (en) | 2011-03-16 | 2021-07-27 | View, Inc. | Monitoring sites containing switchable optical devices and controllers |
USD928729S1 (en) * | 2020-02-28 | 2021-08-24 | Enerlites, Inc. | Duplex paddle switch with changeable LED module |
USD928730S1 (en) * | 2020-02-28 | 2021-08-24 | Enerlites, Inc. | Triple paddle switch with changeable LED module |
USD931238S1 (en) * | 2020-03-30 | 2021-09-21 | Wenzhou Mtlc Electric Appliances Co., Ltd. | Butterfly dimmer |
US11175880B2 (en) | 2018-05-10 | 2021-11-16 | Sonos, Inc. | Systems and methods for voice-assisted media content selection |
US11175888B2 (en) | 2017-09-29 | 2021-11-16 | Sonos, Inc. | Media playback system with concurrent voice assistance |
US11183183B2 (en) | 2018-12-07 | 2021-11-23 | Sonos, Inc. | Systems and methods of operating media playback systems having multiple voice assistant services |
US11184704B2 (en) | 2016-02-22 | 2021-11-23 | Sonos, Inc. | Music service selection |
US11189286B2 (en) | 2019-10-22 | 2021-11-30 | Sonos, Inc. | VAS toggle based on device orientation |
US11200900B2 (en) | 2019-12-20 | 2021-12-14 | Sonos, Inc. | Offline voice control |
US11200889B2 (en) | 2018-11-15 | 2021-12-14 | Sonos, Inc. | Dilated convolutions and gating for efficient keyword spotting |
US11200894B2 (en) | 2019-06-12 | 2021-12-14 | Sonos, Inc. | Network microphone device with command keyword eventing |
US11224112B2 (en) | 2018-01-09 | 2022-01-11 | Lutron Technology Company Llc | Device for powering a modular assembly |
US11237796B2 (en) * | 2018-05-07 | 2022-02-01 | Google Llc | Methods, systems, and apparatus for providing composite graphical assistant interfaces for controlling connected devices |
US11283641B2 (en) * | 2018-04-20 | 2022-03-22 | Baidu Online Network Technology (Beijing) Co., Ltd. | Method, apparatus, device, system and storage medium for controlling smart home |
US11302341B2 (en) * | 2017-01-26 | 2022-04-12 | Yutou Technology (Hangzhou) Co., Ltd. | Microphone array based pickup method and system |
US11302326B2 (en) | 2017-09-28 | 2022-04-12 | Sonos, Inc. | Tone interference cancellation |
US11308958B2 (en) | 2020-02-07 | 2022-04-19 | Sonos, Inc. | Localized wakeword verification |
US11308962B2 (en) | 2020-05-20 | 2022-04-19 | Sonos, Inc. | Input detection windowing |
US11308961B2 (en) | 2016-10-19 | 2022-04-19 | Sonos, Inc. | Arbitration-based voice recognition |
US11315556B2 (en) | 2019-02-08 | 2022-04-26 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification |
US11343614B2 (en) | 2018-01-31 | 2022-05-24 | Sonos, Inc. | Device designation of playback and network microphone device arrangements |
US11354092B2 (en) | 2019-07-31 | 2022-06-07 | Sonos, Inc. | Noise classification for event detection |
US11361756B2 (en) | 2019-06-12 | 2022-06-14 | Sonos, Inc. | Conditional wake word eventing based on environment |
US11380322B2 (en) | 2017-08-07 | 2022-07-05 | Sonos, Inc. | Wake-word detection suppression |
US11381903B2 (en) | 2014-02-14 | 2022-07-05 | Sonic Blocks Inc. | Modular quick-connect A/V system and methods thereof |
US11405430B2 (en) | 2016-02-22 | 2022-08-02 | Sonos, Inc. | Networked microphone device control |
US11432030B2 (en) | 2018-09-14 | 2022-08-30 | Sonos, Inc. | Networked devices, systems, and methods for associating playback devices based on sound codes |
US11437020B2 (en) * | 2016-02-10 | 2022-09-06 | Cerence Operating Company | Techniques for spatially selective wake-up word recognition and related systems and methods |
US11451908B2 (en) | 2017-12-10 | 2022-09-20 | Sonos, Inc. | Network microphone devices with automatic do not disturb actuation capabilities |
US11482978B2 (en) | 2018-08-28 | 2022-10-25 | Sonos, Inc. | Audio notifications |
US11482224B2 (en) | 2020-05-20 | 2022-10-25 | Sonos, Inc. | Command keywords with input detection windowing |
US11500611B2 (en) | 2017-09-08 | 2022-11-15 | Sonos, Inc. | Dynamic computation of system response volume |
US11501773B2 (en) | 2019-06-12 | 2022-11-15 | Sonos, Inc. | Network microphone device with command keyword conditioning |
US11501795B2 (en) | 2018-09-29 | 2022-11-15 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection via multiple network microphone devices |
US11514898B2 (en) | 2016-02-22 | 2022-11-29 | Sonos, Inc. | Voice control of a media playback system |
US11513763B2 (en) | 2016-02-22 | 2022-11-29 | Sonos, Inc. | Audio response playback |
US11516610B2 (en) | 2016-09-30 | 2022-11-29 | Sonos, Inc. | Orientation-based playback device microphone selection |
US11531520B2 (en) | 2016-08-05 | 2022-12-20 | Sonos, Inc. | Playback device supporting concurrent voice assistants |
US11538451B2 (en) | 2017-09-28 | 2022-12-27 | Sonos, Inc. | Multi-channel acoustic echo cancellation |
US11538460B2 (en) | 2018-12-13 | 2022-12-27 | Sonos, Inc. | Networked microphone devices, systems, and methods of localized arbitration |
US11540047B2 (en) | 2018-12-20 | 2022-12-27 | Sonos, Inc. | Optimization of network microphone devices using noise classification |
US11545169B2 (en) | 2016-06-09 | 2023-01-03 | Sonos, Inc. | Dynamic player selection for audio signal processing |
US11551669B2 (en) | 2019-07-31 | 2023-01-10 | Sonos, Inc. | Locally distributed keyword detection |
US11551700B2 (en) | 2021-01-25 | 2023-01-10 | Sonos, Inc. | Systems and methods for power-efficient keyword detection |
US11556307B2 (en) | 2020-01-31 | 2023-01-17 | Sonos, Inc. | Local voice data processing |
US11556306B2 (en) | 2016-02-22 | 2023-01-17 | Sonos, Inc. | Voice controlled media playback system |
US11563842B2 (en) | 2018-08-28 | 2023-01-24 | Sonos, Inc. | Do not disturb feature for audio notifications |
US11562740B2 (en) | 2020-01-07 | 2023-01-24 | Sonos, Inc. | Voice verification for media playback |
US11592723B2 (en) | 2009-12-22 | 2023-02-28 | View, Inc. | Automated commissioning of controllers in a window network |
US11641559B2 (en) | 2016-09-27 | 2023-05-02 | Sonos, Inc. | Audio playback settings for voice interaction |
US11646023B2 (en) | 2019-02-08 | 2023-05-09 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing |
US11646045B2 (en) | 2017-09-27 | 2023-05-09 | Sonos, Inc. | Robust short-time fourier transform acoustic echo cancellation during audio playback |
US11664023B2 (en) | 2016-07-15 | 2023-05-30 | Sonos, Inc. | Voice detection by multiple devices |
US11676590B2 (en) | 2017-12-11 | 2023-06-13 | Sonos, Inc. | Home graph |
US11696074B2 (en) | 2018-06-28 | 2023-07-04 | Sonos, Inc. | Systems and methods for associating playback devices with voice assistant services |
US11698771B2 (en) | 2020-08-25 | 2023-07-11 | Sonos, Inc. | Vocal guidance engines for playback devices |
US11710487B2 (en) | 2019-07-31 | 2023-07-25 | Sonos, Inc. | Locally distributed keyword detection |
US11715489B2 (en) | 2018-05-18 | 2023-08-01 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection |
US11727919B2 (en) | 2020-05-20 | 2023-08-15 | Sonos, Inc. | Memory allocation for keyword spotting engines |
US11726742B2 (en) | 2016-02-22 | 2023-08-15 | Sonos, Inc. | Handling of loss of pairing between networked devices |
CN116627070A (en) * | 2023-05-26 | 2023-08-22 | 中山亿联智能科技有限公司 | A single-fire voice panel realizes a dual-control method |
US11733660B2 (en) | 2014-03-05 | 2023-08-22 | View, Inc. | Monitoring sites containing switchable optical devices and controllers |
US11790937B2 (en) | 2018-09-21 | 2023-10-17 | Sonos, Inc. | Voice detection optimization using sound metadata |
US11792590B2 (en) | 2018-05-25 | 2023-10-17 | Sonos, Inc. | Determining and adapting to changes in microphone performance of playback devices |
US11790911B2 (en) | 2018-09-28 | 2023-10-17 | Sonos, Inc. | Systems and methods for selective wake word detection using neural network models |
US11798553B2 (en) | 2019-05-03 | 2023-10-24 | Sonos, Inc. | Voice assistant persistence across multiple network microphone devices |
US11892810B1 (en) * | 2019-04-23 | 2024-02-06 | Josh.Ai, Inc. | Contextual home automation toggle |
US20240046928A1 (en) * | 2012-04-13 | 2024-02-08 | View, Inc. | Controlling optically-switchable devices |
US11899519B2 (en) | 2018-10-23 | 2024-02-13 | Sonos, Inc. | Multiple stage network microphone device with reduced power consumption and processing load |
US11979960B2 (en) | 2016-07-15 | 2024-05-07 | Sonos, Inc. | Contextualization of voice inputs |
US11984123B2 (en) | 2020-11-12 | 2024-05-14 | Sonos, Inc. | Network device interaction by range |
US12002632B2 (en) | 2020-10-09 | 2024-06-04 | Leviton Manufacturing Co., Inc. | Anywhere wireless switch and/or dimmer |
US12047753B1 (en) | 2017-09-28 | 2024-07-23 | Sonos, Inc. | Three-dimensional beam forming with a microphone array |
US20240407065A1 (en) * | 2023-06-01 | 2024-12-05 | Roku, Inc. | Lighting Device with Audio Signal Control Feature |
US20240431005A1 (en) * | 2021-01-25 | 2024-12-26 | Signify Holding B.V. | Selecting a set of lighting devices based on an identifier of an audio and/or video signal source |
US12217748B2 (en) | 2017-03-27 | 2025-02-04 | Sonos, Inc. | Systems and methods of multiple voice services |
US12244129B2 (en) | 2019-12-24 | 2025-03-04 | Leviton Manufacturing Co., Inc. | Smart wall-plate system |
US12283269B2 (en) | 2020-10-16 | 2025-04-22 | Sonos, Inc. | Intent inference in audiovisual communication sessions |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020018536A1 (en) | 2018-07-17 | 2020-01-23 | iT SpeeX LLC | Method, system, and computer program product for communication with an intelligent industrial assistant and industrial machine |
US11514178B2 (en) | 2018-07-17 | 2022-11-29 | iT SpeeX LLC | Method, system, and computer program product for role- and skill-based privileges for an intelligent industrial assistant |
US11232262B2 (en) | 2018-07-17 | 2022-01-25 | iT SpeeX LLC | Method, system, and computer program product for an intelligent industrial assistant |
TW202046159A (en) | 2019-02-08 | 2020-12-16 | 美商艾提史畢克斯有限責任公司 | Method, system, and computer program product for developing dialogue templates for an intelligent industrial assistant |
CN109991675B (en) * | 2019-03-26 | 2020-11-10 | 中国石油化工股份有限公司 | Method for determining maturity of crude oil by using absolute content of terpene alkane in crude oil |
CN110234043B (en) * | 2019-05-31 | 2020-08-25 | 歌尔科技有限公司 | Sound signal processing method, device and equipment based on microphone array |
JP7367963B2 (en) * | 2019-09-05 | 2023-10-24 | 学校法人上智学院 | Auxiliary equipment for the detection of voices |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7418392B1 (en) * | 2003-09-25 | 2008-08-26 | Sensory, Inc. | System and method for controlling the operation of a device by voice commands |
US20120147594A1 (en) * | 2009-09-15 | 2012-06-14 | Tait Cynthia K | Purselight |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4090032A (en) * | 1976-05-05 | 1978-05-16 | Wm. A. Holmin Corporation | Control system for audio amplifying system having multiple microphones |
US5914826A (en) | 1996-07-24 | 1999-06-22 | Logic Laboratories, Inc. | Light switch cover plate with audio recording and playback feature |
US6188986B1 (en) | 1998-01-02 | 2001-02-13 | Vos Systems, Inc. | Voice activated switch method and apparatus |
US20120105740A1 (en) * | 2000-06-02 | 2012-05-03 | Oakley, Inc. | Eyewear with detachable adjustable electronics module |
US20030179894A1 (en) * | 2002-03-21 | 2003-09-25 | Siemens Hearing Instruments, Inc. | Directional microphone hearing aid system |
US20080218372A1 (en) | 2007-03-07 | 2008-09-11 | Optimal Licensing Corporation | Light switch used as a communication device |
CN102118661A (en) * | 2009-12-30 | 2011-07-06 | 富准精密工业(深圳)有限公司 | Microphone module |
KR101067387B1 (en) * | 2011-04-15 | 2011-09-23 | (주)알고코리아 | Hearing Aid System Using Wireless Optical Communication |
DE102011083209B3 (en) * | 2011-09-22 | 2013-01-17 | Siemens Medical Instruments Pte. Ltd. | IDO hearing instrument with programming socket |
KR101641448B1 (en) * | 2012-03-16 | 2016-07-20 | 뉘앙스 커뮤니케이션즈, 인코포레이티드 | User dedicated automatic speech recognition |
US9060224B1 (en) | 2012-06-01 | 2015-06-16 | Rawles Llc | Voice controlled assistant with coaxial speaker and microphone arrangement |
US9747899B2 (en) | 2013-06-27 | 2017-08-29 | Amazon Technologies, Inc. | Detecting self-generated wake expressions |
US9871974B2 (en) * | 2014-12-05 | 2018-01-16 | Amx, Llc | Computing device camera view controller and shutter |
US9826599B2 (en) | 2015-12-28 | 2017-11-21 | Amazon Technologies, Inc. | Voice-controlled light switches |
-
2015
- 2015-12-28 US US14/980,392 patent/US9826599B2/en active Active
-
2016
- 2016-12-27 EP EP16836191.3A patent/EP3398298B1/en active Active
- 2016-12-27 WO PCT/US2016/068758 patent/WO2017117149A1/en unknown
- 2016-12-27 CN CN201680070020.3A patent/CN108605001B/en active Active
-
2017
- 2017-11-20 US US15/818,524 patent/US10251244B1/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7418392B1 (en) * | 2003-09-25 | 2008-08-26 | Sensory, Inc. | System and method for controlling the operation of a device by voice commands |
US20120147594A1 (en) * | 2009-09-15 | 2012-06-14 | Tait Cynthia K | Purselight |
Cited By (138)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11592723B2 (en) | 2009-12-22 | 2023-02-28 | View, Inc. | Automated commissioning of controllers in a window network |
US11073800B2 (en) | 2011-03-16 | 2021-07-27 | View, Inc. | Monitoring sites containing switchable optical devices and controllers |
US11735183B2 (en) * | 2012-04-13 | 2023-08-22 | View, Inc. | Controlling optically-switchable devices |
US11687045B2 (en) | 2012-04-13 | 2023-06-27 | View, Inc. | Monitoring sites containing switchable optical devices and controllers |
US20240046928A1 (en) * | 2012-04-13 | 2024-02-08 | View, Inc. | Controlling optically-switchable devices |
US20190138704A1 (en) * | 2012-04-13 | 2019-05-09 | View, Inc. | Controlling optically-switchable devices |
US20210174804A1 (en) * | 2012-04-13 | 2021-06-10 | View, Inc. | Controlling optically-switchable devices |
US10964320B2 (en) * | 2012-04-13 | 2021-03-30 | View, Inc. | Controlling optically-switchable devices |
US11381903B2 (en) | 2014-02-14 | 2022-07-05 | Sonic Blocks Inc. | Modular quick-connect A/V system and methods thereof |
US12225344B2 (en) | 2014-02-14 | 2025-02-11 | Sonic Blocks, Inc. | Modular quick-connect A/V system and methods thereof |
US11733660B2 (en) | 2014-03-05 | 2023-08-22 | View, Inc. | Monitoring sites containing switchable optical devices and controllers |
US12130597B2 (en) | 2014-03-05 | 2024-10-29 | View, Inc. | Monitoring sites containing switchable optical devices and controllers |
US10251244B1 (en) | 2015-12-28 | 2019-04-02 | Amazon Technologies, Inc. | Voice-controlled light switches |
US11437020B2 (en) * | 2016-02-10 | 2022-09-06 | Cerence Operating Company | Techniques for spatially selective wake-up word recognition and related systems and methods |
US11513763B2 (en) | 2016-02-22 | 2022-11-29 | Sonos, Inc. | Audio response playback |
US11212612B2 (en) | 2016-02-22 | 2021-12-28 | Sonos, Inc. | Voice control of a media playback system |
US11405430B2 (en) | 2016-02-22 | 2022-08-02 | Sonos, Inc. | Networked microphone device control |
US11832068B2 (en) | 2016-02-22 | 2023-11-28 | Sonos, Inc. | Music service selection |
US11863593B2 (en) | 2016-02-22 | 2024-01-02 | Sonos, Inc. | Networked microphone device control |
US11514898B2 (en) | 2016-02-22 | 2022-11-29 | Sonos, Inc. | Voice control of a media playback system |
US11736860B2 (en) | 2016-02-22 | 2023-08-22 | Sonos, Inc. | Voice control of a media playback system |
US11750969B2 (en) | 2016-02-22 | 2023-09-05 | Sonos, Inc. | Default playback device designation |
US11726742B2 (en) | 2016-02-22 | 2023-08-15 | Sonos, Inc. | Handling of loss of pairing between networked devices |
US12047752B2 (en) | 2016-02-22 | 2024-07-23 | Sonos, Inc. | Content mixing |
US11556306B2 (en) | 2016-02-22 | 2023-01-17 | Sonos, Inc. | Voice controlled media playback system |
US11983463B2 (en) | 2016-02-22 | 2024-05-14 | Sonos, Inc. | Metadata exchange involving a networked playback system and a networked microphone system |
US11184704B2 (en) | 2016-02-22 | 2021-11-23 | Sonos, Inc. | Music service selection |
US11545169B2 (en) | 2016-06-09 | 2023-01-03 | Sonos, Inc. | Dynamic player selection for audio signal processing |
US11979960B2 (en) | 2016-07-15 | 2024-05-07 | Sonos, Inc. | Contextualization of voice inputs |
US11664023B2 (en) | 2016-07-15 | 2023-05-30 | Sonos, Inc. | Voice detection by multiple devices |
US11531520B2 (en) | 2016-08-05 | 2022-12-20 | Sonos, Inc. | Playback device supporting concurrent voice assistants |
US10880284B1 (en) * | 2016-08-19 | 2020-12-29 | Amazon Technologies, Inc. | Repurposing limited functionality devices as authentication factors |
US11641559B2 (en) | 2016-09-27 | 2023-05-02 | Sonos, Inc. | Audio playback settings for voice interaction |
US11516610B2 (en) | 2016-09-30 | 2022-11-29 | Sonos, Inc. | Orientation-based playback device microphone selection |
US11727933B2 (en) | 2016-10-19 | 2023-08-15 | Sonos, Inc. | Arbitration-based voice recognition |
US11308961B2 (en) | 2016-10-19 | 2022-04-19 | Sonos, Inc. | Arbitration-based voice recognition |
US10410651B2 (en) * | 2016-12-29 | 2019-09-10 | Beijing Xiaoniao Tingting Technology Co., LTD. | De-reverberation control method and device of sound producing equipment |
US11302341B2 (en) * | 2017-01-26 | 2022-04-12 | Yutou Technology (Hangzhou) Co., Ltd. | Microphone array based pickup method and system |
US12217748B2 (en) | 2017-03-27 | 2025-02-04 | Sonos, Inc. | Systems and methods of multiple voice services |
US11631403B2 (en) | 2017-07-12 | 2023-04-18 | Universal Electronics Inc. | Apparatus, system and method for directing voice input in a controlling device |
US11489691B2 (en) | 2017-07-12 | 2022-11-01 | Universal Electronics Inc. | Apparatus, system and method for directing voice input in a controlling device |
WO2019013835A1 (en) * | 2017-07-12 | 2019-01-17 | Universal Electronics Inc. | Apparatus, system and method for directing voice input in a controlling device |
US11985003B2 (en) | 2017-07-12 | 2024-05-14 | Universal Electronics Inc. | Apparatus, system and method for directing voice input in a controlling device |
US10930276B2 (en) | 2017-07-12 | 2021-02-23 | Universal Electronics Inc. | Apparatus, system and method for directing voice input in a controlling device |
US10477649B1 (en) * | 2017-07-31 | 2019-11-12 | Amazon Technologies, Inc. | Computing-device lighting schemas |
US11900937B2 (en) | 2017-08-07 | 2024-02-13 | Sonos, Inc. | Wake-word detection suppression |
US11380322B2 (en) | 2017-08-07 | 2022-07-05 | Sonos, Inc. | Wake-word detection suppression |
US11500611B2 (en) | 2017-09-08 | 2022-11-15 | Sonos, Inc. | Dynamic computation of system response volume |
US11646045B2 (en) | 2017-09-27 | 2023-05-09 | Sonos, Inc. | Robust short-time fourier transform acoustic echo cancellation during audio playback |
US11538451B2 (en) | 2017-09-28 | 2022-12-27 | Sonos, Inc. | Multi-channel acoustic echo cancellation |
US12236932B2 (en) | 2017-09-28 | 2025-02-25 | Sonos, Inc. | Multi-channel acoustic echo cancellation |
US12047753B1 (en) | 2017-09-28 | 2024-07-23 | Sonos, Inc. | Three-dimensional beam forming with a microphone array |
US11769505B2 (en) | 2017-09-28 | 2023-09-26 | Sonos, Inc. | Echo of tone interferance cancellation using two acoustic echo cancellers |
US11302326B2 (en) | 2017-09-28 | 2022-04-12 | Sonos, Inc. | Tone interference cancellation |
US11288039B2 (en) | 2017-09-29 | 2022-03-29 | Sonos, Inc. | Media playback system with concurrent voice assistance |
US11893308B2 (en) | 2017-09-29 | 2024-02-06 | Sonos, Inc. | Media playback system with concurrent voice assistance |
US11175888B2 (en) | 2017-09-29 | 2021-11-16 | Sonos, Inc. | Media playback system with concurrent voice assistance |
US11451908B2 (en) | 2017-12-10 | 2022-09-20 | Sonos, Inc. | Network microphone devices with automatic do not disturb actuation capabilities |
US11676590B2 (en) | 2017-12-11 | 2023-06-13 | Sonos, Inc. | Home graph |
CN111602112A (en) * | 2017-12-20 | 2020-08-28 | 豪倍公司 | Voice response wall-in device |
US11513764B2 (en) | 2017-12-27 | 2022-11-29 | Leviton Manufacturing Co., Inc. | Wireless enabled load control device with voice controller |
USD883234S1 (en) | 2017-12-27 | 2020-05-05 | Leviton Manufacturing Co., Inc. | Wiring device |
US12022594B2 (en) | 2018-01-09 | 2024-06-25 | Lutron Technology Company Llc | Device for powering a modular assembly |
US11224112B2 (en) | 2018-01-09 | 2022-01-11 | Lutron Technology Company Llc | Device for powering a modular assembly |
CN108347813A (en) * | 2018-01-29 | 2018-07-31 | 安徽建筑大学 | A kind of intelligent building field building lighting control system |
US11689858B2 (en) | 2018-01-31 | 2023-06-27 | Sonos, Inc. | Device designation of playback and network microphone device arrangements |
US11343614B2 (en) | 2018-01-31 | 2022-05-24 | Sonos, Inc. | Device designation of playback and network microphone device arrangements |
US20210029803A1 (en) * | 2018-04-02 | 2021-01-28 | Signify Holding B.V. | System and methods for augmenting voice commands using connected lighting systems |
CN111903194A (en) * | 2018-04-02 | 2020-11-06 | 昕诺飞控股有限公司 | System and method for enhancing voice commands using connected lighting systems |
EP3550939A1 (en) | 2018-04-02 | 2019-10-09 | Signify Holding B.V. | System and methods for augmenting voice commands using connected lighting systems |
WO2019192869A1 (en) | 2018-04-02 | 2019-10-10 | Signify Holding B.V. | System and methods for augmenting voice commands using connected lighting systems |
US11283641B2 (en) * | 2018-04-20 | 2022-03-22 | Baidu Online Network Technology (Beijing) Co., Ltd. | Method, apparatus, device, system and storage medium for controlling smart home |
US11237796B2 (en) * | 2018-05-07 | 2022-02-01 | Google Llc | Methods, systems, and apparatus for providing composite graphical assistant interfaces for controlling connected devices |
US10711958B2 (en) | 2018-05-09 | 2020-07-14 | Samsung Electronics Co., Ltd. | LED device and LED lamp using the same |
US11175880B2 (en) | 2018-05-10 | 2021-11-16 | Sonos, Inc. | Systems and methods for voice-assisted media content selection |
US11797263B2 (en) | 2018-05-10 | 2023-10-24 | Sonos, Inc. | Systems and methods for voice-assisted media content selection |
US11715489B2 (en) | 2018-05-18 | 2023-08-01 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection |
CN110536213A (en) * | 2018-05-24 | 2019-12-03 | 英飞凌科技股份有限公司 | System and method for monitoring |
US11792590B2 (en) | 2018-05-25 | 2023-10-17 | Sonos, Inc. | Determining and adapting to changes in microphone performance of playback devices |
US11696074B2 (en) | 2018-06-28 | 2023-07-04 | Sonos, Inc. | Systems and methods for associating playback devices with voice assistant services |
US11563842B2 (en) | 2018-08-28 | 2023-01-24 | Sonos, Inc. | Do not disturb feature for audio notifications |
US11482978B2 (en) | 2018-08-28 | 2022-10-25 | Sonos, Inc. | Audio notifications |
US11778259B2 (en) | 2018-09-14 | 2023-10-03 | Sonos, Inc. | Networked devices, systems and methods for associating playback devices based on sound codes |
US11432030B2 (en) | 2018-09-14 | 2022-08-30 | Sonos, Inc. | Networked devices, systems, and methods for associating playback devices based on sound codes |
US11790937B2 (en) | 2018-09-21 | 2023-10-17 | Sonos, Inc. | Voice detection optimization using sound metadata |
US12230291B2 (en) | 2018-09-21 | 2025-02-18 | Sonos, Inc. | Voice detection optimization using sound metadata |
US11031014B2 (en) * | 2018-09-25 | 2021-06-08 | Sonos, Inc. | Voice detection optimization based on selected voice assistant service |
US11727936B2 (en) | 2018-09-25 | 2023-08-15 | Sonos, Inc. | Voice detection optimization based on selected voice assistant service |
US12165651B2 (en) | 2018-09-25 | 2024-12-10 | Sonos, Inc. | Voice detection optimization based on selected voice assistant service |
US11790911B2 (en) | 2018-09-28 | 2023-10-17 | Sonos, Inc. | Systems and methods for selective wake word detection using neural network models |
US12165644B2 (en) | 2018-09-28 | 2024-12-10 | Sonos, Inc. | Systems and methods for selective wake word detection |
US12062383B2 (en) | 2018-09-29 | 2024-08-13 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection via multiple network microphone devices |
US11501795B2 (en) | 2018-09-29 | 2022-11-15 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection via multiple network microphone devices |
US11899519B2 (en) | 2018-10-23 | 2024-02-13 | Sonos, Inc. | Multiple stage network microphone device with reduced power consumption and processing load |
US11200889B2 (en) | 2018-11-15 | 2021-12-14 | Sonos, Inc. | Dilated convolutions and gating for efficient keyword spotting |
US11741948B2 (en) | 2018-11-15 | 2023-08-29 | Sonos Vox France Sas | Dilated convolutions and gating for efficient keyword spotting |
US11183183B2 (en) | 2018-12-07 | 2021-11-23 | Sonos, Inc. | Systems and methods of operating media playback systems having multiple voice assistant services |
US11557294B2 (en) | 2018-12-07 | 2023-01-17 | Sonos, Inc. | Systems and methods of operating media playback systems having multiple voice assistant services |
US11538460B2 (en) | 2018-12-13 | 2022-12-27 | Sonos, Inc. | Networked microphone devices, systems, and methods of localized arbitration |
US11540047B2 (en) | 2018-12-20 | 2022-12-27 | Sonos, Inc. | Optimization of network microphone devices using noise classification |
US11315556B2 (en) | 2019-02-08 | 2022-04-26 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification |
US11646023B2 (en) | 2019-02-08 | 2023-05-09 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing |
US11892810B1 (en) * | 2019-04-23 | 2024-02-06 | Josh.Ai, Inc. | Contextual home automation toggle |
US11798553B2 (en) | 2019-05-03 | 2023-10-24 | Sonos, Inc. | Voice assistant persistence across multiple network microphone devices |
US11200894B2 (en) | 2019-06-12 | 2021-12-14 | Sonos, Inc. | Network microphone device with command keyword eventing |
US11501773B2 (en) | 2019-06-12 | 2022-11-15 | Sonos, Inc. | Network microphone device with command keyword conditioning |
US11361756B2 (en) | 2019-06-12 | 2022-06-14 | Sonos, Inc. | Conditional wake word eventing based on environment |
US11854547B2 (en) | 2019-06-12 | 2023-12-26 | Sonos, Inc. | Network microphone device with command keyword eventing |
US11710487B2 (en) | 2019-07-31 | 2023-07-25 | Sonos, Inc. | Locally distributed keyword detection |
US11354092B2 (en) | 2019-07-31 | 2022-06-07 | Sonos, Inc. | Noise classification for event detection |
US11714600B2 (en) | 2019-07-31 | 2023-08-01 | Sonos, Inc. | Noise classification for event detection |
US12211490B2 (en) | 2019-07-31 | 2025-01-28 | Sonos, Inc. | Locally distributed keyword detection |
US11551669B2 (en) | 2019-07-31 | 2023-01-10 | Sonos, Inc. | Locally distributed keyword detection |
US11862161B2 (en) | 2019-10-22 | 2024-01-02 | Sonos, Inc. | VAS toggle based on device orientation |
US11189286B2 (en) | 2019-10-22 | 2021-11-30 | Sonos, Inc. | VAS toggle based on device orientation |
US11869503B2 (en) | 2019-12-20 | 2024-01-09 | Sonos, Inc. | Offline voice control |
US11200900B2 (en) | 2019-12-20 | 2021-12-14 | Sonos, Inc. | Offline voice control |
US12244129B2 (en) | 2019-12-24 | 2025-03-04 | Leviton Manufacturing Co., Inc. | Smart wall-plate system |
US11562740B2 (en) | 2020-01-07 | 2023-01-24 | Sonos, Inc. | Voice verification for media playback |
US11556307B2 (en) | 2020-01-31 | 2023-01-17 | Sonos, Inc. | Local voice data processing |
US11961519B2 (en) | 2020-02-07 | 2024-04-16 | Sonos, Inc. | Localized wakeword verification |
US11308958B2 (en) | 2020-02-07 | 2022-04-19 | Sonos, Inc. | Localized wakeword verification |
USD928730S1 (en) * | 2020-02-28 | 2021-08-24 | Enerlites, Inc. | Triple paddle switch with changeable LED module |
USD928729S1 (en) * | 2020-02-28 | 2021-08-24 | Enerlites, Inc. | Duplex paddle switch with changeable LED module |
USD931238S1 (en) * | 2020-03-30 | 2021-09-21 | Wenzhou Mtlc Electric Appliances Co., Ltd. | Butterfly dimmer |
US11694689B2 (en) | 2020-05-20 | 2023-07-04 | Sonos, Inc. | Input detection windowing |
US11727919B2 (en) | 2020-05-20 | 2023-08-15 | Sonos, Inc. | Memory allocation for keyword spotting engines |
US11308962B2 (en) | 2020-05-20 | 2022-04-19 | Sonos, Inc. | Input detection windowing |
US11482224B2 (en) | 2020-05-20 | 2022-10-25 | Sonos, Inc. | Command keywords with input detection windowing |
CN111833874A (en) * | 2020-07-10 | 2020-10-27 | 上海茂声智能科技有限公司 | Man-machine interaction method, system, equipment and storage medium based on identifier |
US11698771B2 (en) | 2020-08-25 | 2023-07-11 | Sonos, Inc. | Vocal guidance engines for playback devices |
US12002632B2 (en) | 2020-10-09 | 2024-06-04 | Leviton Manufacturing Co., Inc. | Anywhere wireless switch and/or dimmer |
US12283269B2 (en) | 2020-10-16 | 2025-04-22 | Sonos, Inc. | Intent inference in audiovisual communication sessions |
US11984123B2 (en) | 2020-11-12 | 2024-05-14 | Sonos, Inc. | Network device interaction by range |
US20240431005A1 (en) * | 2021-01-25 | 2024-12-26 | Signify Holding B.V. | Selecting a set of lighting devices based on an identifier of an audio and/or video signal source |
US11551700B2 (en) | 2021-01-25 | 2023-01-10 | Sonos, Inc. | Systems and methods for power-efficient keyword detection |
CN116627070A (en) * | 2023-05-26 | 2023-08-22 | 中山亿联智能科技有限公司 | A single-fire voice panel realizes a dual-control method |
US20240407065A1 (en) * | 2023-06-01 | 2024-12-05 | Roku, Inc. | Lighting Device with Audio Signal Control Feature |
Also Published As
Publication number | Publication date |
---|---|
EP3398298B1 (en) | 2022-06-08 |
WO2017117149A1 (en) | 2017-07-06 |
EP3398298A1 (en) | 2018-11-07 |
CN108605001B (en) | 2021-10-22 |
US9826599B2 (en) | 2017-11-21 |
US10251244B1 (en) | 2019-04-02 |
CN108605001A (en) | 2018-09-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10251244B1 (en) | Voice-controlled light switches | |
US11902707B1 (en) | Location based device grouping with voice control | |
US12118995B1 (en) | Identifying a location of a voice-input device | |
US12170087B1 (en) | Altering audio to improve automatic speech recognition | |
US11624800B1 (en) | Beam rejection in multi-beam microphone systems | |
US11429345B2 (en) | Remote execution of secondary-device drivers | |
US10672387B2 (en) | Systems and methods for recognizing user speech | |
US10297250B1 (en) | Asynchronous transfer of audio data | |
US10755706B2 (en) | Voice-based user interface with dynamically switchable endpoints | |
US10713007B2 (en) | Architecture for a hub configured to control a second device while a connection to a remote system is unavailable | |
US9729821B1 (en) | Sensor fusion for location based device grouping | |
US10249299B1 (en) | Tailoring beamforming techniques to environments | |
US9595997B1 (en) | Adaption-based reduction of echo and noise | |
US9098467B1 (en) | Accepting voice commands based on user identity | |
US9494683B1 (en) | Audio-based gesture detection | |
US11722985B2 (en) | Object tracking and authentication using modular wall units | |
CN109920419B (en) | Voice control method and device, electronic equipment and computer readable medium | |
US20200301378A1 (en) | Deducing floor plans using modular wall units | |
CN111630413B (en) | Confidence-based application-specific user interaction | |
WO2019183257A1 (en) | Modular wall unit system | |
CN112151013A (en) | Intelligent equipment interaction method | |
US10874011B1 (en) | Communication interface for front panel and power module | |
EP4005229A2 (en) | Acoustic zoning with distributed microphones | |
US20250069619A1 (en) | Method and apparatus for providing noise suppression to an intelligent personal assistant |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: AMAZON TECHNOLOGIES, INC., WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BANTA, WILLIAM CHRISTOPHER;REEL/FRAME:042213/0227 Effective date: 20160218 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |