EP1017252A2 - Hearing aid system - Google Patents

Hearing aid system Download PDF

Info

Publication number
EP1017252A2
EP1017252A2 EP99125134A EP99125134A EP1017252A2 EP 1017252 A2 EP1017252 A2 EP 1017252A2 EP 99125134 A EP99125134 A EP 99125134A EP 99125134 A EP99125134 A EP 99125134A EP 1017252 A2 EP1017252 A2 EP 1017252A2
Authority
EP
European Patent Office
Prior art keywords
hearing aid
aid system
earpiece
module
microphone
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP99125134A
Other languages
German (de)
French (fr)
Other versions
EP1017252A3 (en
Inventor
Thomas D. Bybee
Alexander L. Darbut
Mark S. Gorder
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Resistance Technology Inc
Original Assignee
Resistance Technology Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Resistance Technology Inc filed Critical Resistance Technology Inc
Publication of EP1017252A2 publication Critical patent/EP1017252A2/en
Publication of EP1017252A3 publication Critical patent/EP1017252A3/en
Withdrawn legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/40Arrangements for obtaining a desired directivity characteristic
    • H04R25/407Circuits for combining signals of a plurality of transducers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2225/00Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
    • H04R2225/41Detection or adaptation of hearing aid parameters or programs to listening situation, e.g. pub, forest
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2410/00Microphones
    • H04R2410/01Noise reduction using microphones having different directional characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R25/00Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
    • H04R25/55Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired
    • H04R25/554Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired using a wireless connection, e.g. between microphone and amplifier or using Tcoils

Definitions

  • the present invention relates to a hearing aid system, and more particularly to a personal hearing aid and information system that employs beamforming and other processing techniques to optimize the performance of the hearing aid system in a variety of sound environments.
  • Hearing aids have become a popular and effective tool for enabling a person with a hearing deficiency to detect and recognize audible sounds by filtering and amplifying the sounds and reproducing them in a microphone worn in the ear of the user.
  • Hearing aid devices have become more and more sophisticated, and binaural hearing aid systems have been developed so that a user wearing a hearing aid in each ear may identify the direction from which a sound originated based on differences in the time of arrival of a sound at each ear and differences in loudness level between the ears.
  • the present invention is a hearing aid system that includes at least one earpiece and a module in wireless communication with the earpiece.
  • a plurality of microphones are located in the module to receive incoming sound signals from a listening environment.
  • Processing circuitry in the module analyzes the incoming sounds signals received by the microphones and generates control signals that direct the earpiece to produce sounds representative of the listening environment.
  • the processing and analysis amplify sound signals originating from desired directions and attenuate sound signals originating from undesired directions, resulting in superior listening performance.
  • FIG. 1 is a diagram illustrating a human equipped with a hearing aid system 10 according to the present invention.
  • Hearing aid system 10 includes left earpiece 12, right earpiece 14 and medallion 16 worn round the user's neck or clipped onto the user's shirt, for example.
  • earpieces 12 and 14 and medallion 16 are equipped with microphones, wireless transceivers, processing circuitry, or a combination of these components, enabling hearing aid system 10 to provide performance above and beyond the performance of the individual earpieces or medallion.
  • FIG. 2 is a diagram illustrating the configuration of a first embodiment of hearing aid system 10a.
  • Left earpiece 12 includes FM receiver 22, and right earpiece 14 includes FM receiver 24.
  • Medallion 16 includes microphones 26, 28 and 30 and FM transmitter 32.
  • FM transmitter 32 is operable to wirelessly transmit signals and data to FM receiver 22 of left earpiece 12 via wireless link 24, and to FM receiver 24 of right earpiece 14 via wireless link 36.
  • microphones 26, 28 and 30 operate to receive incoming sound signals from the environment around the hearing aid system user.
  • Medallion 16 includes processing means (not shown in FIG. 2) to interpret the incoming sound signals received by microphones 26, 28 and 30 and generate appropriate control signals for reproducing sounds in the ears of the user at earpieces 12 and 14.
  • the control signals are transmitted from medallion 16 via FM transmitter 32 along wireless links 34 and 36 to FM receivers 22 and 24 in earpieces 12 and 14.
  • Earpieces 12 and 14 are operable in response to the control signals to reproduce the appropriate sounds in the ears of the user.
  • FIG. 3 is a diagram illustrating the configuration of a second embodiment of hearing aid system 10b.
  • Left earpiece 12 includes FM transceiver 22 and microphone 23.
  • Right earpiece 14 includes FM transceiver 24 and microphone 25.
  • Medallion 16 includes microphones 26, 28 and 30 and FM transceiver 32.
  • FM transceivers 22, 24 and 32 are operable to wirelessly transmit signals and data between left earpiece 12, right earpiece 14 and medallion 16 on wireless links 34 and 36.
  • microphones 23 and 25 in earpieces 12 and 14, respectively, and microphones 26, 28 and 30 in medallion 16 operate to receive incoming sound signals from the environment around the hearing aid system user.
  • Data representing the sound signals received by microphone 23 in left earpiece 12 and microphone 25 in right earpiece 14 is sent from FM transceiver 22 on wireless link 34 and from FM transceiver 24 on wireless link 36 to FM transceiver 32 in medallion 16.
  • Medallion 16 includes processing means (not shown in FIG. 3) to interpret the incoming sound signals received by microphones 26, 28 and 30 and the data signal received from earpieces 12 and 14 by FM transceiver 32.
  • the processing means generates appropriate control signals for reproducing sounds in the ears of the user at earpieces 12 and 14.
  • the control signals are transmitted from medallion 16 via FM transceiver 32 along wireless links 34 and 36 to FM transceivers 22 and 24 in earpieces 12 and 14.
  • Earpieces 12 and 14 are operable in response to the control signals to reproduce the appropriate sounds in the ears of the user.
  • FIG. 4 is a block diagram illustrating the functional components of medallion 16 utilized in hearing aid system 10 of the present invention.
  • medallion 16 houses the main processing circuitry of hearing aid system 10, in addition to an array of microphones and an FM transceiver.
  • Microphones 26, 28, 30 and 31 are spatially arranged on medallion 16 to optimize the ability of the microphones to distinguish the direction from which an incoming sound originates. The details of this spatial arrangement may vary; a simplistic example showing the effects of different spatial arrangements is discussed below with respect to FIGS. 6-13. Any number of microphones may be provided in medallion 16, as schematically demonstrated in FIG 4.
  • the signals received by microphones 26, 28, 30 and 31 are amplified by amplifiers 46, 48, 50 and 51, respectively.
  • Analog multiplexer and A/D converter 52 converts the analog signals from amplifiers 46, 48, 50, 51 into digital representations of those signals and sequences the signals into a digital data packet for input to digital signal processor (DSP) 54.
  • DSP digital signal processor
  • signals from remote microphones in earpieces worn by the user are transmitted on wireless links 34 and 36 to FM transceiver 32 in medallion 16. These signals are converted into digital form by FM radio receiver 58 and coupled as input to DSP 54.
  • DSP 54 implements a numerical algorithm to process the signals received from analog multiplexer and A/D converter 52 and FM radio receiver 58 in a predetermined manner. Given a desired processing scheme for the input signals, it is within the expertise of one skilled in the art to configure the DSP 54 to implement the appropriate numerical algorithm to achieve the desired processing scheme. The objectives of such processing schemes are described below with respect to FIGS. 7-13. Executing the algorithm produces control signals for generating appropriate signals in the earpieces worn by the user. The control signals are output from DSP 54 to FM radio transmitter 56 for wireless transmission via FM transceiver 32 on wireless links 34 and 36 to the respective earpieces.
  • FIG. 5 is a block diagram illustrating the functional components of earpiece 12 (or 14) utilized in hearing aid system 10 of the present invention.
  • Incoming sound signals are received by directional microphone 23a and omni-directional microphone 23b.
  • only one of directional microphone 23a and omni-directional microphone 23b is provided in earpiece 12.
  • the signals received by microphones 23a and 23b are amplified by amplifiers 66 and 68, respectively, which are in turn coupled to analog multiplexer and analog-to-digital (A/D) converter 70.
  • A/D analog-to-digital
  • Signals from remote microphones in a medallion worn by the user are transmitted on wireless link 34 to FM transceiver 22. These signals are converted into digital form by FM radio receiver 74 and coupled as input to DSP 72.
  • DSP 72 implements a numerical algorithm to process the signals received from analog multiplexer and A/D converter 70 and FM radio receiver 74 in a predetermined manner. Executing the algorithm produces control signals for generating appropriate signals in the earpieces worn by the user. The control signals are output from DSP 72 directly to receiver 78 in earpiece 12, and also to FM radio transmitter 76 for wireless transmission via FM transceiver 22 on wireless link 34 to the medallion worn by the user, where the signal will be relayed to the other earpiece worn by the user.
  • DSP 72 is shown as being located in earpiece 12.
  • DSP's may be employed in the earpieces only, with a medallion worn by the user serving as an additional location for microphones in the microphone array and as an FM relay station for signals to be transmitted between earpieces.
  • an additional DSP may also be located in the medallion, to distribute the processing functions between the physical components of the hearing aid system, or for redundancy purposes.
  • a DSP may only be located in the medallion (as shown in FIG. 4), with no DSP in the earpieces.
  • Other arrangements of functional components within the physical parts of the hearing aid system are contemplated by the present invention.
  • Beamforming is a technique that works by combining the signals received by an array of sensor elements, such as microphones, to generate a desired composite response.
  • a hearing aid system beamformer boosts the signal arriving from some directions and attenuates signals arriving from other directions, to enable the hearing aid system user to "focus" attention on a particular sound source.
  • the direction of maximum response referred to as the maximum response axis (MRA)
  • MRA maximum response axis
  • FIG. 6 depicts a simple example where multiple sensor elements are positioned in a straight line.
  • the angle perpendicular to a line that passes through the elements is defined as the "boresight" angle, or zero degrees.
  • the physical angle of incident sound is therefore defined within a range of -180 to +180 degrees, where 0 degrees is boresight, ⁇ 180 degrees is straight behind, +90 degrees is to the right, and -90 degrees is to the left.
  • This simple example is illustrated in FIG. 6, where microphones 80, 82 and 84 are linearly aligned at a distance d apart and an incoming sound signal having wavelength ⁇ creates an angle ⁇ with the microphone line.
  • the objective is to maximize the response of the system along the boresight axis and minimize the response in all other directions.
  • the results of this scheme are simplistically shown in FIGS. 7-9, where it is assumed that there are only two sensor elements for illustration purposes.
  • FIG. 7 illustrates that where the incoming sound signals are in phase with each other (that is, the sound signals reach the two sensors at the same time, meaning that their origination direction is "boresight"), and the signals received by the sensors are as illustrated by curves 90 and 92, adding the signals will result in curve 94 having the sum of the amplitudes of curves 90 and 92.
  • FIG. 7 illustrates that where the incoming sound signals are in phase with each other (that is, the sound signals reach the two sensors at the same time, meaning that their origination direction is "boresight"), and the signals received by the sensors are as illustrated by curves 90 and 92, adding the signals
  • FIG. 8 illustrates that where the incoming sound signals are 180 degrees out of phase (that is, the sound signals reach the sensors at times separated by a half cycle), and the signals received by the sensors are as illustrated by curves 100 and 102, adding the signals will result in a zero signal 104 - the signals completely cancel. This will occur when the sound signals are coming from an angle that is ⁇ 90 degrees from boresight.
  • FIG. 9 illustrates that where the incoming sound signals are somewhat out of phase (that is, the sound signals reach the sensors at times separated by less than a half cycle), and the signals received by the sensors are as illustrated by curves 110 and 112, adding the signals will result in curve 114 having an amplitude that is less than the sum of the amplitudes of curves 110 and 112.
  • the angle of arrival of the incoming sound signals can be calculated based on the difference in time of arrival, where the spacing between microphones is known.
  • the spacing of sensor elements is important to the ability to determine the origination direction of incoming sound signals.
  • the ideal spacing of sensor elements is a half wavelength, so that an incoming sound signal coming from 90 degrees to boresight reaches adjacent sensors at times separated by a half cycle and the signals completely cancel one another. With this spacing, the range of combined signal amplitudes extends from 0 (complete cancellation) to twice the amplitude of the incoming signal, for the two-sensor example, which is the maximum range.
  • microphones are spaced more than a half wavelength apart (typically for higher frequencies which have lower wavelengths), spatial aliasing occurs and grating lobes form in the response pattern, which makes accurate determination of the origination of the sound signal more difficult.
  • FIGS. 10-13 show the response of the two-microphone array over the full range of sound signal origination angles, with 0 degrees being boresight, -90 degrees being to the left, and +90 degrees being to the right.
  • FIG. 10 illustrates the scenario where the two microphones are omni-directional microphones spaced 5.25 inches apart and the frequency of incoming sound signals is 1275 Hz (at 1275 Hz, 5.25 inches is a half wavelength).
  • Curve 120 shows that the peak response of the system (adding the signals received at the two microphones) is at angles of 0 and ⁇ 180 degrees, with a deep "null" (attenuated response) at an angle of ⁇ 90 degrees.
  • FIG. 10 illustrates the scenario where the two microphones are omni-directional microphones spaced 5.25 inches apart and the frequency of incoming sound signals is 1275 Hz (at 1275 Hz, 5.25 inches is a half wavelength).
  • Curve 120 shows that the peak response of the system (adding the signals received at the two microphones) is at
  • FIG. 11 illustrates the same microphone scenario where the frequency of the incoming sound signals is 1000 Hz - at this frequency, using the same adding formula, curve 130 shows that the nulls at ⁇ 90 degrees are much less definitive.
  • FIG. 12 illustrates the same microphone scenario where the frequency of the incoming sound signals is 1500 Hz - at this frequency, using the same adding formula, curve 140 shows that side lobes are formed at ⁇ 90 degrees, with lower peak amplitudes, and the nulls are shifted to about ⁇ 60 degrees and ⁇ 120 degrees.
  • FIG. 13 illustrates an alternative scenario where the microphones utilized in the two-element array are directional microphones, spaced 5.25 inches apart, where the frequency of incoming sound signals is 1275 Hz.
  • Curve 150 shows that the peak response of the system is again at the boresight angle, with a lesser peak at ⁇ 180 degrees, and wide nulls in the vicinity of ⁇ 90 degrees.
  • the use of directional microphones can potentially allow the hearing aid user to more effectively focus on frontal sounds while attenuating sounds originating from the rear.
  • FIG. 14 is a diagram illustrating an exemplary beam pattern exhibited by microphone array 200 and an associated beamformer according to an embodiment of the present invention.
  • microphone array 200 exhibits a large frontal or main lobe 202, representing sounds coming from the front of a user.
  • Microphone array 200 also exhibits two side lobes 204 and 206 and a back lobe 208.
  • the beam former outputs are such that sounds in main lobe 202 are amplified the most, while sounds in side lobes 204 and 206 are attenuated somewhat and sounds in back lobe 208 are attenuated even more.
  • This beam pattern is designed to allow the user to effectively focus on sounds originating in front of the user (where the user is typically looking), even in an environment where random noise exists to the user's sides and rear. It should be understood that other beam patterns may be formed according to the present invention, and the pattern shown in FIG. 14 is an example of one particular beam pattern that may be implemented when desirable.
  • FIG. 15 is a diagram illustrating the exemplary beam pattern shown in FIG. 14 in relation to a first exemplary sound environment.
  • the sound environment includes multiple sources of random noise 210 (such as a crowd of people) at several points around the hearing aid user (wearing microphone array 200), and a keynote speaker 212 in front of the user.
  • keynote speaker 212 is located at the center of main lobe 202, and sounds originating from speaker 212 will therefore be maximally amplified.
  • Random noise 210 originates from locations at the sides of main lobe 202, and at side lobes 204 and 206 and back lobe 208, and therefore will be attenuated to a level below that of keynote speaker 212.
  • FIGS. 16A, 16B and 16C are graphs depicting an uncorrupted sound signal from keynote speaker 212 (FIG. 15), the sound signals received at the left and right input channels of the hearing aid system worn by the user, and the beamformer output from the hearing aid system to the ears of the user.
  • the uncorrupted signal is shown in FIG. 16A.
  • the sound signals received at the left and right input channels are shown in FIG. 16B.
  • the original input signal (shown in FIG. 16A) is received at each input channel, with a time-shift that is based on the physical distance between the left and right input channel microphones.
  • the left and right channel signals are also corrupted by some level of random noise that persists throughout the listening time frame.
  • the beamformer output is shown in FIG. 16C.
  • the beamformer recognizes the characteristics of the random noise and of the original input signal, and also recognizes, based on the time-shift of the input signal at the left and right channels, which direction the input signal originates from. Since the signal originates from the user's front (see FIG. 15), the signal is amplified to a higher level, while the random noise is attenuated to a lower level, thereby permitting the hearing aid user to better distinguish between the desired sound (keynote speaker 212, FIG. 15) and random noise (noise sources 210, FIG. 15).
  • FIG. 17 is a diagram illustrating the exemplary beam pattern shown in FIG. 14 in relation to a second exemplary sound environment.
  • the sound environment includes multiple sources of random noise 210 (such as a crowd of people) at several points around the hearing aid user (wearing microphone array 200), a keynote speaker 212 in front of the user, and a relatively loud directional noise source 214.
  • the sound environment may also include reverberation 216 of the desired noise signal from the walls of the room.
  • keynote speaker 212 is located at the center of main lobe 202, and sounds originating from speaker 212 will therefore be maximally amplified.
  • Random noise 210, directional noise 214 and reverberation 216 originate from locations at the sides of main lobe 202, and at side lobes 204 and 206 and back lobe 208, and therefore will be attenuated to a level below that of keynote speaker 212.
  • FIGS. 18A, 18B and 18C are graphs depicting an uncorrupted sound signal from keynote speaker 212 (FIG. 17), the sound signals received at the left and right input channels of the hearing aid system worn by the user, and the beamformer output from the hearing aid system to the ears of the user.
  • the uncorrupted signal is shown in FIG. 18A.
  • the sound signals received at the left and right input channels are shown in FIG. 18B.
  • the original input signal (shown in FIG. 18A) is received at each input channel, with a time-shift that is based on the physical distance between the left and right input channel microphones.
  • the left and right channel signals are also corrupted by some level of random noise that persists throughout the listening time frame, a higher level of directional noise, and reverberation of the input signal from the walls of the room.
  • the beamformer output is shown in FIG. 18C.
  • the beamformer recognizes the characteristics of the random noise, directional noise, reverberation, and the original input signal, and also recognizes, based on the time-shift of the input signal at the left and right channels, which direction the input signal originates from. Since the signal originates from the user's front (see FIG.
  • the signal is amplified to a higher level, while the random noise and the directional noise is attenuated to a lower level, thereby permitting the hearing aid user to better distinguish between the desired sound (keynote speaker 212, FIG. 17) and noise (noise sources 210, directional noise source 214, FIG. 17).
  • the hearing aid system of the present invention provides an extensive sensor array and processing capability that enables significant flexibility and control of the response of the system to particular sound patterns.
  • the shape of the response curves (shown simplistically in FIGS. 10-13 for a two-microphone example where the signals are merely added together) can be adjusted by performing different numerical algorithms on the signals received by the sensors. This adjustment can involve a technique called "adaptive null steering" which operates to move the nulls in the response pattern to a particular angular location, so that a particular noise source may be somewhat blocked out of the user's hearing, enabling better focus on the desired sound source.
  • the provision of wireless communication links between earpieces and a medallion worn around a user's neck allows a number of spacings between microphones in the array to be utilized to provide the optimum half wavelength spacing to accommodate a higher range of frequencies of incoming sound signals.
  • the hearing aid system therefore is able to provide superior performance compared to prior hearing aid devices, effectively enabling a listener to focus attention on a desired sound source in a variety of listening environments.

Abstract

A hearing aid system includes an earpiece for producing sound representative of a listening environment. A module worn by the hearing aid user is in wireless communication with the earpiece. A plurality of microphones are located in the module to receive incoming sound signals from the listening environment. Processing circuitry in the module analyzes the incoming sound signals received by the plurality of microphones and generates control signals to govern the production of sound by the earpiece.

Description

    BACKGROUND OF THE INVENTION
  • The present invention relates to a hearing aid system, and more particularly to a personal hearing aid and information system that employs beamforming and other processing techniques to optimize the performance of the hearing aid system in a variety of sound environments.
  • Hearing aids have become a popular and effective tool for enabling a person with a hearing deficiency to detect and recognize audible sounds by filtering and amplifying the sounds and reproducing them in a microphone worn in the ear of the user. Hearing aid devices have become more and more sophisticated, and binaural hearing aid systems have been developed so that a user wearing a hearing aid in each ear may identify the direction from which a sound originated based on differences in the time of arrival of a sound at each ear and differences in loudness level between the ears.
  • Despite the improvements in hearing aid technology and hearing aid system processing, existing hearing aid systems are still unable to approximate the selective listening patterns that are generated by the human brain in response to particular sound environments. Specifically, hearing aid systems are typically unable to effectively "focus" the attention of the listener onto a particular sound source where other sounds are also present. There is a continuing need in the art for a complete hearing aid system providing improved ability to discriminate among sound levels and sources.
  • BRIEF SUMMARY OF THE INVENTION
  • The present invention is a hearing aid system that includes at least one earpiece and a module in wireless communication with the earpiece. A plurality of microphones are located in the module to receive incoming sound signals from a listening environment. Processing circuitry in the module analyzes the incoming sounds signals received by the microphones and generates control signals that direct the earpiece to produce sounds representative of the listening environment. The processing and analysis amplify sound signals originating from desired directions and attenuate sound signals originating from undesired directions, resulting in superior listening performance.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a diagram illustrating a human outfitted with a hearing aid system according to the present invention
  • FIG. 2 is a diagram illustrating the components of a hearing aid system according to a first embodiment of the present invention.
  • FIG. 3 is a diagram illustrating the components of a hearing aid system according to a second embodiment of the present invention.
  • FIG. 4 is a block diagram illustrating the functional components of a medallion utilized in the hearing aid system of the present invention.
  • FIG. 5 is a block diagram illustrating the functional components of an earpiece utilized in the hearing aid system of the present invention.
  • FIG. 6 is a diagram illustrating a simplistic example of a multi-element sensor array for receiving incoming sound signals.
  • FIG. 7 is a stacked graph illustrating a beamforming example where the left and right sensor signals are in phase.
  • FIG. 8 is a stacked graph illustrating a beamforming example where the left and right sensor signals are 180 degrees out of phase.
  • FIG. 9 is a stacked graph illustrating a beamforming example where the left and right sensor signals are less than 180 degrees out of phase.
  • FIG. 10 is a graph illustrating the response of a hearing aid system employing two omni-directional microphones 5.25 inches apart where incoming sound signals have a frequency of 1275 Hz.
  • FIG. 11 is a graph illustrating the response of a hearing aid system employing two omni-directional microphones 5.25 inches apart where incoming sound signals have a frequency of 1000 Hz.
  • FIG. 12 is a graph illustrating the response of a hearing aid system employing two omni-directional microphones 5.25 inches apart where incoming sound signals have a frequency of 1500 Hz.
  • FIG. 13 is a graph illustrating the response of a hearing aid system employing two directional microphones 5.25 inches apart where incoming sound signals have a frequency of 1275 Hz.
  • FIG. 14 is a diagram illustrating an exemplary beam pattern exhibited by a microphone array and beamformer according to an embodiment of the present invention.
  • FIG. 15 is a diagram illustrating the exemplary beam pattern shown in FIG. 14 in relation to a first exemplary sound environment.
  • FIG. 16A is a graph illustrating an uncorrupted input signal from the speaker shown in FIG. 15 to be listened to with the hearing aid system of the present invention.
  • FIG. 16B is a graph illustrating the input signal of FIG. 16A corrupted with random noise.
  • FIG. 16C is a graph illustrating the beamformer output of the hearing aid system of the present invention.
  • FIG. 17 is a diagram illustrating the exemplary beam pattern shown in FIG. 14 in relation to a second exemplary sound environment.
  • FIG. 18A is a graph illustrating an uncorrupted input signal from the speaker shown in FIG. 17 to be listened to with the hearing aid system of the present invention.
  • FIG. 18B is a graph illustrating the input signal of FIG. 18A corrupted with directional noise, random noise and reverberation.
  • FIG. 18C is a graph illustrating the beamformer output of the hearing aid system of the present invention.
  • DETAILED DESCRIPTION
  • FIG. 1 is a diagram illustrating a human equipped with a hearing aid system 10 according to the present invention. Hearing aid system 10 includes left earpiece 12, right earpiece 14 and medallion 16 worn round the user's neck or clipped onto the user's shirt, for example. As will become apparent in the detailed description of hearing aid system 10 to follow, earpieces 12 and 14 and medallion 16 are equipped with microphones, wireless transceivers, processing circuitry, or a combination of these components, enabling hearing aid system 10 to provide performance above and beyond the performance of the individual earpieces or medallion.
  • FIG. 2 is a diagram illustrating the configuration of a first embodiment of hearing aid system 10a. Left earpiece 12 includes FM receiver 22, and right earpiece 14 includes FM receiver 24. Medallion 16 includes microphones 26, 28 and 30 and FM transmitter 32. FM transmitter 32 is operable to wirelessly transmit signals and data to FM receiver 22 of left earpiece 12 via wireless link 24, and to FM receiver 24 of right earpiece 14 via wireless link 36.
  • In an exemplary mode of operation, microphones 26, 28 and 30 operate to receive incoming sound signals from the environment around the hearing aid system user. Medallion 16 includes processing means (not shown in FIG. 2) to interpret the incoming sound signals received by microphones 26, 28 and 30 and generate appropriate control signals for reproducing sounds in the ears of the user at earpieces 12 and 14. The control signals are transmitted from medallion 16 via FM transmitter 32 along wireless links 34 and 36 to FM receivers 22 and 24 in earpieces 12 and 14. Earpieces 12 and 14 are operable in response to the control signals to reproduce the appropriate sounds in the ears of the user.
  • FIG. 3 is a diagram illustrating the configuration of a second embodiment of hearing aid system 10b. Left earpiece 12 includes FM transceiver 22 and microphone 23. Right earpiece 14 includes FM transceiver 24 and microphone 25. Medallion 16 includes microphones 26, 28 and 30 and FM transceiver 32. FM transceivers 22, 24 and 32 are operable to wirelessly transmit signals and data between left earpiece 12, right earpiece 14 and medallion 16 on wireless links 34 and 36.
  • In an exemplary mode of operation, microphones 23 and 25 in earpieces 12 and 14, respectively, and microphones 26, 28 and 30 in medallion 16 operate to receive incoming sound signals from the environment around the hearing aid system user. Data representing the sound signals received by microphone 23 in left earpiece 12 and microphone 25 in right earpiece 14 is sent from FM transceiver 22 on wireless link 34 and from FM transceiver 24 on wireless link 36 to FM transceiver 32 in medallion 16. Medallion 16 includes processing means (not shown in FIG. 3) to interpret the incoming sound signals received by microphones 26, 28 and 30 and the data signal received from earpieces 12 and 14 by FM transceiver 32. The processing means generates appropriate control signals for reproducing sounds in the ears of the user at earpieces 12 and 14. The control signals are transmitted from medallion 16 via FM transceiver 32 along wireless links 34 and 36 to FM transceivers 22 and 24 in earpieces 12 and 14. Earpieces 12 and 14 are operable in response to the control signals to reproduce the appropriate sounds in the ears of the user.
  • FIG. 4 is a block diagram illustrating the functional components of medallion 16 utilized in hearing aid system 10 of the present invention. In an exemplary embodiment, medallion 16 houses the main processing circuitry of hearing aid system 10, in addition to an array of microphones and an FM transceiver. Microphones 26, 28, 30 and 31 are spatially arranged on medallion 16 to optimize the ability of the microphones to distinguish the direction from which an incoming sound originates. The details of this spatial arrangement may vary; a simplistic example showing the effects of different spatial arrangements is discussed below with respect to FIGS. 6-13. Any number of microphones may be provided in medallion 16, as schematically demonstrated in FIG 4. The signals received by microphones 26, 28, 30 and 31 are amplified by amplifiers 46, 48, 50 and 51, respectively. The outputs of amplifiers 46, 48, 50 and 51 are coupled to analog multiplexer and analog-to-digital (A/D) converter 52. Analog multiplexer and A/D converter 52 converts the analog signals from amplifiers 46, 48, 50, 51 into digital representations of those signals and sequences the signals into a digital data packet for input to digital signal processor (DSP) 54. In an embodiment such as hearing aid system 10b shown in FIG. 3, signals from remote microphones in earpieces worn by the user are transmitted on wireless links 34 and 36 to FM transceiver 32 in medallion 16. These signals are converted into digital form by FM radio receiver 58 and coupled as input to DSP 54.
  • DSP 54 implements a numerical algorithm to process the signals received from analog multiplexer and A/D converter 52 and FM radio receiver 58 in a predetermined manner. Given a desired processing scheme for the input signals, it is within the expertise of one skilled in the art to configure the DSP 54 to implement the appropriate numerical algorithm to achieve the desired processing scheme. The objectives of such processing schemes are described below with respect to FIGS. 7-13. Executing the algorithm produces control signals for generating appropriate signals in the earpieces worn by the user. The control signals are output from DSP 54 to FM radio transmitter 56 for wireless transmission via FM transceiver 32 on wireless links 34 and 36 to the respective earpieces.
  • FIG. 5 is a block diagram illustrating the functional components of earpiece 12 (or 14) utilized in hearing aid system 10 of the present invention. Incoming sound signals are received by directional microphone 23a and omni-directional microphone 23b. In an alternative embodiment, only one of directional microphone 23a and omni-directional microphone 23b is provided in earpiece 12. The signals received by microphones 23a and 23b are amplified by amplifiers 66 and 68, respectively, which are in turn coupled to analog multiplexer and analog-to-digital (A/D) converter 70. Signals from remote microphones in a medallion worn by the user are transmitted on wireless link 34 to FM transceiver 22. These signals are converted into digital form by FM radio receiver 74 and coupled as input to DSP 72.
  • As with DSP 54 shown in FIG. 4, DSP 72 implements a numerical algorithm to process the signals received from analog multiplexer and A/D converter 70 and FM radio receiver 74 in a predetermined manner. Executing the algorithm produces control signals for generating appropriate signals in the earpieces worn by the user. The control signals are output from DSP 72 directly to receiver 78 in earpiece 12, and also to FM radio transmitter 76 for wireless transmission via FM transceiver 22 on wireless link 34 to the medallion worn by the user, where the signal will be relayed to the other earpiece worn by the user.
  • In the embodiment schematically pictured in FIG. 5, DSP 72 is shown as being located in earpiece 12. In this embodiment, DSP's may be employed in the earpieces only, with a medallion worn by the user serving as an additional location for microphones in the microphone array and as an FM relay station for signals to be transmitted between earpieces. Alternatively, an additional DSP may also be located in the medallion, to distribute the processing functions between the physical components of the hearing aid system, or for redundancy purposes. In another embodiment, a DSP may only be located in the medallion (as shown in FIG. 4), with no DSP in the earpieces. Other arrangements of functional components within the physical parts of the hearing aid system are contemplated by the present invention.
  • With the components provided as shown in FIGS. 1-5, the hearing aid system of the present invention is able to achieve high performance operation and listening features that were not possible with prior art hearing aids. To understand the performance advantages enabled by the present invention, a simplified description of a multi-sensor processing technique called beamforming is beneficial. Beamforming is a technique that works by combining the signals received by an array of sensor elements, such as microphones, to generate a desired composite response. For example, a hearing aid system beamformer boosts the signal arriving from some directions and attenuates signals arriving from other directions, to enable the hearing aid system user to "focus" attention on a particular sound source. The direction of maximum response, referred to as the maximum response axis (MRA), can be steered in different directions by properly combining the signals from each sensor element.
  • FIG. 6 depicts a simple example where multiple sensor elements are positioned in a straight line. The angle perpendicular to a line that passes through the elements is defined as the "boresight" angle, or zero degrees. The physical angle of incident sound is therefore defined within a range of -180 to +180 degrees, where 0 degrees is boresight, ±180 degrees is straight behind, +90 degrees is to the right, and -90 degrees is to the left. This simple example is illustrated in FIG. 6, where microphones 80, 82 and 84 are linearly aligned at a distance d apart and an incoming sound signal having wavelength λ creates an angle  with the microphone line. The difference in time of the sound signals received by adjacent microphones in the array is defined by: x = 360*d*sinλ degrees of phase or d*sinc seconds, where c is the speed of the waves. This difference in time may be utilized to determine the direction from which sound signals are originating.
  • In one scenario, the objective is to maximize the response of the system along the boresight axis and minimize the response in all other directions. This is a simple processing scenario - the signals from the sensors need only be added together with equal weight given to the signals from each sensor. The results of this scheme are simplistically shown in FIGS. 7-9, where it is assumed that there are only two sensor elements for illustration purposes. FIG. 7 illustrates that where the incoming sound signals are in phase with each other (that is, the sound signals reach the two sensors at the same time, meaning that their origination direction is "boresight"), and the signals received by the sensors are as illustrated by curves 90 and 92, adding the signals will result in curve 94 having the sum of the amplitudes of curves 90 and 92. FIG. 8 illustrates that where the incoming sound signals are 180 degrees out of phase (that is, the sound signals reach the sensors at times separated by a half cycle), and the signals received by the sensors are as illustrated by curves 100 and 102, adding the signals will result in a zero signal 104 - the signals completely cancel. This will occur when the sound signals are coming from an angle that is ±90 degrees from boresight. FIG. 9 illustrates that where the incoming sound signals are somewhat out of phase (that is, the sound signals reach the sensors at times separated by less than a half cycle), and the signals received by the sensors are as illustrated by curves 110 and 112, adding the signals will result in curve 114 having an amplitude that is less than the sum of the amplitudes of curves 110 and 112. The angle of arrival of the incoming sound signals can be calculated based on the difference in time of arrival, where the spacing between microphones is known.
  • It can be appreciated from the discussion above that the spacing of sensor elements is important to the ability to determine the origination direction of incoming sound signals. The ideal spacing of sensor elements is a half wavelength, so that an incoming sound signal coming from 90 degrees to boresight reaches adjacent sensors at times separated by a half cycle and the signals completely cancel one another. With this spacing, the range of combined signal amplitudes extends from 0 (complete cancellation) to twice the amplitude of the incoming signal, for the two-sensor example, which is the maximum range. Where microphones are spaced more than a half wavelength apart (typically for higher frequencies which have lower wavelengths), spatial aliasing occurs and grating lobes form in the response pattern, which makes accurate determination of the origination of the sound signal more difficult.
  • FIGS. 10-13 show the response of the two-microphone array over the full range of sound signal origination angles, with 0 degrees being boresight, -90 degrees being to the left, and +90 degrees being to the right. FIG. 10 illustrates the scenario where the two microphones are omni-directional microphones spaced 5.25 inches apart and the frequency of incoming sound signals is 1275 Hz (at 1275 Hz, 5.25 inches is a half wavelength). Curve 120 shows that the peak response of the system (adding the signals received at the two microphones) is at angles of 0 and ±180 degrees, with a deep "null" (attenuated response) at an angle of ±90 degrees. FIG. 11 illustrates the same microphone scenario where the frequency of the incoming sound signals is 1000 Hz - at this frequency, using the same adding formula, curve 130 shows that the nulls at ±90 degrees are much less definitive. FIG. 12 illustrates the same microphone scenario where the frequency of the incoming sound signals is 1500 Hz - at this frequency, using the same adding formula, curve 140 shows that side lobes are formed at ±90 degrees, with lower peak amplitudes, and the nulls are shifted to about ±60 degrees and ±120 degrees. These curves illustrate the importance of having microphone elements that are a half wavelength apart; where the microphone spacing is more or less than a half wavelength, the response curves are not as well defined, which tends to result in less ability for the user of a hearing aid system to focus on particular sounds.
  • FIG. 13 illustrates an alternative scenario where the microphones utilized in the two-element array are directional microphones, spaced 5.25 inches apart, where the frequency of incoming sound signals is 1275 Hz. Curve 150 shows that the peak response of the system is again at the boresight angle, with a lesser peak at ±180 degrees, and wide nulls in the vicinity of ±90 degrees. The use of directional microphones can potentially allow the hearing aid user to more effectively focus on frontal sounds while attenuating sounds originating from the rear.
  • FIG. 14 is a diagram illustrating an exemplary beam pattern exhibited by microphone array 200 and an associated beamformer according to an embodiment of the present invention. In the embodiment depicted in FIG. 14, microphone array 200 exhibits a large frontal or main lobe 202, representing sounds coming from the front of a user. Microphone array 200 also exhibits two side lobes 204 and 206 and a back lobe 208. The beam former outputs are such that sounds in main lobe 202 are amplified the most, while sounds in side lobes 204 and 206 are attenuated somewhat and sounds in back lobe 208 are attenuated even more. This beam pattern is designed to allow the user to effectively focus on sounds originating in front of the user (where the user is typically looking), even in an environment where random noise exists to the user's sides and rear. It should be understood that other beam patterns may be formed according to the present invention, and the pattern shown in FIG. 14 is an example of one particular beam pattern that may be implemented when desirable.
  • FIG. 15 is a diagram illustrating the exemplary beam pattern shown in FIG. 14 in relation to a first exemplary sound environment. The sound environment includes multiple sources of random noise 210 (such as a crowd of people) at several points around the hearing aid user (wearing microphone array 200), and a keynote speaker 212 in front of the user. As can be seen in FIG. 15, keynote speaker 212 is located at the center of main lobe 202, and sounds originating from speaker 212 will therefore be maximally amplified. Random noise 210 originates from locations at the sides of main lobe 202, and at side lobes 204 and 206 and back lobe 208, and therefore will be attenuated to a level below that of keynote speaker 212.
  • FIGS. 16A, 16B and 16C are graphs depicting an uncorrupted sound signal from keynote speaker 212 (FIG. 15), the sound signals received at the left and right input channels of the hearing aid system worn by the user, and the beamformer output from the hearing aid system to the ears of the user. The uncorrupted signal is shown in FIG. 16A. The sound signals received at the left and right input channels are shown in FIG. 16B. The original input signal (shown in FIG. 16A) is received at each input channel, with a time-shift that is based on the physical distance between the left and right input channel microphones. The left and right channel signals are also corrupted by some level of random noise that persists throughout the listening time frame. The beamformer output is shown in FIG. 16C. The beamformer recognizes the characteristics of the random noise and of the original input signal, and also recognizes, based on the time-shift of the input signal at the left and right channels, which direction the input signal originates from. Since the signal originates from the user's front (see FIG. 15), the signal is amplified to a higher level, while the random noise is attenuated to a lower level, thereby permitting the hearing aid user to better distinguish between the desired sound (keynote speaker 212, FIG. 15) and random noise (noise sources 210, FIG. 15).
  • FIG. 17 is a diagram illustrating the exemplary beam pattern shown in FIG. 14 in relation to a second exemplary sound environment. The sound environment includes multiple sources of random noise 210 (such as a crowd of people) at several points around the hearing aid user (wearing microphone array 200), a keynote speaker 212 in front of the user, and a relatively loud directional noise source 214. The sound environment may also include reverberation 216 of the desired noise signal from the walls of the room. As can be seen in FIG. 15, keynote speaker 212 is located at the center of main lobe 202, and sounds originating from speaker 212 will therefore be maximally amplified. Random noise 210, directional noise 214 and reverberation 216 originate from locations at the sides of main lobe 202, and at side lobes 204 and 206 and back lobe 208, and therefore will be attenuated to a level below that of keynote speaker 212.
  • FIGS. 18A, 18B and 18C are graphs depicting an uncorrupted sound signal from keynote speaker 212 (FIG. 17), the sound signals received at the left and right input channels of the hearing aid system worn by the user, and the beamformer output from the hearing aid system to the ears of the user. The uncorrupted signal is shown in FIG. 18A. The sound signals received at the left and right input channels are shown in FIG. 18B. The original input signal (shown in FIG. 18A) is received at each input channel, with a time-shift that is based on the physical distance between the left and right input channel microphones. The left and right channel signals are also corrupted by some level of random noise that persists throughout the listening time frame, a higher level of directional noise, and reverberation of the input signal from the walls of the room. The beamformer output is shown in FIG. 18C. The beamformer recognizes the characteristics of the random noise, directional noise, reverberation, and the original input signal, and also recognizes, based on the time-shift of the input signal at the left and right channels, which direction the input signal originates from. Since the signal originates from the user's front (see FIG. 17), the signal is amplified to a higher level, while the random noise and the directional noise is attenuated to a lower level, thereby permitting the hearing aid user to better distinguish between the desired sound (keynote speaker 212, FIG. 17) and noise (noise sources 210, directional noise source 214, FIG. 17).
  • The hearing aid system of the present invention, as shown in FIGS. 1-5, provides an extensive sensor array and processing capability that enables significant flexibility and control of the response of the system to particular sound patterns. The shape of the response curves (shown simplistically in FIGS. 10-13 for a two-microphone example where the signals are merely added together) can be adjusted by performing different numerical algorithms on the signals received by the sensors. This adjustment can involve a technique called "adaptive null steering" which operates to move the nulls in the response pattern to a particular angular location, so that a particular noise source may be somewhat blocked out of the user's hearing, enabling better focus on the desired sound source. The provision of wireless communication links between earpieces and a medallion worn around a user's neck allows a number of spacings between microphones in the array to be utilized to provide the optimum half wavelength spacing to accommodate a higher range of frequencies of incoming sound signals. The hearing aid system therefore is able to provide superior performance compared to prior hearing aid devices, effectively enabling a listener to focus attention on a desired sound source in a variety of listening environments.
  • Although the present invention has been described with reference to preferred embodiments, workers skilled in the art will recognize that changes may be made in form and detail without departing from the spirit and scope of the invention.

Claims (19)

  1. A hearing aid system comprising:
    an earpiece for producing sound representative of a listening environment;
    a module in wireless communication with the earpiece;
    a plurality of microphones located in the module to receive incoming sound signals from the listening environment;
    processing circuitry in the module for analyzing the incoming sound signals received by the plurality of microphones and generating control signals to govern the production of sound by the earpiece.
  2. The hearing aid system of claim 1, further comprising:
    at least one microphone located in the earpiece to receive incoming sound signals from the listening environment.
  3. The hearing aid system of claim 2, wherein the at least one microphone located in the earpiece comprises an omni-directional microphone.
  4. The hearing aid system of claim 2, wherein the at least one microphone located in the earpiece comprises a directional microphone.
  5. The hearing aid system of claim 2, wherein the at least one microphone located in the earpiece comprises an omni-directional microphone and a directional microphone.
  6. The hearing aid system of claim 2, wherein the earpiece and the module each include a wireless transceiver for effecting communication between the earpiece and the module.
  7. The hearing aid system of claim 1, wherein the earpiece includes a wireless receiver and the module includes a wireless transmitter for transmitting information from the module to the earpiece.
  8. The hearing aid system of claim 1, further comprising a second earpiece for producing sound representative of the listening environment, the module being in wireless communication wit the second earpiece.
  9. The hearing aid system of claim 8, further comprising:
    at least one microphone located in the first-named and second earpieces to receive incoming sound signals from the listening environment.
  10. The hearing aid system of claim 9, wherein the first-named and second earpieces and the module each include a wireless transceiver for effecting communication between the earpieces and the module.
  11. The hearing aid system of claim 9, wherein the at least one microphone in the first-named and second earpieces comprises an omni-directional microphone.
  12. The hearing aid system of claim 9, wherein the at least one microphone in the first-named and second earpieces comprises a directional microphone.
  13. The hearing aid system of claim 9, wherein the at least one microphone in the first-named and second earpieces comprises an omni-directional microphone and a directional microphone.
  14. The hearing aid system of claim 8, wherein the first-named and second earpieces each include a wireless receiver and the module includes a wireless transmitter for transmitting information from the module to the earpieces.
  15. The hearing aid system of claim 1, wherein the processing circuitry comprises a digital signal processor for performing a numerical algorithm on the incoming sound signals received by the microphones to shape a response of the hearing aid system that is represented by the sound produced by the earpiece.
  16. A method of producing sounds in the ears of a user that represent a listening environment, the method comprising:
    providing a plurality of microphones in a module worn by the user to receive incoming sound signals from the listening environment;
    processing the incoming sound signals received by the microphones to amplify sound signals originating from desired directions and to attenuate sound signals originating from undesired directions;
    generating control signals to govern the production of sound in the user's ears according to the result of the processing step; and
    communicating the control signals to earpieces located in the user's ears to govern the production of sound therein.
  17. The method of claim 16, further comprising:
    providing at least one microphone in the earpieces located in the user's ears to receive incoming sound signals from the listening environment.
  18. The method of claim 16, wherein the processing step comprises:
    performing a numerical algorithm on the incoming sound signals received by the microphones to modify the signals according to a predetermined scheme for locating peaks and nulls in the response of the hearing aid system.
  19. The method of claim 16, wherein the communicating step comprises wirelessly transmitting the control signals from the module to the earpieces.
EP99125134A 1998-12-31 1999-12-16 Hearing aid system Withdrawn EP1017252A3 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US11451398P 1998-12-31 1998-12-31
US114513P 1998-12-31

Publications (2)

Publication Number Publication Date
EP1017252A2 true EP1017252A2 (en) 2000-07-05
EP1017252A3 EP1017252A3 (en) 2006-05-31

Family

ID=22355676

Family Applications (1)

Application Number Title Priority Date Filing Date
EP99125134A Withdrawn EP1017252A3 (en) 1998-12-31 1999-12-16 Hearing aid system

Country Status (1)

Country Link
EP (1) EP1017252A3 (en)

Cited By (125)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002028143A2 (en) * 2000-09-29 2002-04-04 Siemens Audiologische Technik Gmbh Method for operating a hearing aid system and hearing aid system
NL1021485C2 (en) * 2002-09-18 2004-03-22 Stichting Tech Wetenschapp Hearing glasses assembly.
EP1683392A2 (en) * 2003-11-12 2006-07-26 Oticon A/S Microphone system
WO2008107359A1 (en) * 2007-03-05 2008-09-12 Siemens Audiologische Technik Gmbh Hearing system with distributed signal processing and corresponding method
EP2568719A1 (en) * 2011-09-07 2013-03-13 ITI Infotech Industries Limited Wearable sound amplification apparatus for the hearing impaired
CN101778330B (en) * 2009-12-30 2013-04-03 南京信息工程大学 Mobile phone platform-based array microphone hearing aid and control method thereof
US8527280B2 (en) * 2001-12-13 2013-09-03 Peter V. Boesen Voice communication device with foreign language translation
US8641596B2 (en) 2008-07-02 2014-02-04 Cochlear Limited Wireless communication in a multimodal auditory prosthesis
CN103873972A (en) * 2012-12-12 2014-06-18 森海塞尔通信公司 Microphone boom
US20140185845A1 (en) * 2012-12-28 2014-07-03 Gn Resound A/S Spectacle hearing device system
EP2840807A1 (en) * 2013-08-19 2015-02-25 Oticon A/s External microphone array and hearing aid using it
US9755704B2 (en) 2015-08-29 2017-09-05 Bragi GmbH Multimodal communication system induction and radio and method
US9800966B2 (en) 2015-08-29 2017-10-24 Bragi GmbH Smart case power utilization control system and method
US9813826B2 (en) 2015-08-29 2017-11-07 Bragi GmbH Earpiece with electronic environmental sound pass-through system
USD805060S1 (en) 2016-04-07 2017-12-12 Bragi GmbH Earphone
US9843853B2 (en) 2015-08-29 2017-12-12 Bragi GmbH Power control for battery powered personal area network device system and method
US9854372B2 (en) 2015-08-29 2017-12-26 Bragi GmbH Production line PCB serial programming and testing method and system
US9866941B2 (en) 2015-10-20 2018-01-09 Bragi GmbH Multi-point multiple sensor array for data sensing and processing system and method
US9866282B2 (en) 2015-08-29 2018-01-09 Bragi GmbH Magnetic induction antenna for use in a wearable device
US9905088B2 (en) 2015-08-29 2018-02-27 Bragi GmbH Responsive visual communication system and method
US9939891B2 (en) 2015-12-21 2018-04-10 Bragi GmbH Voice dictation systems using earpiece microphone system and method
US9949008B2 (en) 2015-08-29 2018-04-17 Bragi GmbH Reproduction of ambient environmental sound for acoustic transparency of ear canal device system and method
US9944295B2 (en) 2015-11-27 2018-04-17 Bragi GmbH Vehicle with wearable for identifying role of one or more users and adjustment of user settings
US9949013B2 (en) 2015-08-29 2018-04-17 Bragi GmbH Near field gesture control system and method
US9972895B2 (en) 2015-08-29 2018-05-15 Bragi GmbH Antenna for use in a wearable device
US9980033B2 (en) 2015-12-21 2018-05-22 Bragi GmbH Microphone natural speech capture voice dictation system and method
US9980189B2 (en) 2015-10-20 2018-05-22 Bragi GmbH Diversity bluetooth system and method
US9978278B2 (en) 2015-11-27 2018-05-22 Bragi GmbH Vehicle to vehicle communications using ear pieces
USD819438S1 (en) 2016-04-07 2018-06-05 Bragi GmbH Package
US10013542B2 (en) 2016-04-28 2018-07-03 Bragi GmbH Biometric interface system and method
US10015579B2 (en) 2016-04-08 2018-07-03 Bragi GmbH Audio accelerometric feedback through bilateral ear worn device system and method
USD821970S1 (en) 2016-04-07 2018-07-03 Bragi GmbH Wearable device charger
USD823835S1 (en) 2016-04-07 2018-07-24 Bragi GmbH Earphone
USD824371S1 (en) 2016-05-06 2018-07-31 Bragi GmbH Headphone
US10045112B2 (en) 2016-11-04 2018-08-07 Bragi GmbH Earpiece with added ambient environment
US10045116B2 (en) 2016-03-14 2018-08-07 Bragi GmbH Explosive sound pressure level active noise cancellation utilizing completely wireless earpieces system and method
US10045117B2 (en) 2016-11-04 2018-08-07 Bragi GmbH Earpiece with modified ambient environment over-ride function
US10040423B2 (en) 2015-11-27 2018-08-07 Bragi GmbH Vehicle with wearable for identifying one or more vehicle occupants
US10045110B2 (en) 2016-07-06 2018-08-07 Bragi GmbH Selective sound field environment processing system and method
US10049184B2 (en) 2016-10-07 2018-08-14 Bragi GmbH Software application transmission via body interface using a wearable device in conjunction with removable body sensor arrays system and method
US10045736B2 (en) 2016-07-06 2018-08-14 Bragi GmbH Detection of metabolic disorders using wireless earpieces
US10052065B2 (en) 2016-03-23 2018-08-21 Bragi GmbH Earpiece life monitor with capability of automatic notification system and method
US10063957B2 (en) 2016-11-04 2018-08-28 Bragi GmbH Earpiece with source selection within ambient environment
US10062373B2 (en) 2016-11-03 2018-08-28 Bragi GmbH Selective audio isolation from body generated sound system and method
US10058282B2 (en) 2016-11-04 2018-08-28 Bragi GmbH Manual operation assistance with earpiece with 3D sound cues
US10085082B2 (en) 2016-03-11 2018-09-25 Bragi GmbH Earpiece with GPS receiver
US10085091B2 (en) 2016-02-09 2018-09-25 Bragi GmbH Ambient volume modification through environmental microphone feedback loop system and method
US10099636B2 (en) 2015-11-27 2018-10-16 Bragi GmbH System and method for determining a user role and user settings associated with a vehicle
US10104486B2 (en) 2016-01-25 2018-10-16 Bragi GmbH In-ear sensor calibration and detecting system and method
US10104458B2 (en) 2015-10-20 2018-10-16 Bragi GmbH Enhanced biometric control systems for detection of emergency events system and method
US10099374B2 (en) 2015-12-01 2018-10-16 Bragi GmbH Robotic safety using wearables
US10104460B2 (en) 2015-11-27 2018-10-16 Bragi GmbH Vehicle with interaction between entertainment systems and wearable devices
US10117604B2 (en) 2016-11-02 2018-11-06 Bragi GmbH 3D sound positioning with distributed sensors
US10122421B2 (en) 2015-08-29 2018-11-06 Bragi GmbH Multimodal communication system using induction and radio and method
US10129620B2 (en) 2016-01-25 2018-11-13 Bragi GmbH Multilayer approach to hydrophobic and oleophobic system and method
US10154332B2 (en) 2015-12-29 2018-12-11 Bragi GmbH Power management for wireless earpieces utilizing sensor measurements
USD836089S1 (en) 2016-05-06 2018-12-18 Bragi GmbH Headphone
US10158934B2 (en) 2016-07-07 2018-12-18 Bragi GmbH Case for multiple earpiece pairs
US10165350B2 (en) 2016-07-07 2018-12-25 Bragi GmbH Earpiece with app environment
US10175753B2 (en) 2015-10-20 2019-01-08 Bragi GmbH Second screen devices utilizing data from ear worn device system and method
US10194232B2 (en) 2015-08-29 2019-01-29 Bragi GmbH Responsive packaging system for managing display actions
US10194228B2 (en) 2015-08-29 2019-01-29 Bragi GmbH Load balancing to maximize device function in a personal area network device system and method
US10200790B2 (en) 2016-01-15 2019-02-05 Bragi GmbH Earpiece with cellular connectivity
US10203773B2 (en) 2015-08-29 2019-02-12 Bragi GmbH Interactive product packaging system and method
US10206052B2 (en) 2015-12-22 2019-02-12 Bragi GmbH Analytical determination of remote battery temperature through distributed sensor array system and method
US10205814B2 (en) 2016-11-03 2019-02-12 Bragi GmbH Wireless earpiece with walkie-talkie functionality
US10206042B2 (en) 2015-10-20 2019-02-12 Bragi GmbH 3D sound field using bilateral earpieces system and method
US10216474B2 (en) 2016-07-06 2019-02-26 Bragi GmbH Variable computing engine for interactive media based upon user biometrics
US10225638B2 (en) 2016-11-03 2019-03-05 Bragi GmbH Ear piece with pseudolite connectivity
US10234133B2 (en) 2015-08-29 2019-03-19 Bragi GmbH System and method for prevention of LED light spillage
US10313779B2 (en) 2016-08-26 2019-06-04 Bragi GmbH Voice assistant system for wireless earpieces
US10327082B2 (en) 2016-03-02 2019-06-18 Bragi GmbH Location based tracking using a wireless earpiece device, system, and method
US10334346B2 (en) 2016-03-24 2019-06-25 Bragi GmbH Real-time multivariable biometric analysis and display system and method
US10334345B2 (en) 2015-12-29 2019-06-25 Bragi GmbH Notification and activation system utilizing onboard sensors of wireless earpieces
US10344960B2 (en) 2017-09-19 2019-07-09 Bragi GmbH Wireless earpiece controlled medical headlight
US10342428B2 (en) 2015-10-20 2019-07-09 Bragi GmbH Monitoring pulse transmissions using radar
US10397686B2 (en) 2016-08-15 2019-08-27 Bragi GmbH Detection of movement adjacent an earpiece device
US10405081B2 (en) 2017-02-08 2019-09-03 Bragi GmbH Intelligent wireless headset system
US10409394B2 (en) 2015-08-29 2019-09-10 Bragi GmbH Gesture based control system based upon device orientation system and method
US10409091B2 (en) 2016-08-25 2019-09-10 Bragi GmbH Wearable with lenses
US10453450B2 (en) 2015-10-20 2019-10-22 Bragi GmbH Wearable earpiece voice command control system and method
US10455313B2 (en) 2016-10-31 2019-10-22 Bragi GmbH Wireless earpiece with force feedback
US10460095B2 (en) 2016-09-30 2019-10-29 Bragi GmbH Earpiece with biometric identifiers
US10469931B2 (en) 2016-07-07 2019-11-05 Bragi GmbH Comparative analysis of sensors to control power status for wireless earpieces
US10506322B2 (en) 2015-10-20 2019-12-10 Bragi GmbH Wearable device onboard applications system and method
US10506327B2 (en) 2016-12-27 2019-12-10 Bragi GmbH Ambient environmental sound field manipulation based on user defined voice and audio recognition pattern analysis system and method
US10542340B2 (en) 2015-11-30 2020-01-21 Bragi GmbH Power management for wireless earpieces
US10555700B2 (en) 2016-07-06 2020-02-11 Bragi GmbH Combined optical sensor for audio and pulse oximetry system and method
US10575086B2 (en) 2017-03-22 2020-02-25 Bragi GmbH System and method for sharing wireless earpieces
US10575083B2 (en) 2015-12-22 2020-02-25 Bragi GmbH Near field based earpiece data transfer system and method
US10580282B2 (en) 2016-09-12 2020-03-03 Bragi GmbH Ear based contextual environment and biometric pattern recognition system and method
US10582290B2 (en) 2017-02-21 2020-03-03 Bragi GmbH Earpiece with tap functionality
US10582328B2 (en) 2016-07-06 2020-03-03 Bragi GmbH Audio response based on user worn microphones to direct or adapt program responses system and method
US10587943B2 (en) 2016-07-09 2020-03-10 Bragi GmbH Earpiece with wirelessly recharging battery
US10598506B2 (en) 2016-09-12 2020-03-24 Bragi GmbH Audio navigation using short range bilateral earpieces
US10621583B2 (en) 2016-07-07 2020-04-14 Bragi GmbH Wearable earpiece multifactorial biometric analysis system and method
US10617297B2 (en) 2016-11-02 2020-04-14 Bragi GmbH Earpiece with in-ear electrodes
US10635385B2 (en) 2015-11-13 2020-04-28 Bragi GmbH Method and apparatus for interfacing with wireless earpieces
US10667033B2 (en) 2016-03-02 2020-05-26 Bragi GmbH Multifactorial unlocking function for smart wearable device and method
CN111327984A (en) * 2020-02-27 2020-06-23 北京声加科技有限公司 Earphone auxiliary listening method based on null filtering and ear-worn equipment
US10698983B2 (en) 2016-10-31 2020-06-30 Bragi GmbH Wireless earpiece with a medical engine
US10708699B2 (en) 2017-05-03 2020-07-07 Bragi GmbH Hearing aid with added functionality
US10747337B2 (en) 2016-04-26 2020-08-18 Bragi GmbH Mechanical detection of a touch movement using a sensor and a special surface pattern system and method
US10771881B2 (en) 2017-02-27 2020-09-08 Bragi GmbH Earpiece with audio 3D menu
US10771877B2 (en) 2016-10-31 2020-09-08 Bragi GmbH Dual earpieces for same ear
US10821361B2 (en) 2016-11-03 2020-11-03 Bragi GmbH Gaming with earpiece 3D audio
US10852829B2 (en) 2016-09-13 2020-12-01 Bragi GmbH Measurement of facial muscle EMG potentials for predictive analysis using a smart wearable system and method
US10856809B2 (en) 2016-03-24 2020-12-08 Bragi GmbH Earpiece with glucose sensor and system
US10887679B2 (en) 2016-08-26 2021-01-05 Bragi GmbH Earpiece for audiograms
US10888039B2 (en) 2016-07-06 2021-01-05 Bragi GmbH Shielded case for wireless earpieces
US10942701B2 (en) 2016-10-31 2021-03-09 Bragi GmbH Input and edit functions utilizing accelerometer based earpiece movement system and method
US10977348B2 (en) 2016-08-24 2021-04-13 Bragi GmbH Digital signature using phonometry and compiled biometric data system and method
US11013445B2 (en) 2017-06-08 2021-05-25 Bragi GmbH Wireless earpiece with transcranial stimulation
CN112911480A (en) * 2021-01-22 2021-06-04 成都市舒听医疗器械有限责任公司 Hearing aid sound amplification method and hearing aid
US11086593B2 (en) 2016-08-26 2021-08-10 Bragi GmbH Voice assistant for wireless earpieces
US11085871B2 (en) 2016-07-06 2021-08-10 Bragi GmbH Optical vibration detection system and method
US11116415B2 (en) 2017-06-07 2021-09-14 Bragi GmbH Use of body-worn radar for biometric measurements, contextual awareness and identification
US11200026B2 (en) 2016-08-26 2021-12-14 Bragi GmbH Wireless earpiece with a passive virtual assistant
US11272367B2 (en) 2017-09-20 2022-03-08 Bragi GmbH Wireless earpieces for hub communications
US11283742B2 (en) 2016-09-27 2022-03-22 Bragi GmbH Audio-based social media platform
US11380430B2 (en) 2017-03-22 2022-07-05 Bragi GmbH System and method for populating electronic medical records with wireless earpieces
US11490858B2 (en) 2016-08-31 2022-11-08 Bragi GmbH Disposable sensor array wearable device sleeve system and method
US11544104B2 (en) 2017-03-22 2023-01-03 Bragi GmbH Load sharing between wireless earpieces
US11694771B2 (en) 2017-03-22 2023-07-04 Bragi GmbH System and method for populating electronic health records with wireless earpieces
US11799852B2 (en) 2016-03-29 2023-10-24 Bragi GmbH Wireless dongle for communications with wireless earpieces

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10200780B2 (en) 2016-08-29 2019-02-05 Bragi GmbH Method and apparatus for conveying battery life of wireless earpiece
USD822645S1 (en) 2016-09-03 2018-07-10 Bragi GmbH Headphone

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4751738A (en) * 1984-11-29 1988-06-14 The Board Of Trustees Of The Leland Stanford Junior University Directional hearing aid
US5479522A (en) * 1993-09-17 1995-12-26 Audiologic, Inc. Binaural hearing aid
US5524056A (en) * 1993-04-13 1996-06-04 Etymotic Research, Inc. Hearing aid having plural microphones and a microphone switching system
US5680466A (en) * 1994-10-06 1997-10-21 Zelikovitz; Joseph Omnidirectional hearing aid
US5793875A (en) * 1996-04-22 1998-08-11 Cardinal Sound Labs, Inc. Directional hearing system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4751738A (en) * 1984-11-29 1988-06-14 The Board Of Trustees Of The Leland Stanford Junior University Directional hearing aid
US5524056A (en) * 1993-04-13 1996-06-04 Etymotic Research, Inc. Hearing aid having plural microphones and a microphone switching system
US5479522A (en) * 1993-09-17 1995-12-26 Audiologic, Inc. Binaural hearing aid
US5680466A (en) * 1994-10-06 1997-10-21 Zelikovitz; Joseph Omnidirectional hearing aid
US5793875A (en) * 1996-04-22 1998-08-11 Cardinal Sound Labs, Inc. Directional hearing system

Cited By (192)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2002028143A3 (en) * 2000-09-29 2003-07-24 Siemens Audiologische Technik Method for operating a hearing aid system and hearing aid system
US7020296B2 (en) 2000-09-29 2006-03-28 Siemens Audiologische Technik Gmbh Method for operating a hearing aid system and hearing aid system
WO2002028143A2 (en) * 2000-09-29 2002-04-04 Siemens Audiologische Technik Gmbh Method for operating a hearing aid system and hearing aid system
US8527280B2 (en) * 2001-12-13 2013-09-03 Peter V. Boesen Voice communication device with foreign language translation
US20170011027A1 (en) * 2001-12-13 2017-01-12 Peter V. Boesen Voice communication device with foreign language translation
US9438294B2 (en) 2001-12-13 2016-09-06 Peter V. Boesen Voice communication device with foreign language translation
US20160065259A1 (en) * 2001-12-13 2016-03-03 Peter V. Boesen Voice communication device with foreign language translation
NL1021485C2 (en) * 2002-09-18 2004-03-22 Stichting Tech Wetenschapp Hearing glasses assembly.
WO2004028203A2 (en) * 2002-09-18 2004-04-01 Stichting Voor De Technische Wetenschappen Spectacle hearing aid
WO2004028203A3 (en) * 2002-09-18 2004-07-08 Stichting Tech Wetenschapp Spectacle hearing aid
US7609842B2 (en) 2002-09-18 2009-10-27 Varibel B.V. Spectacle hearing aid
CN1701634B (en) * 2002-09-18 2010-05-26 万利宝有限公司 Spectacle hearing aid
EP1683392A2 (en) * 2003-11-12 2006-07-26 Oticon A/S Microphone system
EP1683392A4 (en) * 2003-11-12 2007-10-31 Oticon As Microphone system
WO2008107359A1 (en) * 2007-03-05 2008-09-12 Siemens Audiologische Technik Gmbh Hearing system with distributed signal processing and corresponding method
US8641596B2 (en) 2008-07-02 2014-02-04 Cochlear Limited Wireless communication in a multimodal auditory prosthesis
CN101778330B (en) * 2009-12-30 2013-04-03 南京信息工程大学 Mobile phone platform-based array microphone hearing aid and control method thereof
EP2568719A1 (en) * 2011-09-07 2013-03-13 ITI Infotech Industries Limited Wearable sound amplification apparatus for the hearing impaired
CN103873972A (en) * 2012-12-12 2014-06-18 森海塞尔通信公司 Microphone boom
CN103873972B (en) * 2012-12-12 2019-03-05 森海塞尔通信公司 Mike boom
US20140185845A1 (en) * 2012-12-28 2014-07-03 Gn Resound A/S Spectacle hearing device system
US10231065B2 (en) * 2012-12-28 2019-03-12 Gn Hearing A/S Spectacle hearing device system
EP2840807A1 (en) * 2013-08-19 2015-02-25 Oticon A/s External microphone array and hearing aid using it
US9510112B2 (en) 2013-08-19 2016-11-29 Oticon A/S External microphone array and hearing aid using it
US9813826B2 (en) 2015-08-29 2017-11-07 Bragi GmbH Earpiece with electronic environmental sound pass-through system
US10439679B2 (en) 2015-08-29 2019-10-08 Bragi GmbH Multimodal communication system using induction and radio and method
US9843853B2 (en) 2015-08-29 2017-12-12 Bragi GmbH Power control for battery powered personal area network device system and method
US9854372B2 (en) 2015-08-29 2017-12-26 Bragi GmbH Production line PCB serial programming and testing method and system
US9800966B2 (en) 2015-08-29 2017-10-24 Bragi GmbH Smart case power utilization control system and method
US9866282B2 (en) 2015-08-29 2018-01-09 Bragi GmbH Magnetic induction antenna for use in a wearable device
US9905088B2 (en) 2015-08-29 2018-02-27 Bragi GmbH Responsive visual communication system and method
US10672239B2 (en) 2015-08-29 2020-06-02 Bragi GmbH Responsive visual communication system and method
US9949008B2 (en) 2015-08-29 2018-04-17 Bragi GmbH Reproduction of ambient environmental sound for acoustic transparency of ear canal device system and method
US9755704B2 (en) 2015-08-29 2017-09-05 Bragi GmbH Multimodal communication system induction and radio and method
US9949013B2 (en) 2015-08-29 2018-04-17 Bragi GmbH Near field gesture control system and method
US9972895B2 (en) 2015-08-29 2018-05-15 Bragi GmbH Antenna for use in a wearable device
US10234133B2 (en) 2015-08-29 2019-03-19 Bragi GmbH System and method for prevention of LED light spillage
US10203773B2 (en) 2015-08-29 2019-02-12 Bragi GmbH Interactive product packaging system and method
US10194228B2 (en) 2015-08-29 2019-01-29 Bragi GmbH Load balancing to maximize device function in a personal area network device system and method
US10412478B2 (en) 2015-08-29 2019-09-10 Bragi GmbH Reproduction of ambient environmental sound for acoustic transparency of ear canal device system and method
US10194232B2 (en) 2015-08-29 2019-01-29 Bragi GmbH Responsive packaging system for managing display actions
US10297911B2 (en) 2015-08-29 2019-05-21 Bragi GmbH Antenna for use in a wearable device
US10409394B2 (en) 2015-08-29 2019-09-10 Bragi GmbH Gesture based control system based upon device orientation system and method
US10122421B2 (en) 2015-08-29 2018-11-06 Bragi GmbH Multimodal communication system using induction and radio and method
US10397688B2 (en) 2015-08-29 2019-08-27 Bragi GmbH Power control for battery powered personal area network device system and method
US10117014B2 (en) 2015-08-29 2018-10-30 Bragi GmbH Power control for battery powered personal area network device system and method
US10104487B2 (en) 2015-08-29 2018-10-16 Bragi GmbH Production line PCB serial programming and testing method and system
US10382854B2 (en) 2015-08-29 2019-08-13 Bragi GmbH Near field gesture control system and method
US10506322B2 (en) 2015-10-20 2019-12-10 Bragi GmbH Wearable device onboard applications system and method
US10212505B2 (en) 2015-10-20 2019-02-19 Bragi GmbH Multi-point multiple sensor array for data sensing and processing system and method
US11683735B2 (en) 2015-10-20 2023-06-20 Bragi GmbH Diversity bluetooth system and method
US9866941B2 (en) 2015-10-20 2018-01-09 Bragi GmbH Multi-point multiple sensor array for data sensing and processing system and method
US10342428B2 (en) 2015-10-20 2019-07-09 Bragi GmbH Monitoring pulse transmissions using radar
US11419026B2 (en) 2015-10-20 2022-08-16 Bragi GmbH Diversity Bluetooth system and method
US10453450B2 (en) 2015-10-20 2019-10-22 Bragi GmbH Wearable earpiece voice command control system and method
US11064408B2 (en) 2015-10-20 2021-07-13 Bragi GmbH Diversity bluetooth system and method
US10206042B2 (en) 2015-10-20 2019-02-12 Bragi GmbH 3D sound field using bilateral earpieces system and method
US9980189B2 (en) 2015-10-20 2018-05-22 Bragi GmbH Diversity bluetooth system and method
US10175753B2 (en) 2015-10-20 2019-01-08 Bragi GmbH Second screen devices utilizing data from ear worn device system and method
US10582289B2 (en) 2015-10-20 2020-03-03 Bragi GmbH Enhanced biometric control systems for detection of emergency events system and method
US10104458B2 (en) 2015-10-20 2018-10-16 Bragi GmbH Enhanced biometric control systems for detection of emergency events system and method
US10635385B2 (en) 2015-11-13 2020-04-28 Bragi GmbH Method and apparatus for interfacing with wireless earpieces
US9978278B2 (en) 2015-11-27 2018-05-22 Bragi GmbH Vehicle to vehicle communications using ear pieces
US10099636B2 (en) 2015-11-27 2018-10-16 Bragi GmbH System and method for determining a user role and user settings associated with a vehicle
US9944295B2 (en) 2015-11-27 2018-04-17 Bragi GmbH Vehicle with wearable for identifying role of one or more users and adjustment of user settings
US10104460B2 (en) 2015-11-27 2018-10-16 Bragi GmbH Vehicle with interaction between entertainment systems and wearable devices
US10040423B2 (en) 2015-11-27 2018-08-07 Bragi GmbH Vehicle with wearable for identifying one or more vehicle occupants
US10155524B2 (en) 2015-11-27 2018-12-18 Bragi GmbH Vehicle with wearable for identifying role of one or more users and adjustment of user settings
US10542340B2 (en) 2015-11-30 2020-01-21 Bragi GmbH Power management for wireless earpieces
US10099374B2 (en) 2015-12-01 2018-10-16 Bragi GmbH Robotic safety using wearables
US10620698B2 (en) 2015-12-21 2020-04-14 Bragi GmbH Voice dictation systems using earpiece microphone system and method
US10904653B2 (en) 2015-12-21 2021-01-26 Bragi GmbH Microphone natural speech capture voice dictation system and method
US9939891B2 (en) 2015-12-21 2018-04-10 Bragi GmbH Voice dictation systems using earpiece microphone system and method
US11496827B2 (en) 2015-12-21 2022-11-08 Bragi GmbH Microphone natural speech capture voice dictation system and method
US9980033B2 (en) 2015-12-21 2018-05-22 Bragi GmbH Microphone natural speech capture voice dictation system and method
US10206052B2 (en) 2015-12-22 2019-02-12 Bragi GmbH Analytical determination of remote battery temperature through distributed sensor array system and method
US10575083B2 (en) 2015-12-22 2020-02-25 Bragi GmbH Near field based earpiece data transfer system and method
US10334345B2 (en) 2015-12-29 2019-06-25 Bragi GmbH Notification and activation system utilizing onboard sensors of wireless earpieces
US10154332B2 (en) 2015-12-29 2018-12-11 Bragi GmbH Power management for wireless earpieces utilizing sensor measurements
US10200790B2 (en) 2016-01-15 2019-02-05 Bragi GmbH Earpiece with cellular connectivity
US10104486B2 (en) 2016-01-25 2018-10-16 Bragi GmbH In-ear sensor calibration and detecting system and method
US10129620B2 (en) 2016-01-25 2018-11-13 Bragi GmbH Multilayer approach to hydrophobic and oleophobic system and method
US10085091B2 (en) 2016-02-09 2018-09-25 Bragi GmbH Ambient volume modification through environmental microphone feedback loop system and method
US10412493B2 (en) 2016-02-09 2019-09-10 Bragi GmbH Ambient volume modification through environmental microphone feedback loop system and method
US10667033B2 (en) 2016-03-02 2020-05-26 Bragi GmbH Multifactorial unlocking function for smart wearable device and method
US10327082B2 (en) 2016-03-02 2019-06-18 Bragi GmbH Location based tracking using a wireless earpiece device, system, and method
US11336989B2 (en) 2016-03-11 2022-05-17 Bragi GmbH Earpiece with GPS receiver
US10893353B2 (en) 2016-03-11 2021-01-12 Bragi GmbH Earpiece with GPS receiver
US11700475B2 (en) 2016-03-11 2023-07-11 Bragi GmbH Earpiece with GPS receiver
US10085082B2 (en) 2016-03-11 2018-09-25 Bragi GmbH Earpiece with GPS receiver
US10506328B2 (en) 2016-03-14 2019-12-10 Bragi GmbH Explosive sound pressure level active noise cancellation
US10045116B2 (en) 2016-03-14 2018-08-07 Bragi GmbH Explosive sound pressure level active noise cancellation utilizing completely wireless earpieces system and method
US10433788B2 (en) 2016-03-23 2019-10-08 Bragi GmbH Earpiece life monitor with capability of automatic notification system and method
US10052065B2 (en) 2016-03-23 2018-08-21 Bragi GmbH Earpiece life monitor with capability of automatic notification system and method
US10856809B2 (en) 2016-03-24 2020-12-08 Bragi GmbH Earpiece with glucose sensor and system
US10334346B2 (en) 2016-03-24 2019-06-25 Bragi GmbH Real-time multivariable biometric analysis and display system and method
US11799852B2 (en) 2016-03-29 2023-10-24 Bragi GmbH Wireless dongle for communications with wireless earpieces
USD821970S1 (en) 2016-04-07 2018-07-03 Bragi GmbH Wearable device charger
USD823835S1 (en) 2016-04-07 2018-07-24 Bragi GmbH Earphone
USD819438S1 (en) 2016-04-07 2018-06-05 Bragi GmbH Package
USD805060S1 (en) 2016-04-07 2017-12-12 Bragi GmbH Earphone
US10313781B2 (en) 2016-04-08 2019-06-04 Bragi GmbH Audio accelerometric feedback through bilateral ear worn device system and method
US10015579B2 (en) 2016-04-08 2018-07-03 Bragi GmbH Audio accelerometric feedback through bilateral ear worn device system and method
US10747337B2 (en) 2016-04-26 2020-08-18 Bragi GmbH Mechanical detection of a touch movement using a sensor and a special surface pattern system and method
US10169561B2 (en) 2016-04-28 2019-01-01 Bragi GmbH Biometric interface system and method
US10013542B2 (en) 2016-04-28 2018-07-03 Bragi GmbH Biometric interface system and method
USD949130S1 (en) 2016-05-06 2022-04-19 Bragi GmbH Headphone
USD824371S1 (en) 2016-05-06 2018-07-31 Bragi GmbH Headphone
USD836089S1 (en) 2016-05-06 2018-12-18 Bragi GmbH Headphone
US10045110B2 (en) 2016-07-06 2018-08-07 Bragi GmbH Selective sound field environment processing system and method
US10216474B2 (en) 2016-07-06 2019-02-26 Bragi GmbH Variable computing engine for interactive media based upon user biometrics
US10448139B2 (en) 2016-07-06 2019-10-15 Bragi GmbH Selective sound field environment processing system and method
US10888039B2 (en) 2016-07-06 2021-01-05 Bragi GmbH Shielded case for wireless earpieces
US11085871B2 (en) 2016-07-06 2021-08-10 Bragi GmbH Optical vibration detection system and method
US10470709B2 (en) 2016-07-06 2019-11-12 Bragi GmbH Detection of metabolic disorders using wireless earpieces
US10045736B2 (en) 2016-07-06 2018-08-14 Bragi GmbH Detection of metabolic disorders using wireless earpieces
US11781971B2 (en) 2016-07-06 2023-10-10 Bragi GmbH Optical vibration detection system and method
US11770918B2 (en) 2016-07-06 2023-09-26 Bragi GmbH Shielded case for wireless earpieces
US11497150B2 (en) 2016-07-06 2022-11-08 Bragi GmbH Shielded case for wireless earpieces
US10582328B2 (en) 2016-07-06 2020-03-03 Bragi GmbH Audio response based on user worn microphones to direct or adapt program responses system and method
US10555700B2 (en) 2016-07-06 2020-02-11 Bragi GmbH Combined optical sensor for audio and pulse oximetry system and method
US10201309B2 (en) 2016-07-06 2019-02-12 Bragi GmbH Detection of physiological data using radar/lidar of wireless earpieces
US10621583B2 (en) 2016-07-07 2020-04-14 Bragi GmbH Wearable earpiece multifactorial biometric analysis system and method
US10158934B2 (en) 2016-07-07 2018-12-18 Bragi GmbH Case for multiple earpiece pairs
US10469931B2 (en) 2016-07-07 2019-11-05 Bragi GmbH Comparative analysis of sensors to control power status for wireless earpieces
US10165350B2 (en) 2016-07-07 2018-12-25 Bragi GmbH Earpiece with app environment
US10516930B2 (en) 2016-07-07 2019-12-24 Bragi GmbH Comparative analysis of sensors to control power status for wireless earpieces
US10587943B2 (en) 2016-07-09 2020-03-10 Bragi GmbH Earpiece with wirelessly recharging battery
US10397686B2 (en) 2016-08-15 2019-08-27 Bragi GmbH Detection of movement adjacent an earpiece device
US10977348B2 (en) 2016-08-24 2021-04-13 Bragi GmbH Digital signature using phonometry and compiled biometric data system and method
US11620368B2 (en) 2016-08-24 2023-04-04 Bragi GmbH Digital signature using phonometry and compiled biometric data system and method
US10409091B2 (en) 2016-08-25 2019-09-10 Bragi GmbH Wearable with lenses
US11086593B2 (en) 2016-08-26 2021-08-10 Bragi GmbH Voice assistant for wireless earpieces
US10313779B2 (en) 2016-08-26 2019-06-04 Bragi GmbH Voice assistant system for wireless earpieces
US10887679B2 (en) 2016-08-26 2021-01-05 Bragi GmbH Earpiece for audiograms
US11200026B2 (en) 2016-08-26 2021-12-14 Bragi GmbH Wireless earpiece with a passive virtual assistant
US11573763B2 (en) 2016-08-26 2023-02-07 Bragi GmbH Voice assistant for wireless earpieces
US11861266B2 (en) 2016-08-26 2024-01-02 Bragi GmbH Voice assistant for wireless earpieces
US11490858B2 (en) 2016-08-31 2022-11-08 Bragi GmbH Disposable sensor array wearable device sleeve system and method
US10598506B2 (en) 2016-09-12 2020-03-24 Bragi GmbH Audio navigation using short range bilateral earpieces
US10580282B2 (en) 2016-09-12 2020-03-03 Bragi GmbH Ear based contextual environment and biometric pattern recognition system and method
US10852829B2 (en) 2016-09-13 2020-12-01 Bragi GmbH Measurement of facial muscle EMG potentials for predictive analysis using a smart wearable system and method
US11294466B2 (en) 2016-09-13 2022-04-05 Bragi GmbH Measurement of facial muscle EMG potentials for predictive analysis using a smart wearable system and method
US11675437B2 (en) 2016-09-13 2023-06-13 Bragi GmbH Measurement of facial muscle EMG potentials for predictive analysis using a smart wearable system and method
US11283742B2 (en) 2016-09-27 2022-03-22 Bragi GmbH Audio-based social media platform
US11627105B2 (en) 2016-09-27 2023-04-11 Bragi GmbH Audio-based social media platform
US10460095B2 (en) 2016-09-30 2019-10-29 Bragi GmbH Earpiece with biometric identifiers
US10049184B2 (en) 2016-10-07 2018-08-14 Bragi GmbH Software application transmission via body interface using a wearable device in conjunction with removable body sensor arrays system and method
US10942701B2 (en) 2016-10-31 2021-03-09 Bragi GmbH Input and edit functions utilizing accelerometer based earpiece movement system and method
US10698983B2 (en) 2016-10-31 2020-06-30 Bragi GmbH Wireless earpiece with a medical engine
US11947874B2 (en) 2016-10-31 2024-04-02 Bragi GmbH Input and edit functions utilizing accelerometer based earpiece movement system and method
US10771877B2 (en) 2016-10-31 2020-09-08 Bragi GmbH Dual earpieces for same ear
US10455313B2 (en) 2016-10-31 2019-10-22 Bragi GmbH Wireless earpiece with force feedback
US11599333B2 (en) 2016-10-31 2023-03-07 Bragi GmbH Input and edit functions utilizing accelerometer based earpiece movement system and method
US10617297B2 (en) 2016-11-02 2020-04-14 Bragi GmbH Earpiece with in-ear electrodes
US10117604B2 (en) 2016-11-02 2018-11-06 Bragi GmbH 3D sound positioning with distributed sensors
US11417307B2 (en) 2016-11-03 2022-08-16 Bragi GmbH Selective audio isolation from body generated sound system and method
US11325039B2 (en) 2016-11-03 2022-05-10 Bragi GmbH Gaming with earpiece 3D audio
US10821361B2 (en) 2016-11-03 2020-11-03 Bragi GmbH Gaming with earpiece 3D audio
US10205814B2 (en) 2016-11-03 2019-02-12 Bragi GmbH Wireless earpiece with walkie-talkie functionality
US11908442B2 (en) 2016-11-03 2024-02-20 Bragi GmbH Selective audio isolation from body generated sound system and method
US11806621B2 (en) 2016-11-03 2023-11-07 Bragi GmbH Gaming with earpiece 3D audio
US10225638B2 (en) 2016-11-03 2019-03-05 Bragi GmbH Ear piece with pseudolite connectivity
US10896665B2 (en) 2016-11-03 2021-01-19 Bragi GmbH Selective audio isolation from body generated sound system and method
US10062373B2 (en) 2016-11-03 2018-08-28 Bragi GmbH Selective audio isolation from body generated sound system and method
US10045117B2 (en) 2016-11-04 2018-08-07 Bragi GmbH Earpiece with modified ambient environment over-ride function
US10397690B2 (en) 2016-11-04 2019-08-27 Bragi GmbH Earpiece with modified ambient environment over-ride function
US10063957B2 (en) 2016-11-04 2018-08-28 Bragi GmbH Earpiece with source selection within ambient environment
US10058282B2 (en) 2016-11-04 2018-08-28 Bragi GmbH Manual operation assistance with earpiece with 3D sound cues
US10681450B2 (en) 2016-11-04 2020-06-09 Bragi GmbH Earpiece with source selection within ambient environment
US10681449B2 (en) 2016-11-04 2020-06-09 Bragi GmbH Earpiece with added ambient environment
US10045112B2 (en) 2016-11-04 2018-08-07 Bragi GmbH Earpiece with added ambient environment
US10398374B2 (en) 2016-11-04 2019-09-03 Bragi GmbH Manual operation assistance with earpiece with 3D sound cues
US10506327B2 (en) 2016-12-27 2019-12-10 Bragi GmbH Ambient environmental sound field manipulation based on user defined voice and audio recognition pattern analysis system and method
US10405081B2 (en) 2017-02-08 2019-09-03 Bragi GmbH Intelligent wireless headset system
US10582290B2 (en) 2017-02-21 2020-03-03 Bragi GmbH Earpiece with tap functionality
US10771881B2 (en) 2017-02-27 2020-09-08 Bragi GmbH Earpiece with audio 3D menu
US10575086B2 (en) 2017-03-22 2020-02-25 Bragi GmbH System and method for sharing wireless earpieces
US11380430B2 (en) 2017-03-22 2022-07-05 Bragi GmbH System and method for populating electronic medical records with wireless earpieces
US11694771B2 (en) 2017-03-22 2023-07-04 Bragi GmbH System and method for populating electronic health records with wireless earpieces
US11710545B2 (en) 2017-03-22 2023-07-25 Bragi GmbH System and method for populating electronic medical records with wireless earpieces
US11544104B2 (en) 2017-03-22 2023-01-03 Bragi GmbH Load sharing between wireless earpieces
US10708699B2 (en) 2017-05-03 2020-07-07 Bragi GmbH Hearing aid with added functionality
US11116415B2 (en) 2017-06-07 2021-09-14 Bragi GmbH Use of body-worn radar for biometric measurements, contextual awareness and identification
US11013445B2 (en) 2017-06-08 2021-05-25 Bragi GmbH Wireless earpiece with transcranial stimulation
US11911163B2 (en) 2017-06-08 2024-02-27 Bragi GmbH Wireless earpiece with transcranial stimulation
US10344960B2 (en) 2017-09-19 2019-07-09 Bragi GmbH Wireless earpiece controlled medical headlight
US11272367B2 (en) 2017-09-20 2022-03-08 Bragi GmbH Wireless earpieces for hub communications
US11711695B2 (en) 2017-09-20 2023-07-25 Bragi GmbH Wireless earpieces for hub communications
CN111327984A (en) * 2020-02-27 2020-06-23 北京声加科技有限公司 Earphone auxiliary listening method based on null filtering and ear-worn equipment
CN111327984B (en) * 2020-02-27 2022-04-22 北京声加科技有限公司 Earphone auxiliary listening method based on null filtering and ear-worn equipment
CN112911480A (en) * 2021-01-22 2021-06-04 成都市舒听医疗器械有限责任公司 Hearing aid sound amplification method and hearing aid

Also Published As

Publication number Publication date
EP1017252A3 (en) 2006-05-31

Similar Documents

Publication Publication Date Title
EP1017252A2 (en) Hearing aid system
US8184823B2 (en) Headphone device, sound reproduction system, and sound reproduction method
US9949040B2 (en) Peer to peer hearing system
US6389142B1 (en) In-the-ear hearing aid with directional microphone system
EP0855130B1 (en) Digital hearing aid system
EP2088802B1 (en) Method of estimating weighting function of audio signals in a hearing aid
EP2360943B1 (en) Beamforming in hearing aids
US4751738A (en) Directional hearing aid
US5289544A (en) Method and apparatus for reducing background noise in communication systems and for enhancing binaural hearing systems for the hearing impaired
US9980055B2 (en) Hearing device and a hearing system configured to localize a sound source
US20080008339A1 (en) Audio processing system and method
US7340073B2 (en) Hearing aid and operating method with switching among different directional characteristics
EP2537353A1 (en) Device and method for direction dependent spatial noise reduction
EP1326478A2 (en) Method for producing control signals, method of controlling signal transfer and a hearing device
US9723403B2 (en) Wearable directional microphone array apparatus and system
JP3254789B2 (en) Hearing aid
US10911877B2 (en) Hearing device with adaptive binaural auditory steering and related method
EP3148217B1 (en) Method for operating a binaural hearing system
KR101271517B1 (en) The Acoustic Multipole Array and Packaging and Controlling Method of thereof
CN115967883A (en) Earphone, user equipment and method for processing signal

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

AX Request for extension of the european patent

Free format text: AL;LT;LV;MK;RO;SI

PUAL Search report despatched

Free format text: ORIGINAL CODE: 0009013

AK Designated contracting states

Kind code of ref document: A3

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

AX Request for extension of the european patent

Extension state: AL LT LV MK RO SI

17P Request for examination filed

Effective date: 20060914

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION HAS BEEN WITHDRAWN

18W Application withdrawn

Effective date: 20061129