US20160119725A1 - Packet loss concealment techniques for phone-to-hearing-aid streaming - Google Patents
Packet loss concealment techniques for phone-to-hearing-aid streaming Download PDFInfo
- Publication number
- US20160119725A1 US20160119725A1 US14/854,704 US201514854704A US2016119725A1 US 20160119725 A1 US20160119725 A1 US 20160119725A1 US 201514854704 A US201514854704 A US 201514854704A US 2016119725 A1 US2016119725 A1 US 2016119725A1
- Authority
- US
- United States
- Prior art keywords
- frame
- cross
- samples
- faded
- hearing assistance
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 102
- 238000005562 fading Methods 0.000 claims abstract description 18
- 230000003044 adaptive effect Effects 0.000 claims description 6
- 230000005540 biological transmission Effects 0.000 claims description 3
- 238000004891 communication Methods 0.000 description 24
- 230000015654 memory Effects 0.000 description 17
- 238000012545 processing Methods 0.000 description 14
- 238000013213 extrapolation Methods 0.000 description 13
- 230000000694 effects Effects 0.000 description 4
- 239000000945 filler Substances 0.000 description 4
- 230000005291 magnetic effect Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 230000003068 static effect Effects 0.000 description 3
- 230000015572 biosynthetic process Effects 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000003786 synthesis reaction Methods 0.000 description 2
- 239000003826 tablet Substances 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 1
- 230000003321 amplification Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000003139 buffering effect Effects 0.000 description 1
- 230000010267 cellular communication Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000000875 corresponding effect Effects 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000001934 delay Effects 0.000 description 1
- 210000000613 ear canal Anatomy 0.000 description 1
- 210000005069 ears Anatomy 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000007274 generation of a signal involved in cell-cell signaling Effects 0.000 description 1
- 208000016354 hearing loss disease Diseases 0.000 description 1
- 239000007943 implant Substances 0.000 description 1
- 230000001939 inductive effect Effects 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 230000005923 long-lasting effect Effects 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000003199 nucleic acid amplification method Methods 0.000 description 1
- 230000000737 periodic effect Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- 238000002604 ultrasonography Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/55—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired
- H04R25/554—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception using an external connection, either wireless or wired using a wireless connection, e.g. between microphone and amplifier or using Tcoils
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/005—Correction of errors induced by the transmission channel, if related to the coding algorithm
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/30—Monitoring or testing of hearing aids, e.g. functioning, settings, battery power
- H04R25/305—Self-monitoring or self-testing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R25/00—Deaf-aid sets, i.e. electro-acoustic or electro-mechanical hearing aids; Electric tinnitus maskers providing an auditory perception
- H04R25/43—Electronic input selection or mixing based on input signal analysis, e.g. mixing or selection between microphone and telecoil or between microphones with different directivity characteristics
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/03—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
- G10L25/12—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being prediction coefficients
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2225/00—Details of deaf aids covered by H04R25/00, not provided for in any of its subgroups
- H04R2225/43—Signal processing in hearing aids to enhance the speech intelligibility
Definitions
- ADPCM Adaptive differential pulse-code modulation
- ADPCM has a low latency, good quality, a low bitrate, and low computational requirements.
- ADPCM has a low latency, good quality, a low bitrate, and low computational requirements.
- ADPCM is negatively affected by packet-loss.
- the negative impact on resulting audio quality when packet-loss occurs with ADPCM is not limited to the dropped packet, but also up to several dozens of milliseconds after the dropped packet.
- the encoder and the decoder both maintain a certain state based on the encoded signal which, under normal operation and after initial convergence, is the same.
- a packet drop causes the encoder and the decoder states to depart from one another, and the decoder state will take time to converge back to the encoder state once valid data is available again after a drop.
- Packet-loss-concealment (PLC) techniques mitigate the error caused by packet loss. While there are multiple single-channel PLC techniques currently used, they are often slow and costly in terms of instructions per second used, and thus can be infeasible in a hearing assistance device setting.
- a method for packet loss concealment includes receiving a first frame at a hearing assistance device, determining, at the hearing assistance device, that a second frame was not received within a predetermined time, and determining a first set of sequential samples that match the first frame.
- the method can include cross-fading the first frame and the first set of sequential samples to create a first cross-faded frame and extrapolating a third frame to replace the second frame using the first set of sequential samples and an autoregressive model.
- FIG. 1 illustrates generally a graph showing effects of a packet loss on an encoded signal in accordance with some embodiments of the present subject matter.
- FIG. 2 illustrates generally a graph showing an autoregressive-based technique for extrapolation of an audio signal in accordance with some embodiments of the present subject matter.
- FIG. 3 illustrates generally graphs of cross-faded waveforms with filled in frames in accordance with some embodiments of the present subject matter.
- FIGS. 4A-4D illustrate generally frame by frame replacements in accordance with some embodiments of the present subject matter.
- FIGS. 5A-5D illustrate generally cross-faded samples for replacement frames in accordance with some embodiments of the present subject matter.
- FIG. 6 illustrates generally a hearing assistance device in accordance with some embodiments of the present subject matter.
- FIG. 7 illustrates generally a flowchart for packet loss concealment in accordance with some embodiments of the present subject matter.
- FIG. 8 illustrates generally an example of a block diagram of a machine upon which any one or more of the techniques discussed herein can perform in accordance with some embodiments of the present subject matter.
- Adaptive differential pulse-code modulation is useful for improving hearing assistance device functionality when streaming from device-to-device, but is particularly susceptible to packet loss issues.
- Packet-loss-concealment (PLC) techniques are therefore used to mitigate the error caused by packet loss.
- a hearing assistance device can use a codec to decode audio.
- a codec for example, a G.722 audio codec according to an International Telecommunication Union Telecommunication Standardization Sector (ITU-T) standard can be used in an ADPCM system to decode audio.
- ITU-T International Telecommunication Union Telecommunication Standardization Sector
- Other codec apparatus and methods can be used without departing from the scope of the present subject matter.
- FIG. 1 illustrates generally a graph 100 showing effects of a packet loss on an encoded signal in accordance with some embodiments of the present subject matter.
- the graph 100 includes signals encoded using an ADPCM-based codec.
- the graph 100 includes a first encoded signal 102 over time with no packet loss and a second signal 104 over time with a packet loss.
- the packet loss is highlighted on the graph 100 of the second signal 104 with a box 106 .
- the packet loss affects the signal not only at, but also after, the packet loss highlighted by box 106 .
- the effect of the packet loss is not confined to just the window of the lost packets, but beyond.
- a PLC technique includes artificially “filling” the blank caused by the packet drop with a filler signal that is thought to naturally extend the signal with one that is similar enough to the original signal that was encoded at the encoder.
- This replacement signal is sometimes called a concealment frame, and there exists a wide range of solutions to generate it.
- the mere generation of a filler does not always remove the state inconsistencies, which can potentially create long-lasting and highly audible artifacts.
- re-encoding the synthetic concealment frame at the decoder can be used to avoid artifacts. This allows for the decoder's state to keep updating, and with an appropriate “filler” signal, its state will not drift too far from the encoder's state at the end of the frame.
- One method to generate a concealment frame is the so-called “pitch-repetition” method.
- the last 20 milliseconds of audio are stored in a circular buffer, and from this data an estimate of the pitch is tracked (essentially by doing a brute-force search from circular delays of 5 to 15 milliseconds, using a maximum correlation-based method).
- a 10-millisecond concealment frame is generated synthetically from a repetition of the last pitch period.
- Some windowing and overlap-add can also be employed to ensure smooth transitions between the real and synthetic signals, and slight structural changes in the decoder can be made. However, repeating the last full pitch period can have a negative impact in quality for erroneously detected pitch.
- an “enhanced” pitch repetition method can be used, where multiple additional tricks are used to improve the resulting quality, including: using the so-called “ringing signal” (e.g., the first few samples coming out of the synthesis filter with zero-input, as seen right after the beginning of box 106 at the bottom of FIG.
- ringing signal e.g., the first few samples coming out of the synthesis filter with zero-input, as seen right after the beginning of box 106 at the bottom of FIG.
- a low-complexity PLC technique can be used, which can include a tradeoff of lower quality for less complexity. It can introduce short and long-term predictors that it uses to generate the extrapolated signal, it can “over-extrapolate” to cover the next available frame for cross-fading, or the like. It can include higher quality for music signals.
- a technique for packet loss concealment can include all-pole modelling an arbitrary length of buffered audio data for packet loss concealment.
- the technique can include an appropriate filter initialization, and letting the resulting filter “ring” for the duration of the blank.
- the Burg method can be used for a high output quality
- the Yule-Walker method can be used for a lighter computational burden to calculate or estimate autoregressive parameters. There are many methods to determine autoregressive parameters, and other methods besides Yule-Walker method and Burg method can be used without departing from the scope of the present subject matter.
- the internal state of the decoder can be updated by re-encoding it. In an example, this re-encoding is well indicated as the extrapolated waveform is typically very close to the true waveform (at least for the first few milliseconds of data), and therefore the decoder's state is adequately updated.
- FIG. 2 illustrates generally a graph 200 showing an autoregressive-based technique for extrapolation of an audio signal in accordance with some embodiments of the present subject matter.
- graph 200 can include an autoregressive extrapolation of a 16 kHz audio signal.
- the audio is interrupted at cut-off point 202 , at which point the extrapolation occurs.
- the autoregressive (AR) order is 100 , and the first 200 samples are used to determine the AR model.
- the extrapolation length is 500 samples.
- the extrapolated signal 204 and the true signal 206 are shown.
- the extrapolated signal 204 closely tracks the true signal 206 . In a packet loss situation, the true signal 206 would be missing, and the extrapolated signal 204 would be substituted for the missing true signal 206 .
- this technique when the model order is large-enough and well initialized, this technique produces high quality extrapolation and can be easily augmented to “latch” onto future available data by cross-fading. This technique also does not induce additional latency since the forward extension is very natural-sounding for several dozens of milliseconds.
- This technique can have a large computational complexity involved in determining autoregressive parameters.
- the technique can include large AR model orders (e.g., more than 50), which makes it difficult to include in current fixed-point processors. Using the large AR model orders, the technique produces results that are of markedly higher quality than previous techniques. Estimations techniques, increased processor capabilities, and modeling data can be used to implement this technique in hearing assistance devices, including at lower orders using less computational complexity.
- Various embodiments include concealing a source of artifacts and non-naturalness in discontinuities in the reconstructed signal as described below.
- One component of the technique includes pattern matching, and includes further operations to enforce a smooth and natural-sounding extrapolation. An assumption can be made that the decoder is aware of whether or not the next frame is valid or dropped, and an “extrapolated data” buffer can be available.
- FIG. 3 illustrates generally graphs 300 A- 300 D of cross-faded waveforms with filled in frames in accordance with some embodiments of the present subject matter.
- the technique can include, if the current frame 304 is valid (i.e., not dropped), first decoding the frame normally (without playing it back yet). If the previous frame 302 was not valid (i.e., dropped), the current frame can be cross-faded using the data just decoded above with the data present in the extrapolated data buffer over the entire frame. In the example shown in graph 300 A, the previous frame 302 is valid, so no cross-fading need take place.
- the technique can continue by determining if the next frame 306 is dropped, and if so (as is the case, as shown in graph 300 A), then beginning a specific form of extrapolation that will generate artificial data over the 2 next future frames (e.g., 308 and 310 ), and modify the current frame 304 (as shown in graph 300 B, where the current frame 304 is cross-faded with a matching segment from history). This can be done whether the next frame 306 is freshly decoded from a valid frame or from past extrapolations. If the current frame 304 is dropped, the technique can proceed without additional decoding and playback the audio data currently available in the “extrapolated data” buffer at the correct indices.
- FIG. 3 shows a technique for smoothly extending a waveform from time index 200 to 500 (e.g., over frames 306 , 308 , and 310 ), using 3 iterations as shown in graphs 300 B, 300 C, and 300 D.
- the 2 later matching tasks (those shown in graphs 300 C and 300 D), can be simplified as discussed below. For example, an assumption can be made that the locations of the later matching segments are likely known from the first search results.
- frames 308 and 310 can be extrapolated with less computational effort than that used to extrapolate frame 306 .
- frame 308 can include a matched segment that immediately follows the matched segment for extrapolated frame 306 and frame 310 can include a matched segment following the one used for frame 308 .
- the computation to determine the matched segments for frames 308 and 310 is simplified compared to the extrapolation computation for frame 306 .
- an audio extrapolation technique can minimize artifacts when used in the context described above.
- the technique of audio extrapolation described in details in the paragraphs below, can include extrapolating using overlapping blocks, with multiple and far-reaching cross-fadings when a new block is added.
- a match on the lower audio bands can be extrapolated, and the matching positions can be replicated in the higher audio bands without a noticeable quality impact.
- An extrapolation technique can include operations to extend an existing audio segment of length N to a length N+T.
- the technique can include operations to perform some arbitrary type of pattern matching to find a waveform in history that resembles the one in positions N ⁇ L+1 to N. The search can begin from N ⁇ 2L+1 and go backwards in history. Once a matching waveform is found, the current data in N ⁇ L+1 to N can be cross-faded with the matching waveform, and the L samples in history that follow the best match can be used to fill the positions N+1 to N+L. The current blank position can then include N+L+1. The exact same procedure can be repeated P times until data up to N+T is covered. In an example, there are P matching tasks and P cross-fading operations to fill the audio up until N+T.
- FIGS. 4A-4D illustrate generally frame by frame replacements in accordance with some embodiments of the present subject matter.
- the following frames e.g., frames 402 A-D, 404 A-D, and 406 A-D
- the frames in the places of D and E are dropped:
- Each frame can contain 3 samples:
- Frames A, B, and C are decoded normally, and before playing C back, the decoder is informed that the next frame (i.e., frame D) is dropped.
- the technique includes finding a match for frame C from its saved history, testing several positions up to the beginning of frame B.
- a new frame can be constructed C 0 .
- an appropriate match includes the samples [a 3 , b 1 , b 2 ] (see FIG. 4D ).
- the history has slightly changed, and matches within the new history can be slightly different.
- the best match to [c 3 , d 1 , d 2 ] can include [ ⁇ tilde over (c) ⁇ 3 , ⁇ tilde over (d) ⁇ 1 , ⁇ tilde over (d) ⁇ 2 ], which is exactly one frame forward from the match for new frame D 0 [b 3 , ⁇ tilde over (c) ⁇ 1 , ⁇ tilde over (c) ⁇ 2 ].
- Cross-fading of [ ⁇ tilde over (c) ⁇ 3 , ⁇ tilde over (d) ⁇ 1 , ⁇ tilde over (d) ⁇ 2 ] and [c 3 , d 1 , d 2 ] can be performed, which can ensure maximal smoothness of the extended waveform.
- the cross-fading operation can either partially be necessary (if the first few samples to be cross-faded are equal), or not necessary at all (if all samples are equal). As shown in FIGS. 4A and 4D , if the next frame F is valid (i.e., not dropped), it can be cross-faded with the next sequential match samples after those used for frame E. If frame G is also valid, then the technique can continue with no cross-fading for frame G.
- FIGS. 5A-5D illustrate generally cross-faded samples for replacement frames in accordance with some embodiments of the present subject matter.
- the original frames can be replaced by new frames that include extrapolated or original samples cross-faded with matched samples:
- FIG. 5A shows frame ⁇ tilde over (C) ⁇ as: [c 1 ′, c 2 ′, c 3 ′] including the original samples that are cross-faded to result in ⁇ tilde over (C) ⁇ .
- FIGS. 5B-5C show frames ⁇ tilde over (D) ⁇ , ⁇ tilde over (E) ⁇ , and ⁇ tilde over (F) ⁇ respectively, including the original samples that are cross-faded to result in those frames.
- an assumption can be made that one search is sufficient, and the next matches can be located one frame ahead from the initial match position, which can greatly reduce computational burden.
- the pattern match can be done over a central portion of a segment, where the cross-fading will have the most effect.
- T can be equal to 2L where L is the frame length, and if the computational requirements allow, other combinations can be used (e.g., 4L′ where L′ is half of L, etc.).
- the extension can cover an extra frame, which can correspond to a non-dropped frame. If the extra frame is a non-dropped frame, another full-frame cross-fading operation can take place (e.g., for frame F above).
- multiple iterations can be used, instead of a single search far enough back to cover enough frames.
- the multiple iterations can increase the chances of finding a highly correlated concealment frame since the search can be started close to the dropped frame (e.g., due to short-time stationarity assumptions).
- the multiple iterations can also cover situations where multiple frames are dropped in the future or in a row.
- Using multiple iterations also decreases the computational complexity of determining subsequent searches through approximation to subsequent samples after a first set of matching samples is determined for a first cross-faded frame.
- FIG. 6 illustrates generally a hearing assistance device 602 in accordance with some embodiments of the present subject matter.
- the hearing assistance device 602 can include a wireless receiver 604 to receive a first frame at the hearing assistance device 602 .
- the receiver 604 can be in communication, such as using wireless signals, with a mobile device or music player, such as a phone.
- the hearing assistance device 602 can include a processor 606 to determine that a second frame was not received by the receiver within a predetermined time and determine a first set of sequential samples that match the first frame.
- the processor 606 can cross-fade the first frame and the first set of sequential samples to create a first cross-faded frame and extrapolate a third frame to replace the second frame using the first set of sequential samples and an autoregressive model.
- the processor 606 can determine that a fourth frame was not received within a predetermined time.
- the processor 606 can cross-fade the third frame with the second set of samples to create a second cross-faded frame to replace the fourth frame.
- the processor 606 can determine parameters for the autoregressive model using one of a Burg method or a Yule-Walker method. Other methods to determine autoregressive parameters can be used without departing from the scope of the present subject matter.
- the processor 606 can encode the first frame using an adaptive differential pulse-code modulation.
- the hearing assistance device 602 can include a speaker 608 to play an audio signal corresponding to the first cross-faded frame.
- the speaker 608 can play the third frame after the first cross-faded frame.
- the speaker 608 can play the second cross-faded frame.
- the hearing assistance device 602 can include memory 610 to store a buffer including samples, flags (e.g., a dropped frame flag), etc.
- the third frame includes a second set of samples, and wherein the second set of samples includes samples starting one frame after the first set of sequential samples.
- the autoregressive model includes a model order of at least 50.
- FIG. 7 illustrates generally a flowchart 700 for packet loss concealment in accordance with some embodiments of the present subject matter.
- the flowchart 700 can include a method for packet loss concealment, the method comprising an operation 702 to receive a first frame at a hearing assistance device.
- the first frame can be received from a mobile device, such as a phone, a tablet, a dedicated music player, or the like.
- the method can include an operation 704 to determine, at the hearing assistance device, that a second frame was not received within a predetermined time.
- the method can include an operation 706 to determine a first set of sequential samples that match the first frame.
- the method can include an operation 708 to cross-fade the first frame and the first set of sequential samples to create a first cross-faded frame.
- the method can include an operation 710 to extrapolate a third frame to replace the second frame using the first set of sequential samples and an autoregressive model.
- the method can further include operations to play the first cross-faded frame or play the third frame after the first cross-faded frame.
- the method can include an operation to determine, at the hearing assistance device, that a fourth frame was not received within a predetermined time.
- the method can include an operation to cross-fade the third frame with the second set of samples to create a second cross-faded frame to replace the fourth frame.
- the method can include an operation to play the second cross-faded frame.
- the method includes an operation to determine parameters for the autoregressive model using a Burg method, a Yule-Walker method, and/or other methods for determining autoregressive parameters.
- the method can include an operation to encode the first frame using an adaptive differential pulse-code modulation (ADPCM).
- ADPCM adaptive differential pulse-code modulation
- the third frame can include a second set of samples, and wherein the second set of samples include samples starting one frame after the first set of sequential samples.
- the autoregressive model can include a model order, such as an order of 50 or greater.
- any type of pattern matching can be used with the techniques described herein.
- the multiple cross-fadings performed over entire segments can include sound that is smooth and natural compared to other packet loss concealment techniques.
- two or more of the techniques described herein can be used together for packet loss concealment.
- FIG. 8 illustrates generally an example of a block diagram of a machine 800 upon which any one or more of the techniques discussed herein can perform in accordance with some embodiments of the present subject matter.
- the machine 800 can operate as a standalone device or can be connected (e.g., networked) to other machines.
- the machine 800 can operate in the capacity of a server machine, a client machine, or both in server-client network environments.
- the machine 800 can act as a peer machine in peer-to-peer (P2P) (or other distributed) network environment.
- P2P peer-to-peer
- the machine 800 can be a personal computer (PC), a tablet PC, a set-top box (STB), a personal digital assistant (PDA), a mobile telephone, a web appliance, a network router, switch or bridge, or any machine capable of executing instructions (sequential or otherwise) that specify actions to be taken by that machine.
- Machine 800 can include processor 606 of hearing assistance device 602 in FIG. 6 .
- the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methods described herein, such as cloud computing, software as a service (SaaS), other computer cluster configurations.
- Examples, as described herein, can include, or can operate on, logic or a number of components, modules, or mechanisms.
- Modules are tangible entities (e.g., hardware) capable of performing specified operations when operating.
- a module includes hardware.
- the hardware can be specifically configured to carry out a specific operation (e.g., hardwired).
- the hardware can include configurable execution units (e.g., transistors, circuits, etc.) and a computer readable medium containing instructions, where the instructions configure the execution units to carry out a specific operation when in operation. The configuring can occur under the direction of the executions units or a loading mechanism. Accordingly, the execution units are communicatively coupled to the computer readable medium when the device is operating.
- the execution units can be a member of more than one module.
- the execution units can be configured by a first set of instructions to implement a first module at one point in time and reconfigured by a second set of instructions to implement a second module.
- Machine 800 can include a hardware processor 802 (e.g., a central processing unit (CPU), a graphics processing unit (GPU), a hardware processor core, or any combination thereof), a main memory 804 and a static memory 806 , some or all of which can communicate with each other via an interlink (e.g., bus) 808 .
- the machine 800 can further include a display unit 810 , an alphanumeric input device 812 (e.g., a keyboard), and a user interface (UI) navigation device 814 (e.g., a mouse).
- the display unit 810 , alphanumeric input device 812 and UI navigation device 814 can be a touch screen display.
- the machine 800 can additionally include a storage device (e.g., drive unit) 816 , a signal generation device 818 (e.g., a speaker), a network interface device 820 , and one or more sensors 821 , such as a global positioning system (GPS) sensor, compass, accelerometer, or other sensor.
- the machine 800 can include an output controller 828 , such as a serial (e.g., universal serial bus (USB), parallel, or other wired or wireless (e.g., infrared (IR), near field communication (NFC), etc.) connection to communicate or control one or more peripheral devices (e.g., a printer, card reader, etc.).
- a serial e.g., universal serial bus (USB), parallel, or other wired or wireless (e.g., infrared (IR), near field communication (NFC), etc.) connection to communicate or control one or more peripheral devices (e.g., a printer, card reader, etc.).
- USB universal serial bus
- the storage device 816 can include a machine readable medium 822 that is non-transitory on which is stored one or more sets of data structures or instructions 824 (e.g., software) embodying or utilized by any one or more of the techniques or functions described herein.
- the instructions 824 can also reside, completely or at least partially, within the main memory 804 , within static memory 806 , or within the hardware processor 802 during execution thereof by the machine 800 .
- one or any combination of the hardware processor 802 , the main memory 804 , the static memory 806 , or the storage device 816 can constitute machine readable media.
- machine readable medium 822 is illustrated as a single medium, the term “machine readable medium” can include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) configured to store the one or more instructions 824 .
- machine readable medium can include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) configured to store the one or more instructions 824 .
- machine readable medium can include any medium that is capable of storing, encoding, or carrying instructions for execution by the machine 800 and that cause the machine 800 to perform any one or more of the techniques of the present disclosure, or that is capable of storing, encoding or carrying data structures used by or associated with such instructions.
- Non-limiting machine readable medium examples can include solid-state memories, and optical and magnetic media.
- machine readable media can include: non-volatile memory, such as semiconductor memory devices (e.g., Electrically Programmable Read-Only Memory (EPROM), Electrically Erasable Programmable Read-Only Memory (EEPROM)) and flash memory devices; magnetic disks, such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks.
- non-volatile memory such as semiconductor memory devices (e.g., Electrically Programmable Read-Only Memory (EPROM), Electrically Erasable Programmable Read-Only Memory (EEPROM)) and flash memory devices
- EPROM Electrically Programmable Read-Only Memory
- EEPROM Electrically Erasable Programmable Read-Only Memory
- flash memory devices e.g., electrically Erasable Programmable Read-Only Memory (EEPROM)
- EPROM Electrically Programmable Read-Only Memory
- EEPROM Electrically Erasable Programmable Read-Only Memory
- flash memory devices e.g., electrically Era
- the instructions 824 can further be transmitted or received over a communications network 826 using a transmission medium via the network interface device 820 utilizing any one of a number of transfer protocols (e.g., frame relay, internet protocol (IP), transmission control protocol (TCP), user datagram protocol (UDP), hypertext transfer protocol (HTTP), etc.).
- Example communication networks can include a local area network (LAN), a wide area network (WAN), a packet data network (e.g., the Internet), mobile telephone networks (e.g., cellular networks), Plain Old Telephone (POTS) networks, and wireless data networks (e.g., Institute of Electrical and Electronics Engineers (IEEE) 802 .
- the network interface device 820 can include one or more physical jacks (e.g., Ethernet, coaxial, or phone jacks) or one or more antennas to connect to the communications network 826 .
- the network interface device 820 can include a plurality of antennas to wirelessly communicate using at least one of single-input multiple-output (SIMO), multiple-input multiple-output (MIMO), or multiple-input single-output (MISO) techniques.
- SIMO single-input multiple-output
- MIMO multiple-input multiple-output
- MISO multiple-input single-output
- Hearing assistance devices typically include at least one enclosure or housing, a microphone, hearing assistance device electronics including processing electronics, and a speaker or “receiver.”
- Hearing assistance devices can include a power source, such as a battery.
- the battery can be rechargeable.
- multiple energy sources can be employed.
- the microphone is optional.
- the receiver is optional.
- Antenna configurations can vary and can be included within an enclosure for the electronics or be external to an enclosure for the electronics.
- digital hearing aids include a processor.
- programmable gains can be employed to adjust the hearing aid output to a wearer's particular hearing impairment.
- the processor can be a digital signal processor (DSP), microprocessor, microcontroller, other digital logic, or combinations thereof.
- DSP digital signal processor
- the processing can be done by a single processor, or can be distributed over different devices.
- the processing of signals referenced in this application can be performed using the processor or over different devices.
- Processing can be done in the digital domain, the analog domain, or combinations thereof.
- Processing can be done using subband processing techniques. Processing can be done using frequency domain or time domain approaches. Some processing can involve both frequency and time domain aspects.
- drawings can omit certain blocks that perform frequency synthesis, frequency analysis, analog-to-digital conversion, digital-to-analog conversion, amplification, buffering, and certain types of filtering and processing.
- the processor is adapted to perform instructions stored in one or more memories, which can or cannot be explicitly shown. Various types of memory can be used, including volatile and nonvolatile forms of memory.
- the processor or other processing devices execute instructions to perform a number of signal processing tasks. Such embodiments can include analog components in communication with the processor to perform signal processing tasks, such as sound reception by a microphone, or playing of sound using a receiver (i.e., in applications where such transducers are used).
- different realizations of the block diagrams, circuits, and processes set forth herein can be created by one of skill in the art without departing from the scope of the present subject matter.
- the wireless communications can include standard or nonstandard communications.
- standard wireless communications include, but not limited to, BluetoothTM, low energy Bluetooth, IEEE 802.11(wireless LANs), 802.15 (WPANs), and 802.16 (WiMAX).
- Cellular communications can include, but not limited to, CDMA, GSM, ZigBee, and ultra-wideband (UWB) technologies.
- the communications are radio frequency communications.
- the communications are optical communications, such as infrared communications.
- the communications are inductive communications.
- the communications are ultrasound communications.
- the wireless communications support a connection from other devices.
- Such connections include, but are not limited to, one or more mono or stereo connections or digital connections having link protocols including, but not limited to 802.3 (Ethernet), 802.4, 802.5, USB, ATM, Fibre-channel, Firewire or 1394, InfiniBand, or a native streaming interface.
- link protocols including, but not limited to 802.3 (Ethernet), 802.4, 802.5, USB, ATM, Fibre-channel, Firewire or 1394, InfiniBand, or a native streaming interface.
- link protocols including, but not limited to 802.3 (Ethernet), 802.4, 802.5, USB, ATM, Fibre-channel, Firewire or 1394, InfiniBand, or a native streaming interface.
- such connections include all past and present link protocols. It is also contemplated that future versions of these protocols and new protocols can be employed without departing from the scope of the present subject matter.
- the present subject matter is used in hearing assistance devices that are configured to communicate with mobile phones.
- the hearing assistance device can be operable to perform one or more of the following: answer incoming calls, hang up on calls, and/or provide two way telephone communications.
- the present subject matter is used in hearing assistance devices configured to communicate with packet-based devices.
- the present subject matter includes hearing assistance devices configured to communicate with streaming audio devices.
- the present subject matter includes hearing assistance devices configured to communicate with Wi-Fi devices.
- the present subject matter includes hearing assistance devices capable of being controlled by remote control devices.
- hearing assistance devices can embody the present subject matter without departing from the scope of the present disclosure.
- the devices depicted in the figures are intended to demonstrate the subject matter, but not necessarily in a limited, exhaustive, or exclusive sense. It is also understood that the present subject matter can be used with a device designed for use in the right ear or the left ear or both ears of the wearer.
- the present subject matter can be employed in hearing assistance devices, such as headsets, headphones, and similar hearing devices.
- hearing assistance devices including hearing aids, including but not limited to, behind-the-ear (BTE), in-the-ear (ITE), in-the-canal (ITC), receiver-in-canal (RIC), or completely-in-the-canal (CIC) type hearing aids.
- BTE behind-the-ear
- ITE in-the-ear
- ITC in-the-canal
- RIC receiver-in-canal
- CIC completely-in-the-canal
- hearing assistance devices including but not limited to, behind-the-ear (BTE), in-the-ear (ITE), in-the-canal (ITC), receiver-in-canal (RIC), or completely-in-the-canal (CIC) type hearing aids.
- BTE behind-the-ear
- ITE in-the-ear
- ITC in-the-canal
- RIC receiver-in-canal
- CIC completely-in-the-canal
- hearing assistance devices including but not limited to, behind-the-ear (BTE), in
- the present subject matter can also be used in hearing assistance devices generally, such as cochlear implant type hearing devices and such as deep insertion devices having a transducer, such as a receiver or microphone, whether custom fitted, standard fitted, open fitted and/or occlusive fitted. It is understood that other hearing assistance devices not expressly stated herein can be used in conjunction with the present subject matter.
- Method examples described herein can be machine or computer-implemented at least in part. Some examples can include a computer-readable medium or machine-readable medium encoded with instructions operable to configure an electronic device to perform methods as described in the above examples.
- An implementation of such methods can include code, such as microcode, assembly language code, a higher-level language code, or the like. Such code can include computer readable instructions for performing various methods. The code can form portions of computer program products. Further, in an example, the code can be tangibly stored on one or more volatile, non-transitory, or non-volatile tangible computer-readable media, such as during execution or at other times.
- Examples of these tangible computer-readable media can include, but are not limited to, hard disks, removable magnetic disks, removable optical disks (e.g., compact disks and digital video disks), magnetic cassettes, memory cards or sticks, random access memories (RAMs), read only memories (ROMs), and the like.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- General Health & Medical Sciences (AREA)
- Otolaryngology (AREA)
- Neurosurgery (AREA)
- Computer Networks & Wireless Communication (AREA)
- Computational Linguistics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Mobile Radio Communication Systems (AREA)
Abstract
Description
- This patent application claims the benefit of U.S. Provisional Patent Application No. 62/068,404, filed Oct. 24, 2014, entitled “PACKET LOSS CONCEALMENT TECHNIQUES FOR PHONE-TO-HEARING-AID STREAMING”, which is incorporated by reference herein in full.
- Disclosed herein are devices and methods for packet loss concealment in device-to-device streaming, and in particular for streaming to a hearing aid.
- Adaptive differential pulse-code modulation (ADPCM) is used in the context of audio streaming to improve hearing assistance device functionality when streaming from a remote device to a hearing assistance device. ADPCM has a low latency, good quality, a low bitrate, and low computational requirements. However, one drawback to using ADPCM is that it is negatively affected by packet-loss. The negative impact on resulting audio quality when packet-loss occurs with ADPCM is not limited to the dropped packet, but also up to several dozens of milliseconds after the dropped packet.
- When using ADPCM, the encoder and the decoder both maintain a certain state based on the encoded signal which, under normal operation and after initial convergence, is the same. A packet drop causes the encoder and the decoder states to depart from one another, and the decoder state will take time to converge back to the encoder state once valid data is available again after a drop.
- Packet-loss-concealment (PLC) techniques mitigate the error caused by packet loss. While there are multiple single-channel PLC techniques currently used, they are often slow and costly in terms of instructions per second used, and thus can be infeasible in a hearing assistance device setting.
- Disclosed herein are devices and methods for packet loss concealment for streaming to a hearing assistance device. In various embodiments, a method for packet loss concealment includes receiving a first frame at a hearing assistance device, determining, at the hearing assistance device, that a second frame was not received within a predetermined time, and determining a first set of sequential samples that match the first frame. The method can include cross-fading the first frame and the first set of sequential samples to create a first cross-faded frame and extrapolating a third frame to replace the second frame using the first set of sequential samples and an autoregressive model.
- This Summary is an overview of some of the teachings of the present application and not intended to be an exclusive or exhaustive treatment of the present subject matter. Further details about the present subject matter are found in the detailed description and appended claims. The scope of the present invention is defined by the appended claims and their legal equivalents.
- In the drawings, which are not necessarily drawn to scale, like numerals can describe similar components in different views. Like numerals having different letter suffixes can represent different instances of similar components. The drawings illustrate generally, by way of example, but not by way of limitation, various embodiments discussed in the present document.
-
FIG. 1 illustrates generally a graph showing effects of a packet loss on an encoded signal in accordance with some embodiments of the present subject matter. -
FIG. 2 illustrates generally a graph showing an autoregressive-based technique for extrapolation of an audio signal in accordance with some embodiments of the present subject matter. -
FIG. 3 illustrates generally graphs of cross-faded waveforms with filled in frames in accordance with some embodiments of the present subject matter. -
FIGS. 4A-4D illustrate generally frame by frame replacements in accordance with some embodiments of the present subject matter. -
FIGS. 5A-5D illustrate generally cross-faded samples for replacement frames in accordance with some embodiments of the present subject matter. -
FIG. 6 illustrates generally a hearing assistance device in accordance with some embodiments of the present subject matter. -
FIG. 7 illustrates generally a flowchart for packet loss concealment in accordance with some embodiments of the present subject matter. -
FIG. 8 illustrates generally an example of a block diagram of a machine upon which any one or more of the techniques discussed herein can perform in accordance with some embodiments of the present subject matter. - The following detailed description of the present subject matter refers to subject matter in the accompanying drawings which show, by way of illustration, specific aspects and embodiments in which the present subject matter can be practiced. These embodiments are described in sufficient detail to enable those skilled in the art to practice the present subject matter. References to “an”, “one”, or “various” embodiments in this disclosure are not necessarily to the same embodiment, and such references contemplate more than one embodiment. The following detailed description is demonstrative and not to be taken in a limiting sense. The scope of the present subject matter is defined by the appended claims, along with the full scope of legal equivalents to which such claims are entitled.
- Adaptive differential pulse-code modulation (ADPCM) is useful for improving hearing assistance device functionality when streaming from device-to-device, but is particularly susceptible to packet loss issues. Packet-loss-concealment (PLC) techniques are therefore used to mitigate the error caused by packet loss.
- In an example, a hearing assistance device can use a codec to decode audio. For example, a G.722 audio codec according to an International Telecommunication Union Telecommunication Standardization Sector (ITU-T) standard can be used in an ADPCM system to decode audio. Other codec apparatus and methods can be used without departing from the scope of the present subject matter.
-
FIG. 1 illustrates generally agraph 100 showing effects of a packet loss on an encoded signal in accordance with some embodiments of the present subject matter. Thegraph 100 includes signals encoded using an ADPCM-based codec. Thegraph 100 includes a first encodedsignal 102 over time with no packet loss and asecond signal 104 over time with a packet loss. The packet loss is highlighted on thegraph 100 of thesecond signal 104 with abox 106. As is evident from thesecond signal 104, the packet loss affects the signal not only at, but also after, the packet loss highlighted bybox 106. The effect of the packet loss is not confined to just the window of the lost packets, but beyond. In an example, a PLC technique includes artificially “filling” the blank caused by the packet drop with a filler signal that is thought to naturally extend the signal with one that is similar enough to the original signal that was encoded at the encoder. This replacement signal is sometimes called a concealment frame, and there exists a wide range of solutions to generate it. - In an ADPCM-based system, the mere generation of a filler does not always remove the state inconsistencies, which can potentially create long-lasting and highly audible artifacts. In an example, re-encoding the synthetic concealment frame at the decoder can be used to avoid artifacts. This allows for the decoder's state to keep updating, and with an appropriate “filler” signal, its state will not drift too far from the encoder's state at the end of the frame.
- One method to generate a concealment frame is the so-called “pitch-repetition” method. In an example, the last 20 milliseconds of audio are stored in a circular buffer, and from this data an estimate of the pitch is tracked (essentially by doing a brute-force search from circular delays of 5 to 15 milliseconds, using a maximum correlation-based method). When a dropped frame is detected by the decoder, a 10-millisecond concealment frame is generated synthetically from a repetition of the last pitch period. Some windowing and overlap-add can also be employed to ensure smooth transitions between the real and synthetic signals, and slight structural changes in the decoder can be made. However, repeating the last full pitch period can have a negative impact in quality for erroneously detected pitch. Moreover, repeating the exact same pitch period for more than 10 milliseconds (e.g. if multiple frames are dropped) will also typically sound very synthetic, and even more so if there are significant discontinuities at the frame borders (e.g., due to badly estimated pitch). To address the latter issue, if additional latency is acceptable then some type of cross-fading between the filler and the previous or the next correct frame(s) can be introduced.
- In an example, an “enhanced” pitch repetition method can be used, where multiple additional tricks are used to improve the resulting quality, including: using the so-called “ringing signal” (e.g., the first few samples coming out of the synthesis filter with zero-input, as seen right after the beginning of
box 106 at the bottom ofFIG. 1 ) to overlap-adds it to a periodic waveform extension of the previous frame; time-warping and re-phasing (to improve alignment between concealment frame and the previous/next valid data); classification of the previous frame just before the loss as either voice-like or noise-like; gradual gain attenuation as more and more frames are lost (so that after 60 milliseconds, the output is in fact practically muted), and other tricks that can be used if possible, such as re-alignment once the next frame is present. - In another example, a low-complexity PLC technique can be used, which can include a tradeoff of lower quality for less complexity. It can introduce short and long-term predictors that it uses to generate the extrapolated signal, it can “over-extrapolate” to cover the next available frame for cross-fading, or the like. It can include higher quality for music signals.
- In yet another example, in order to minimize state disparities following a frame loss, additional information can be sent with a frame, indicative of the state of the encoder. This greatly speeds up the re-synchronizing process and reduces artifacts when a correct frame is received again. This example technique can also employ pitch repetition to fill the blank left during the frame loss. However there is an additional bandwidth cost with this method and it is not compatible with typical ADPCM encoders.
- In an example, a technique for packet loss concealment can include all-pole modelling an arbitrary length of buffered audio data for packet loss concealment. The technique can include an appropriate filter initialization, and letting the resulting filter “ring” for the duration of the blank. In various examples, the Burg method can be used for a high output quality, and the Yule-Walker method can be used for a lighter computational burden to calculate or estimate autoregressive parameters. There are many methods to determine autoregressive parameters, and other methods besides Yule-Walker method and Burg method can be used without departing from the scope of the present subject matter. After the waveform has been generated, the internal state of the decoder can be updated by re-encoding it. In an example, this re-encoding is well indicated as the extrapolated waveform is typically very close to the true waveform (at least for the first few milliseconds of data), and therefore the decoder's state is adequately updated.
-
FIG. 2 illustrates generally agraph 200 showing an autoregressive-based technique for extrapolation of an audio signal in accordance with some embodiments of the present subject matter. In a specific example,graph 200 can include an autoregressive extrapolation of a 16 kHz audio signal. The audio is interrupted at cut-off point 202, at which point the extrapolation occurs. In this example, the autoregressive (AR) order is 100, and the first 200 samples are used to determine the AR model. The extrapolation length is 500 samples. After the cut-off point 202, the extrapolatedsignal 204 and thetrue signal 206 are shown. The extrapolatedsignal 204 closely tracks thetrue signal 206. In a packet loss situation, thetrue signal 206 would be missing, and the extrapolatedsignal 204 would be substituted for the missingtrue signal 206. - In one embodiment, when the model order is large-enough and well initialized, this technique produces high quality extrapolation and can be easily augmented to “latch” onto future available data by cross-fading. This technique also does not induce additional latency since the forward extension is very natural-sounding for several dozens of milliseconds.
- This technique can have a large computational complexity involved in determining autoregressive parameters. The technique can include large AR model orders (e.g., more than 50), which makes it difficult to include in current fixed-point processors. Using the large AR model orders, the technique produces results that are of markedly higher quality than previous techniques. Estimations techniques, increased processor capabilities, and modeling data can be used to implement this technique in hearing assistance devices, including at lower orders using less computational complexity.
- Various embodiments include concealing a source of artifacts and non-naturalness in discontinuities in the reconstructed signal as described below. One component of the technique includes pattern matching, and includes further operations to enforce a smooth and natural-sounding extrapolation. An assumption can be made that the decoder is aware of whether or not the next frame is valid or dropped, and an “extrapolated data” buffer can be available.
-
FIG. 3 illustrates generally graphs 300A-300D of cross-faded waveforms with filled in frames in accordance with some embodiments of the present subject matter. For an incoming frame, the technique can include, if thecurrent frame 304 is valid (i.e., not dropped), first decoding the frame normally (without playing it back yet). If theprevious frame 302 was not valid (i.e., dropped), the current frame can be cross-faded using the data just decoded above with the data present in the extrapolated data buffer over the entire frame. In the example shown in graph 300A, theprevious frame 302 is valid, so no cross-fading need take place. The technique can continue by determining if thenext frame 306 is dropped, and if so (as is the case, as shown in graph 300A), then beginning a specific form of extrapolation that will generate artificial data over the 2 next future frames (e.g., 308 and 310), and modify the current frame 304 (as shown ingraph 300B, where thecurrent frame 304 is cross-faded with a matching segment from history). This can be done whether thenext frame 306 is freshly decoded from a valid frame or from past extrapolations. If thecurrent frame 304 is dropped, the technique can proceed without additional decoding and playback the audio data currently available in the “extrapolated data” buffer at the correct indices. -
FIG. 3 shows a technique for smoothly extending a waveform fromtime index 200 to 500 (e.g., overframes graphs graphs frame 306. For example,frame 308 can include a matched segment that immediately follows the matched segment for extrapolatedframe 306 andframe 310 can include a matched segment following the one used forframe 308. By using following matched segments, the computation to determine the matched segments forframes frame 306. - In an example, an audio extrapolation technique can minimize artifacts when used in the context described above. The technique of audio extrapolation, described in details in the paragraphs below, can include extrapolating using overlapping blocks, with multiple and far-reaching cross-fadings when a new block is added. In an example, such as when using codec G.722, a match on the lower audio bands can be extrapolated, and the matching positions can be replicated in the higher audio bands without a noticeable quality impact.
- An extrapolation technique can include operations to extend an existing audio segment of length N to a length N+T. An operation can include decomposing T=PL where L can represent a certain base block length and P is an integer, and then extrapolating one block at a time while modifying the length—L block just before the current blank to maximize smoothness. The technique can include operations to perform some arbitrary type of pattern matching to find a waveform in history that resembles the one in positions N−
L+ 1 to N. The search can begin from N−2L+1 and go backwards in history. Once a matching waveform is found, the current data in N−L+ 1 to N can be cross-faded with the matching waveform, and the L samples in history that follow the best match can be used to fill the positions N+1 to N+L. The current blank position can then include N+L+1. The exact same procedure can be repeated P times until data up to N+T is covered. In an example, there are P matching tasks and P cross-fading operations to fill the audio up until N+T. -
FIGS. 4A-4D illustrate generally frame by frame replacements in accordance with some embodiments of the present subject matter. In an example, the following frames (e.g., frames 402A-D, 404A-D, and 406A-D) are received and the frames in the places of D and E are dropped: -
[A], [B], [C], [Ø], [Ø], [F], [G], . . . - Each frame can contain 3 samples:
-
[a1, a2, a3], [b1, b2, b3], [c1, c2, c3], [0, 0, 0], [0, 0, 0], [f1, f2, f3], [g1, g2, g3], . . . - Frames A, B, and C, are decoded normally, and before playing C back, the decoder is informed that the next frame (i.e., frame D) is dropped. The technique includes finding a match for frame C from its saved history, testing several positions up to the beginning of frame B. A new frame can be constructed C0. In an example, an appropriate match includes the samples [a3, b1, b2] (see
FIG. 4D ). Frame C can be replaced with the cross-fading of C and C0, yielding frame {tilde over (C)}=C{circle around (X)}C0, which can then be played back. From here, a new frame D can be constructed, namely frame {tilde over (D)}=[b3, c1, c2]=[{tilde over (d)}1, {tilde over (d)}2, {tilde over (d)}3] (seeFIG. 4C ) and that can be placed in the extrapolated data buffer. - As the technique proceeds to the frame where {tilde over (D)} is currently present, the decoder can also be informed that the next frame is dropped again. This time, a match from history to {tilde over (E)}=[c3, d1, d2] is determined, and the decoder is allowed to look all the way from the beginning of {tilde over (D)}. The history has slightly changed, and matches within the new history can be slightly different. In an example, the best match to [c3, d1, d2] can include [{tilde over (c)}3, {tilde over (d)}1, {tilde over (d)}2], which is exactly one frame forward from the match for new frame D0 [b3, {tilde over (c)}1, {tilde over (c)}2]. Cross-fading of [{tilde over (c)}3, {tilde over (d)}1, {tilde over (d)}2] and [c3, d1, d2] can be performed, which can ensure maximal smoothness of the extended waveform. In an example, the cross-fading operation can either partially be necessary (if the first few samples to be cross-faded are equal), or not necessary at all (if all samples are equal). As shown in
FIGS. 4A and 4D , if the next frame F is valid (i.e., not dropped), it can be cross-faded with the next sequential match samples after those used for frame E. If frame G is also valid, then the technique can continue with no cross-fading for frame G. -
FIGS. 5A-5D illustrate generally cross-faded samples for replacement frames in accordance with some embodiments of the present subject matter. As shown inFIG. 4B , the original frames can be replaced by new frames that include extrapolated or original samples cross-faded with matched samples: -
[a1, a2, a3], [b1, b2, b3], [c1′, c2′, c3′], [d1′, d2′, d3′], [e1′, e2′, e3′], [f1′, f2′, f3′], [g1, g2, g3], . . . -
FIG. 5A shows frame {tilde over (C)} as: [c1′, c2′, c3′] including the original samples that are cross-faded to result in {tilde over (C)}. Similarly,FIGS. 5B-5C show frames {tilde over (D)}, {tilde over (E)}, and {tilde over (F)} respectively, including the original samples that are cross-faded to result in those frames. In an example, an assumption can be made that one search is sufficient, and the next matches can be located one frame ahead from the initial match position, which can greatly reduce computational burden. In another example, while the cross-fade operations take place over a whole segment, the pattern match can be done over a central portion of a segment, where the cross-fading will have the most effect. - In practice, and in the context of packet loss concealment technique as described above, T can be equal to 2L where L is the frame length, and if the computational requirements allow, other combinations can be used (e.g., 4L′ where L′ is half of L, etc.). The extension can cover an extra frame, which can correspond to a non-dropped frame. If the extra frame is a non-dropped frame, another full-frame cross-fading operation can take place (e.g., for frame F above).
- In the technique described above, multiple iterations can be used, instead of a single search far enough back to cover enough frames. The multiple iterations can increase the chances of finding a highly correlated concealment frame since the search can be started close to the dropped frame (e.g., due to short-time stationarity assumptions). The multiple iterations can also cover situations where multiple frames are dropped in the future or in a row. Using multiple iterations also decreases the computational complexity of determining subsequent searches through approximation to subsequent samples after a first set of matching samples is determined for a first cross-faded frame.
-
FIG. 6 illustrates generally ahearing assistance device 602 in accordance with some embodiments of the present subject matter. In an example, thehearing assistance device 602 can include awireless receiver 604 to receive a first frame at thehearing assistance device 602. Thereceiver 604 can be in communication, such as using wireless signals, with a mobile device or music player, such as a phone. - The
hearing assistance device 602 can include aprocessor 606 to determine that a second frame was not received by the receiver within a predetermined time and determine a first set of sequential samples that match the first frame. Theprocessor 606 can cross-fade the first frame and the first set of sequential samples to create a first cross-faded frame and extrapolate a third frame to replace the second frame using the first set of sequential samples and an autoregressive model. Theprocessor 606 can determine that a fourth frame was not received within a predetermined time. In an example, theprocessor 606 can cross-fade the third frame with the second set of samples to create a second cross-faded frame to replace the fourth frame. In another example, theprocessor 606 can determine parameters for the autoregressive model using one of a Burg method or a Yule-Walker method. Other methods to determine autoregressive parameters can be used without departing from the scope of the present subject matter. In yet another example, theprocessor 606 can encode the first frame using an adaptive differential pulse-code modulation. - The
hearing assistance device 602 can include aspeaker 608 to play an audio signal corresponding to the first cross-faded frame. In an example, thespeaker 608 can play the third frame after the first cross-faded frame. In another example, thespeaker 608 can play the second cross-faded frame. Thehearing assistance device 602 can includememory 610 to store a buffer including samples, flags (e.g., a dropped frame flag), etc. - In an example, the third frame includes a second set of samples, and wherein the second set of samples includes samples starting one frame after the first set of sequential samples. In another example, the autoregressive model includes a model order of at least 50.
-
FIG. 7 illustrates generally aflowchart 700 for packet loss concealment in accordance with some embodiments of the present subject matter. Theflowchart 700 can include a method for packet loss concealment, the method comprising anoperation 702 to receive a first frame at a hearing assistance device. The first frame can be received from a mobile device, such as a phone, a tablet, a dedicated music player, or the like. - The method can include an
operation 704 to determine, at the hearing assistance device, that a second frame was not received within a predetermined time. The method can include anoperation 706 to determine a first set of sequential samples that match the first frame. The method can include anoperation 708 to cross-fade the first frame and the first set of sequential samples to create a first cross-faded frame. The method can include anoperation 710 to extrapolate a third frame to replace the second frame using the first set of sequential samples and an autoregressive model. - The method can further include operations to play the first cross-faded frame or play the third frame after the first cross-faded frame. The method can include an operation to determine, at the hearing assistance device, that a fourth frame was not received within a predetermined time. The method can include an operation to cross-fade the third frame with the second set of samples to create a second cross-faded frame to replace the fourth frame. The method can include an operation to play the second cross-faded frame. In an example, the method includes an operation to determine parameters for the autoregressive model using a Burg method, a Yule-Walker method, and/or other methods for determining autoregressive parameters. The method can include an operation to encode the first frame using an adaptive differential pulse-code modulation (ADPCM).
- The third frame can include a second set of samples, and wherein the second set of samples include samples starting one frame after the first set of sequential samples. The autoregressive model can include a model order, such as an order of 50 or greater.
- In another example, any type of pattern matching can be used with the techniques described herein. The multiple cross-fadings performed over entire segments can include sound that is smooth and natural compared to other packet loss concealment techniques. In yet another example, two or more of the techniques described herein can be used together for packet loss concealment.
-
FIG. 8 illustrates generally an example of a block diagram of amachine 800 upon which any one or more of the techniques discussed herein can perform in accordance with some embodiments of the present subject matter. In various embodiments, themachine 800 can operate as a standalone device or can be connected (e.g., networked) to other machines. In a networked deployment, themachine 800 can operate in the capacity of a server machine, a client machine, or both in server-client network environments. In an example, themachine 800 can act as a peer machine in peer-to-peer (P2P) (or other distributed) network environment. Themachine 800 can be a personal computer (PC), a tablet PC, a set-top box (STB), a personal digital assistant (PDA), a mobile telephone, a web appliance, a network router, switch or bridge, or any machine capable of executing instructions (sequential or otherwise) that specify actions to be taken by that machine.Machine 800 can includeprocessor 606 of hearingassistance device 602 inFIG. 6 . Further, while only a single machine is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute a set (or multiple sets) of instructions to perform any one or more of the methods described herein, such as cloud computing, software as a service (SaaS), other computer cluster configurations. - Examples, as described herein, can include, or can operate on, logic or a number of components, modules, or mechanisms. Modules are tangible entities (e.g., hardware) capable of performing specified operations when operating. A module includes hardware. In an example, the hardware can be specifically configured to carry out a specific operation (e.g., hardwired). In an example, the hardware can include configurable execution units (e.g., transistors, circuits, etc.) and a computer readable medium containing instructions, where the instructions configure the execution units to carry out a specific operation when in operation. The configuring can occur under the direction of the executions units or a loading mechanism. Accordingly, the execution units are communicatively coupled to the computer readable medium when the device is operating. In this example, the execution units can be a member of more than one module. For example, under operation, the execution units can be configured by a first set of instructions to implement a first module at one point in time and reconfigured by a second set of instructions to implement a second module.
- Machine (e.g., computer system) 800 can include a hardware processor 802 (e.g., a central processing unit (CPU), a graphics processing unit (GPU), a hardware processor core, or any combination thereof), a
main memory 804 and astatic memory 806, some or all of which can communicate with each other via an interlink (e.g., bus) 808. Themachine 800 can further include adisplay unit 810, an alphanumeric input device 812 (e.g., a keyboard), and a user interface (UI) navigation device 814 (e.g., a mouse). In an example, thedisplay unit 810,alphanumeric input device 812 andUI navigation device 814 can be a touch screen display. Themachine 800 can additionally include a storage device (e.g., drive unit) 816, a signal generation device 818 (e.g., a speaker), anetwork interface device 820, and one ormore sensors 821, such as a global positioning system (GPS) sensor, compass, accelerometer, or other sensor. Themachine 800 can include anoutput controller 828, such as a serial (e.g., universal serial bus (USB), parallel, or other wired or wireless (e.g., infrared (IR), near field communication (NFC), etc.) connection to communicate or control one or more peripheral devices (e.g., a printer, card reader, etc.). - The
storage device 816 can include a machinereadable medium 822 that is non-transitory on which is stored one or more sets of data structures or instructions 824 (e.g., software) embodying or utilized by any one or more of the techniques or functions described herein. Theinstructions 824 can also reside, completely or at least partially, within themain memory 804, withinstatic memory 806, or within thehardware processor 802 during execution thereof by themachine 800. In an example, one or any combination of thehardware processor 802, themain memory 804, thestatic memory 806, or thestorage device 816 can constitute machine readable media. - While the machine
readable medium 822 is illustrated as a single medium, the term “machine readable medium” can include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) configured to store the one ormore instructions 824. - The term “machine readable medium” can include any medium that is capable of storing, encoding, or carrying instructions for execution by the
machine 800 and that cause themachine 800 to perform any one or more of the techniques of the present disclosure, or that is capable of storing, encoding or carrying data structures used by or associated with such instructions. Non-limiting machine readable medium examples can include solid-state memories, and optical and magnetic media. Specific examples of machine readable media can include: non-volatile memory, such as semiconductor memory devices (e.g., Electrically Programmable Read-Only Memory (EPROM), Electrically Erasable Programmable Read-Only Memory (EEPROM)) and flash memory devices; magnetic disks, such as internal hard disks and removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks. - The
instructions 824 can further be transmitted or received over acommunications network 826 using a transmission medium via thenetwork interface device 820 utilizing any one of a number of transfer protocols (e.g., frame relay, internet protocol (IP), transmission control protocol (TCP), user datagram protocol (UDP), hypertext transfer protocol (HTTP), etc.). Example communication networks can include a local area network (LAN), a wide area network (WAN), a packet data network (e.g., the Internet), mobile telephone networks (e.g., cellular networks), Plain Old Telephone (POTS) networks, and wireless data networks (e.g., Institute of Electrical and Electronics Engineers (IEEE) 802.11 family of standards known as Wi-Fi®, IEEE 802.16 family of standards known as WiMax®), IEEE 802.15.4 family of standards, peer-to-peer (P2P) networks, among others. In an example, thenetwork interface device 820 can include one or more physical jacks (e.g., Ethernet, coaxial, or phone jacks) or one or more antennas to connect to thecommunications network 826. In an example, thenetwork interface device 820 can include a plurality of antennas to wirelessly communicate using at least one of single-input multiple-output (SIMO), multiple-input multiple-output (MIMO), or multiple-input single-output (MISO) techniques. - Hearing assistance devices typically include at least one enclosure or housing, a microphone, hearing assistance device electronics including processing electronics, and a speaker or “receiver.” Hearing assistance devices can include a power source, such as a battery. In various embodiments, the battery can be rechargeable. In various embodiments multiple energy sources can be employed. It is understood that in various embodiments the microphone is optional. It is understood that in various embodiments the receiver is optional. It is understood that variations in communications protocols, antenna configurations, and combinations of components can be employed without departing from the scope of the present subject matter. Antenna configurations can vary and can be included within an enclosure for the electronics or be external to an enclosure for the electronics. Thus, the examples set forth herein are intended to be demonstrative and not a limiting or exhaustive depiction of variations.
- It is understood that digital hearing aids include a processor. In digital hearing aids with a processor, programmable gains can be employed to adjust the hearing aid output to a wearer's particular hearing impairment. The processor can be a digital signal processor (DSP), microprocessor, microcontroller, other digital logic, or combinations thereof. The processing can be done by a single processor, or can be distributed over different devices. The processing of signals referenced in this application can be performed using the processor or over different devices. Processing can be done in the digital domain, the analog domain, or combinations thereof. Processing can be done using subband processing techniques. Processing can be done using frequency domain or time domain approaches. Some processing can involve both frequency and time domain aspects. For brevity, in some examples drawings can omit certain blocks that perform frequency synthesis, frequency analysis, analog-to-digital conversion, digital-to-analog conversion, amplification, buffering, and certain types of filtering and processing. In various embodiments the processor is adapted to perform instructions stored in one or more memories, which can or cannot be explicitly shown. Various types of memory can be used, including volatile and nonvolatile forms of memory. In various embodiments, the processor or other processing devices execute instructions to perform a number of signal processing tasks. Such embodiments can include analog components in communication with the processor to perform signal processing tasks, such as sound reception by a microphone, or playing of sound using a receiver (i.e., in applications where such transducers are used). In various embodiments, different realizations of the block diagrams, circuits, and processes set forth herein can be created by one of skill in the art without departing from the scope of the present subject matter.
- Various embodiments of the present subject matter support wireless communications with a hearing assistance device. In various embodiments the wireless communications can include standard or nonstandard communications. Some examples of standard wireless communications include, but not limited to, Bluetooth™, low energy Bluetooth, IEEE 802.11(wireless LANs), 802.15 (WPANs), and 802.16 (WiMAX). Cellular communications can include, but not limited to, CDMA, GSM, ZigBee, and ultra-wideband (UWB) technologies. In various embodiments, the communications are radio frequency communications. In various embodiments the communications are optical communications, such as infrared communications. In various embodiments, the communications are inductive communications. In various embodiments, the communications are ultrasound communications. Although embodiments of the present system can be demonstrated as radio communication systems, it is possible that other forms of wireless communications can be used. It is understood that past and present standards can be used. It is also contemplated that future versions of these standards and new future standards can be employed without departing from the scope of the present subject matter.
- The wireless communications support a connection from other devices. Such connections include, but are not limited to, one or more mono or stereo connections or digital connections having link protocols including, but not limited to 802.3 (Ethernet), 802.4, 802.5, USB, ATM, Fibre-channel, Firewire or 1394, InfiniBand, or a native streaming interface. In various embodiments, such connections include all past and present link protocols. It is also contemplated that future versions of these protocols and new protocols can be employed without departing from the scope of the present subject matter.
- In various embodiments, the present subject matter is used in hearing assistance devices that are configured to communicate with mobile phones. In such embodiments, the hearing assistance device can be operable to perform one or more of the following: answer incoming calls, hang up on calls, and/or provide two way telephone communications. In various embodiments, the present subject matter is used in hearing assistance devices configured to communicate with packet-based devices. In various embodiments, the present subject matter includes hearing assistance devices configured to communicate with streaming audio devices. In various embodiments, the present subject matter includes hearing assistance devices configured to communicate with Wi-Fi devices. In various embodiments, the present subject matter includes hearing assistance devices capable of being controlled by remote control devices.
- It is further understood that different hearing assistance devices can embody the present subject matter without departing from the scope of the present disclosure. The devices depicted in the figures are intended to demonstrate the subject matter, but not necessarily in a limited, exhaustive, or exclusive sense. It is also understood that the present subject matter can be used with a device designed for use in the right ear or the left ear or both ears of the wearer.
- The present subject matter can be employed in hearing assistance devices, such as headsets, headphones, and similar hearing devices.
- The present subject matter is demonstrated for hearing assistance devices, including hearing aids, including but not limited to, behind-the-ear (BTE), in-the-ear (ITE), in-the-canal (ITC), receiver-in-canal (RIC), or completely-in-the-canal (CIC) type hearing aids. It is understood that behind-the-ear type hearing aids can include devices that reside substantially behind the ear or over the ear. Such devices can include hearing aids with receivers associated with the electronics portion of the behind-the-ear device, or hearing aids of the type having receivers in the ear canal of the user, including but not limited to receiver-in-canal (RIC) or receiver-in-the-ear (RITE) designs. The present subject matter can also be used in hearing assistance devices generally, such as cochlear implant type hearing devices and such as deep insertion devices having a transducer, such as a receiver or microphone, whether custom fitted, standard fitted, open fitted and/or occlusive fitted. It is understood that other hearing assistance devices not expressly stated herein can be used in conjunction with the present subject matter.
- This application is intended to cover adaptations or variations of the present subject matter. It is to be understood that the above description is intended to be illustrative, and not restrictive. The scope of the present subject matter should be determined with reference to the appended claims, along with the full scope of legal equivalents to which such claims are entitled.
- Method examples described herein can be machine or computer-implemented at least in part. Some examples can include a computer-readable medium or machine-readable medium encoded with instructions operable to configure an electronic device to perform methods as described in the above examples. An implementation of such methods can include code, such as microcode, assembly language code, a higher-level language code, or the like. Such code can include computer readable instructions for performing various methods. The code can form portions of computer program products. Further, in an example, the code can be tangibly stored on one or more volatile, non-transitory, or non-volatile tangible computer-readable media, such as during execution or at other times. Examples of these tangible computer-readable media can include, but are not limited to, hard disks, removable magnetic disks, removable optical disks (e.g., compact disks and digital video disks), magnetic cassettes, memory cards or sticks, random access memories (RAMs), read only memories (ROMs), and the like.
Claims (22)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/854,704 US9706317B2 (en) | 2014-10-24 | 2015-09-15 | Packet loss concealment techniques for phone-to-hearing-aid streaming |
EP15191332.4A EP3012834B1 (en) | 2014-10-24 | 2015-10-23 | Packet loss concealment techniques for phone-to-hearing-aid streaming |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201462068404P | 2014-10-24 | 2014-10-24 | |
US14/854,704 US9706317B2 (en) | 2014-10-24 | 2015-09-15 | Packet loss concealment techniques for phone-to-hearing-aid streaming |
Publications (2)
Publication Number | Publication Date |
---|---|
US20160119725A1 true US20160119725A1 (en) | 2016-04-28 |
US9706317B2 US9706317B2 (en) | 2017-07-11 |
Family
ID=54608252
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/854,704 Active US9706317B2 (en) | 2014-10-24 | 2015-09-15 | Packet loss concealment techniques for phone-to-hearing-aid streaming |
Country Status (2)
Country | Link |
---|---|
US (1) | US9706317B2 (en) |
EP (1) | EP3012834B1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220408202A1 (en) * | 2021-06-21 | 2022-12-22 | Sonova Ag | Method and system for streaming a multichannel audio signal to a binaural hearing system |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040083110A1 (en) * | 2002-10-23 | 2004-04-29 | Nokia Corporation | Packet loss recovery based on music signal classification and mixing |
US20050091048A1 (en) * | 2003-10-24 | 2005-04-28 | Broadcom Corporation | Method for packet loss and/or frame erasure concealment in a voice communication system |
US20050166124A1 (en) * | 2003-01-30 | 2005-07-28 | Yoshiteru Tsuchinaga | Voice packet loss concealment device, voice packet loss concealment method, receiving terminal, and voice communication system |
US20050240402A1 (en) * | 1999-04-19 | 2005-10-27 | Kapilow David A | Method and apparatus for performing packet loss or frame erasure concealment |
US20070274550A1 (en) * | 2006-05-24 | 2007-11-29 | Phonak Ag | Hearing assistance system and method of operating the same |
US20110022924A1 (en) * | 2007-06-14 | 2011-01-27 | Vladimir Malenovsky | Device and Method for Frame Erasure Concealment in a PCM Codec Interoperable with the ITU-T Recommendation G. 711 |
US20110208517A1 (en) * | 2010-02-23 | 2011-08-25 | Broadcom Corporation | Time-warping of audio signals for packet loss concealment |
US20120314890A1 (en) * | 2010-02-12 | 2012-12-13 | Phonak Ag | Wireless hearing assistance system and method |
US20130129126A1 (en) * | 2010-04-22 | 2013-05-23 | Phonak Ag | Hearing assistance system and method |
US20140169599A1 (en) * | 2012-12-17 | 2014-06-19 | Starkey Laboratories, Inc. | Ear to ear communication using bluetooth low energy transport |
US20140170979A1 (en) * | 2012-12-17 | 2014-06-19 | Qualcomm Incorporated | Contextual power saving in bluetooth audio |
US20150255079A1 (en) * | 2012-09-28 | 2015-09-10 | Dolby Laboratories Licensing Corporation | Position-Dependent Hybrid Domain Packet Loss Concealment |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4966452B2 (en) | 1999-04-19 | 2012-07-04 | エイ・ティ・アンド・ティ・コーポレーション | Frame erasing concealment processor |
US8280728B2 (en) | 2006-08-11 | 2012-10-02 | Broadcom Corporation | Packet loss concealment for a sub-band predictive coder based on extrapolation of excitation waveform |
FR2907586A1 (en) * | 2006-10-20 | 2008-04-25 | France Telecom | Digital audio signal e.g. speech signal, synthesizing method for adaptive differential pulse code modulation type decoder, involves correcting samples of repetition period to limit amplitude of signal, and copying samples in replacing block |
CN101325631B (en) | 2007-06-14 | 2010-10-20 | 华为技术有限公司 | Method and apparatus for estimating tone cycle |
US8185388B2 (en) | 2007-07-30 | 2012-05-22 | Huawei Technologies Co., Ltd. | Apparatus for improving packet loss, frame erasure, or jitter concealment |
EP3031217B1 (en) * | 2013-08-09 | 2017-10-04 | Sonova AG | Hearing instrument with t-coil |
-
2015
- 2015-09-15 US US14/854,704 patent/US9706317B2/en active Active
- 2015-10-23 EP EP15191332.4A patent/EP3012834B1/en active Active
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050240402A1 (en) * | 1999-04-19 | 2005-10-27 | Kapilow David A | Method and apparatus for performing packet loss or frame erasure concealment |
US20040083110A1 (en) * | 2002-10-23 | 2004-04-29 | Nokia Corporation | Packet loss recovery based on music signal classification and mixing |
US20050166124A1 (en) * | 2003-01-30 | 2005-07-28 | Yoshiteru Tsuchinaga | Voice packet loss concealment device, voice packet loss concealment method, receiving terminal, and voice communication system |
US20050091048A1 (en) * | 2003-10-24 | 2005-04-28 | Broadcom Corporation | Method for packet loss and/or frame erasure concealment in a voice communication system |
US20070274550A1 (en) * | 2006-05-24 | 2007-11-29 | Phonak Ag | Hearing assistance system and method of operating the same |
US20110022924A1 (en) * | 2007-06-14 | 2011-01-27 | Vladimir Malenovsky | Device and Method for Frame Erasure Concealment in a PCM Codec Interoperable with the ITU-T Recommendation G. 711 |
US20120314890A1 (en) * | 2010-02-12 | 2012-12-13 | Phonak Ag | Wireless hearing assistance system and method |
US20110208517A1 (en) * | 2010-02-23 | 2011-08-25 | Broadcom Corporation | Time-warping of audio signals for packet loss concealment |
US20130129126A1 (en) * | 2010-04-22 | 2013-05-23 | Phonak Ag | Hearing assistance system and method |
US20150255079A1 (en) * | 2012-09-28 | 2015-09-10 | Dolby Laboratories Licensing Corporation | Position-Dependent Hybrid Domain Packet Loss Concealment |
US20140169599A1 (en) * | 2012-12-17 | 2014-06-19 | Starkey Laboratories, Inc. | Ear to ear communication using bluetooth low energy transport |
US20140170979A1 (en) * | 2012-12-17 | 2014-06-19 | Qualcomm Incorporated | Contextual power saving in bluetooth audio |
Non-Patent Citations (1)
Title |
---|
Etter, Walter. "Restoration of a discrete-time signal segment by interpolation based on the left-sided and right-sided autoregressive parameters." IEEE Transactions on Signal Processing 44.5 (1996): 1124-1135. * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220408202A1 (en) * | 2021-06-21 | 2022-12-22 | Sonova Ag | Method and system for streaming a multichannel audio signal to a binaural hearing system |
US12149892B2 (en) * | 2021-06-21 | 2024-11-19 | Sonova Ag | Method and system for streaming a multichannel audio signal to a binaural hearing system |
Also Published As
Publication number | Publication date |
---|---|
EP3012834A1 (en) | 2016-04-27 |
EP3012834B1 (en) | 2021-02-24 |
US9706317B2 (en) | 2017-07-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US12149890B2 (en) | Neural network-driven frequency translation | |
US9712930B2 (en) | Packet loss concealment for bidirectional ear-to-ear streaming | |
CN102461040B (en) | Systems and methods for preventing the loss of information within a speech frame | |
CA2777601C (en) | A hearing aid with audio codec and method | |
JP6335301B2 (en) | Method and apparatus for encoding stereo phase parameters | |
US9877115B2 (en) | Dynamic relative transfer function estimation using structured sparse Bayesian learning | |
US9980058B2 (en) | Method for selecting transmission direction in a binaural hearing aid | |
WO2010112073A1 (en) | Adaptive feedback cancellation based on inserted and/or intrinsic characteristics and matched retrieval | |
US20230290333A1 (en) | Hearing apparatus with bone conduction sensor | |
WO2019071149A1 (en) | Spatial relation coding of higher order ambisonic coefficients using expanded parameters | |
US9706317B2 (en) | Packet loss concealment techniques for phone-to-hearing-aid streaming | |
KR102587134B1 (en) | Speech signal encoding and decoding | |
WO2018231469A1 (en) | Advanced packet-based sample audio concealment | |
US20170094424A1 (en) | Binaurally coordinated frequency translation in hearing assistance devices | |
US20220148558A1 (en) | Feedback cancellation divergence prevention | |
JPWO2008090970A1 (en) | Stereo encoding apparatus, stereo decoding apparatus, and methods thereof | |
US20220358945A1 (en) | Snr profile adaptive hearing assistance attenuation | |
WO2022231977A1 (en) | Recovery of voice audio quality using a deep learning model | |
KR20230035089A (en) | packet loss concealment | |
JP6718670B2 (en) | ADPCM decoder including packet loss concealment method in ADPCM codec and PLC circuit | |
US20250030990A1 (en) | Telecoil predicted repeating system interference canceller | |
US11570562B2 (en) | Hearing assistance device fitting based on heart rate sensor | |
US20240323618A1 (en) | Reducing comb filtering for hearing devices | |
EP4375999A1 (en) | Audio device with signal parameter-based processing, related methods and systems | |
CN111261175A (en) | Bluetooth audio signal transmission method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: STARKEY LABORATORIES, INC., MINNESOTA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MUSTIERE, FREDERIC PHILIPPE DENIS;MOVAHEDI, MOHAMMED REZA;ZHANG, TAO;AND OTHERS;SIGNING DATES FROM 20161219 TO 20170213;REEL/FRAME:041443/0168 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: CITIBANK, N.A., AS ADMINISTRATIVE AGENT, TEXAS Free format text: NOTICE OF GRANT OF SECURITY INTEREST IN PATENTS;ASSIGNOR:STARKEY LABORATORIES, INC.;REEL/FRAME:046944/0689 Effective date: 20180824 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |