Language selection

Search

Patent 2439427 Summary

Third-party information liability

Some of the information on this Web page has been provided by external sources. The Government of Canada is not responsible for the accuracy, reliability or currency of the information supplied by external sources. Users wishing to rely upon this information should consult directly with the source of the information. Content provided by external sources is not subject to official languages, privacy and accessibility requirements.

Claims and Abstract availability

Any discrepancies in the text and image of the Claims and Abstract are due to differing posting times. Text of the Claims and Abstract are posted:

  • At the time the application is open to public inspection;
  • At the time of issue of the patent (grant).
(12) Patent: (11) CA 2439427
(54) English Title: METHOD FOR DETERMINING AN ACOUSTIC ENVIRONMENT SITUATION, APPLICATION OF THE METHOD AND HEARING AID
(54) French Title: PROCEDE DE DETERMINATION D'UNE SITUATION ACOUSTIQUE ENVIRONNANTE, UTILISATION DU PROCEDE ET APPAREIL DE CORRECTION AUDITIVE
Status: Expired
Bibliographic Data
Abstracts

English Abstract




The invention relates to a method and a device for determining an acoustic
environment situation. The method consists of processing an acoustic input
signal (IN) that is preferably picked up with the help of at least one
microphone in at least two processing stages (S1, ..., Sn). At least one of
the two processing stages (S1, ..., Sn) comprises an extraction phase in which
characteristic features are extracted from the input signal (IN) and each of
the processing stages (S1, ..., Sn) comprises an identification phase in which
extracted characteristic features are classified. Class information (KI1, ...,
KIn; KI1', ..., KIn') characterising or identifying the acoustic environment
situation is generated based on the classification of said features in at
least one processing stage (S1, ..., Sn). The invention also relates to
applications of the inventive method in hearing aids and to a hearing aid.


French Abstract

L'invention concerne un procédé et un dispositif pour la détermination d'une situation acoustique environnante. Le procédé est caractérisé en ce que dans au moins deux étapes de traitement (S1, ..., Sn), on traite un signal d'entrée acoustique (IN) reçu de préférence par au moins un microphone, de telle façon que dans au moins l'une des deux étapes (S1, ..., Sn), une phase d'extraction soit prévue, dans laquelle sont extraites les caractéristiques du signal d'entrée (IN), et en ce qu'il est prévu dans chaque étape de traitement (S1, , Sn), une phase d'identification dans laquelle les caractéristiques extraites sont classifiées. Sur la base de la classification des caractéristiques dans au moins une étape de traitement (S1, , Sn), on produit des informations de classes (KI1, ..., KIn; KI1', ..., KIn') qui caractérisent ou identifient la situation acoustique environnante. L'invention concerne en outre des utilisations du procédé avec des appareils de correction auditive. L'invention concerne également un appareil de correction auditive.

Claims

Note: Claims are shown in the official language in which they were submitted.



19
WHAT IS CLAIMED IS:

1. A method for identifying an acoustic scene, comprising the steps of:
recording an acoustic input signal; and
providing at least two processing stages wherein:
an extraction phase is provided in at least one of the at least two processing
stages, in which said extraction phase characteristic features are extracted
from the
input signal, and wherein
an identification phase is provided in each processing stage, in which said
identification phase the extracted characteristic features are classified, and
further
wherein
class information is generated according to the classification of the features
in
at least one of the processing stages, wherein said class information
characterizes or
identifies the acoustic scene,
wherein a manner of processing in a processing stage is selected according to
the class information obtained in another processing stage.

2. The method according to claim 1, wherein the class information obtained in
the identification phase of a processing stage i determines a processing
manner in
one of the following, inferior processing stages i+1.

3. The method according to claim 2, wherein, according to class information
obtained in the processing stage i, specific features are selected in the
extraction
phase of the following, inferior processing stage i+1 or specific
classification methods
are selected in the identification phase of the following, inferior processing
stage i+1
or both.

4. Method according to claim 2, wherein a post-processing phase is provided in
at least one processing stage subsequent to the extraction phase, in which


20
postprocessing stage the class information are revised in order to generate
revised
class information.

5. Method according to claim 1, wherein a post-processing phase is provided in
at least one processing stage subsequent to the extraction phase, in which
postprocessing stage the class information are revised in order to generate
revised
class information.

6. The method according to claim 1, wherein one or more of the following
classification methods is used in the identification phase: Hidden Markov
Models;
Fuzzy Logic; Bayes Classifier; Rule-based Classifier Neuronal Networks; and
Minimal Distance.

7. Method according to claim 1, wherein technical, auditory-based or both,
features are extracted in the extraction phase.

8. A method for identifying an acoustic scene, comprising the steps of:
recording an acoustic input signal; and
providing at least two processing stages wherein:
an extraction phase is provided in at least one of the at least two processing
stages, in which said extraction phase characteristic features are extracted
from the
input signal, and wherein
an identification phase is provided in each processing stage, in which said
identification phase the extracted characteristic features are classified, and
further
wherein
class information is generated according to the classification of the features
in
at least one of the processing stages, wherein said class information
characterizes or
identifies the acoustic scene, and wherein an extraction phase is provided in
each
processing stage, in which extraction phase characteristic features are
extracted
from the input signal, and further wherein


21
a manner of processing in a processing stage is selected according to the
class information obtained in another processing stage.

9. The method according to claim 8, wherein the class information obtained in
the identification phase of a processing stage i determines a processing
manner in
one of the following, inferior processing stages i+1.

10. The method according to claim 9, wherein, according to class information
obtained in the processing stage i, specific features are selected in the
extraction
phase of the following, inferior processing stage i+1 or specific
classification methods
are selected in the identification phase of the following, inferior processing
stage i+1
or both.

11. Method according to claim 9, wherein a post-processing phase is provided
in
at least one processing stage subsequent to the extraction phase, in which
postprocessing stage the class information are revised in order to generate
revised
class information.

12. Method according to claim 8, wherein a post-processing phase is provided
in
at least one processing stage subsequent to the extraction phase, in which
postprocessing stage the class information are revised in order to generate
revised
class information.

13. A method for identifying an acoustic scene, comprising the steps of:
recording an acoustic input signal; and
providing at least two processing stages wherein:
an extraction phase is provided in at least one of the at least two processing
stages, in which said extraction phase characteristic features are extracted
from the
input signal, and wherein


22
an identification phase is provided in each processing stage, in which said
identification phase the extracted characteristic features are classified, and
further
wherein
class information is generated according to the classification of the features
in
at least one of the processing stages, wherein said class information
characterizes or
identifies the acoustic scene, and wherein an extraction phase is provided in
each
processing stage, in which extraction phase characteristic features are
extracted
from the input signal, and further wherein
the class information obtained in the identification phase of a processing
stage
i determines a processing manner in one of the following, inferior processing
stages
i+1.

14. The method according to claim 13, wherein, according to class information
obtained in the processing stage i, specific features are selected in the
extraction
phase of the following, inferior processing stage i+1 or both specific
classification
methods are selected in the identification phase of the following, inferior
processing
stage i+1 or both.

15. Method according to claim 14, wherein a post-processing phase is provided
in
at least one processing stage subsequent to the extraction phase, in which
postprocessing stage the class information are revised in order to generate
revised
class information.

16. Method according to claim 13, wherein a post-processing phase is provided
in
at least one processing stage subsequent to the extraction phase, in which
postprocessing stage the class information are revised in order to generate
revised
class information.

17. Use of the method according to one of claims 1 to 16 for the adjustment of
at
least one hearing device to a momentary acoustic scene.


23
18. Use of the method according to claim 17, wherein a hearing program or a
transfer function between at least one microphone and a speaker in a hearing
device
is selected according to a determined acoustic scene.

19. Use of the method according to one of claims 1 to 16 for speech analysis
or
speech detection.

20. A device for identifying an acoustic scene in an input signal, the device
comprising:
at least two processing stages;
a feature extraction unit in at least one of the at least two processing
stages;
and
a classification unit in each one of said at least two processing stages,
wherein:
the input signal is fed to the feature extraction unit, an output of which is
at
least fed to one of the at least two classification units, and wherein at
least one of the
at least two classification units is operatively connected to at least another
of the at
least two classification units in order to adjust processing according to
class
information in another processing stage.

21. The device according to claim 20, further comprising a feature extraction
unit
in each processing stage.

22. The device according to claim 21, wherein the class information of a
processing stage i is fed to a following, inferior processing stage i+1.

23. The device according to claim 22, wherein the class information of a
processing stage i is fed to a feature extraction unit of a following,
inferior processing
stage i+1, or wherein the class information of a processing stage i is fed to
a
classification unit of a following, inferior processing stage i+1 or both.


24
24. The device according to claim 20, wherein the class information is fed to
other
processing stages.

25. The device according to claim 24, wherein the class information of a
processing stage i is fed to a following, inferior processing stage M.

26. The device according to claim 25, wherein the class information of a
processing stage i is fed to a feature extraction unit of a following,
inferior processing
stage i+1, or wherein the class information of a processing stage i is fed to
a
classification unit of a following, inferior processing stage i+1 or both.

27. The device according to claim 21, wherein the class information is fed to
other
processing stages.

28. The device according to claim 27, wherein the class information of a
processing stage i is fed to a following, inferior processing stage M.

29. The device according to claim 28, wherein the class information of a
processing stage i is fed to a feature extraction unit of a following,
inferior processing
stage i+1, or wherein the class information of a processing stage i is fed to
a
classification unit of a following, inferior processing stage i+1 or both.

30. The device according to claim 20, wherein the class information of a
processing stage i is fed to a following, inferior processing stage M.

31. The device according to claim 30, wherein the class information of a
processing stage i is fed to a feature extraction unit of a following,
inferior processing
stage i+1, or wherein the class information of a processing stage i is fed to
a
classification unit of a following, inferior processing stage i+1 or both.


25
32. The device according to one of claims 20 to 31, wherein the class
information
obtained in at least one processing stage is fed to a post-processing unit in
order to
generate revised class information.

33. The device according to claim 20 or 21, wherein the class information of
all
processing stages is fed to a decision unit.

34. The device according to claim 33, wherein the decision unit is operatively
connected to at least one of the feature extraction units, or to at least one
of the
classification units or both.

35. A hearing device with a transfer unit operatively connected to at least
one
microphone and to a converter unit, and with a device according to one of
claims 20
to 31 for generating class information, whereas the class information is fed
to the
transfer unit.

36. The hearing device according to claim 35, further comprising an input unit
which is operatively connected to the transfer unit.

37. The hearing device according to claim 36, comprising a wireless link
between
the input unit and the transfer unit.

38. A hearing device according to claim 35, wherein the convecter device is a
speaker.

Description

Note: Descriptions are shown in the official language in which they were submitted.



CA 02439427 2003-08-27
1

METHOD FOR DETERMINING AN ACOUSTIC ENVIRONMENT SITUATION,
APPLICATION OF THE METHOD AND HEARING AID

The present invention is related to a method for
identifying an acoustic scene, a use of the method, a
device for identifying the acoustic scene as well as a
hearing device.

Modern day hearing devices, when employing different
hearing programs, permit their adaptation to varying
acoustic environments or scenes. Therewith, the hearing
device offers the user an optimal use in every situation.

The hearing program can be selected either via a remote
control or by means of a selector switch on the hearing
device itself. For many users, however, having to switch
program settings is a nuisance, or difficult, or even
impossible. Nor it is always easy even for experienced
wearers of hearing devices to determine at what point in
time which program is most comfortable and offers optimal
speech discrimination. An automatic recognition of the
acoustic scene and corresponding automatic switching of the
hearing program settings in the hearing device is therefore
desirable.


CA 02439427 2003-08-27

- 2 -

There exist several different methods for the automatic
classification of acoustic scenes. All of the methods
concerned involve the extraction of different features from
the input signal, which may be derived from one or several
microphones in the hearing device. Based on these features,
a pattern recognition device employing a particular
algorithm makes the determination as to the attribution of
the analyzed signal to a specific acoustic scene. These
various existing methods differ from one another both in
terms of the features on the basis of which they define the
acoustic scene (signal analysis) and with regard to the
pattern recognition device, which serves to classify these
features (signal identification).

From the publication of the international patent
application having the publication No. WO 01/20 965 a
method and a device for identifying an acoustic scene are
known. Described is a single-stage process in which an
acoustic input signal is processed in a feature extraction
unit and, afterwards, in a classification unit, in which
the extracted features are classified to generate class
information. Good results are obtained by this known
teaching in particular if auditory-based features are
extracted. An improvement is desirable particularly in the
field of hearing devices, since in this application field
the classification of acoustic scenes must be very
accurate. At the same time, the occurrence of several very
broad sound classes, as e.g. music or noise, cause greater
difficulties. It corresponds to the nature of these sound


CA 02439427 2010-11-10
3

classes that they are very general and broad, i.e. their
occurrence may be in manifold manner. The sound class
"noise", for example, comprises very different sounds as
e.g. background noise resulting from discussions, train
station noise, hair dryer noise, and the sound class
"music" comprises for example pop music, classic music,
single instruments, singing, etc.

Especially because of the very general nature of these
sound classes, it is very difficult to obtain a good
recognition rate with the aid of the known processing
methods in a feature extraction unit and a following
classification unit. In fact, the robustness of the
recognition system can be improved by the selection of
features as has been described in WO 01/20965 for the first
time, namely by using auditory-based features.
Nevertheless, it is very difficult to separate between
different general sound classes in a clear and doubtless
manner, because of the high variance of these general sound
classes.

It is therefore an object of the present invention to
introduce a method for identifying an acoustic scene, which
is more reliable and more precise compared to prior art
methods.

In accordance with one aspect of the invention, there is provided a method for
identifying an acoustic scene, comprising the steps of:


CA 02439427 2010-11-10
4

recording an acoustic input signal; and
providing at least two processing stages wherein:
an extraction phase is provided in at least one of the at least two processing
stages, in which said extraction phase characteristic features are extracted
from the
input signal, and wherein
an identification phase is provided in each processing stage, in which said
identification phase the extracted characteristic features are classified, and
further
wherein
class information is generated according to the classification of the features
in
at least one of the processing stages, wherein said class information
characterizes or
identifies the acoustic scene,
wherein a manner of processing in a processing stage is selected according to
the class information obtained in another processing stage.

In accordance. with another aspect of the invention, there is provided a
method for
identifying an acoustic scene, comprising the steps of:
recording an acoustic input signal; and
providing at least two processing stages wherein:
an extraction phase is provided in at least one of the at least two processing
stages, in which said extraction phase characteristic features are extracted
from the
input signal, and wherein
an identification phase is provided in each processing stage, in which said
identification phase the extracted characteristic features are classified, and
further
wherein
class information is generated according to the classification of the features
in
at least one of the processing stages, wherein said class information
characterizes or
identifies the acoustic scene, and wherein an extraction phase is provided in
each
processing stage, in which extraction phase characteristic features are
extracted
from the input signal, and further wherein


CA 02439427 2010-11-10

4a
a manner of processing in a processing stage is selected according to the
class information obtained in another processing stage.

Still another aspect of the invention provides a method for identifying an
acoustic
scene, comprising the steps of:
recording an acoustic input signal; and
providing at least two processing stages wherein:
an extraction phase is provided in at least one of the at least two processing
stages, in which said extraction phase characteristic features are extracted
from the
input signal, and wherein
an identification phase is provided in each processing stage, in which said
identification phase the extracted characteristic features are classified, and
further
wherein
class information is generated according to the classification of the features
in
at least one of the processing stages, wherein said class information
characterizes or
identifies the acoustic scene, and wherein an extraction phase is provided in
each
processing stage, in which extraction phase characteristic features are
extracted
from the input signal, and further wherein
the class information obtained in the identification phase of a processing
stage
i determines a processing manner in one of the following, inferior processing
stages
i+1.

Still yet another aspect of the invention provides a device for identifying an
acoustic
scene in an input signal, the device comprising:
at least two processing stages;
a feature extraction unit in at least one of the at least two processing
stages;
and
a classification unit in each one of said at least two processing stages,
wherein:


CA 02439427 2010-11-10
4b

the input signal is fed to the feature extraction unit, an output of which is
at
least fed to one of the at least two classification units, and wherein at
least one of the
at least two classification units is operatively connected to at least another
of the at
least two classification units in order to adjust processing according to
class
information in another processing stage.

Advantageous embodiments of the present invention, a use of the method, a
device
as well as a hearing device are provided.
By processing an acoustic input signal in a multistage
process in which at least two classification stages are
implemented, whereas each stage preferably comprises an
extraction phase and an identification phase. The present
invention has the advantage to obtain a very robust and
precise classification of the momentary acoustic scene. The
present invention allows preventing successfully a wrong
classification of, for example, pop music in the sound
class of "speech in noise". In addition, the present method
allows a breakdown of a general sound class, as for example
noise, in subclasses, as for example traffic noise or
background noise resulting from discussions. Special
situations, as for example in-the-car noise, can also be
recognized. In general, room characteristics can be
identified and taken into consideration correspondingly in
further processing of important signal parts. Furthermore,
the present invention can be used to localize sound
sources, whereby the possibility is obtained to detect the
occurrence of a specific sound source in a mixture of
several other sound sources.


CA 02439427 2010-11-10

4c
In the following, the invention is explained in more detail
by way of an example with reference to drawings. Thereby,
it is shown in:


CA 02439427 2003-08-27

-

Fig. 1 a known single-stage device for identifying an
acoustic scene;

Fig. 2 a first embodiment of a device according to the
5 invention with two processing stages;

Fig. 3 a second, general embodiment of a multistage
device according to the present invention;

Fig. 4 a third, general embodiment of a multistage device
according to the present invention;

Fig. 5 a fourth, general embodiment of a multistage
device according to the present invention;

Fig. 6 an embodiment of the present invention which is
simplified compared to the two-stage embodiment
according to fig. 2, and

Fig. 7 a hearing device with a multistage device
according to figs. 2 to 6.

Fig. 1 shows a known single-stage device for identifying an
acoustic scene, whereby the device comprises a feature
extraction unit F, a classification unit C and a post-
processing unit 2 connected together in sequence.


CA 02439427 2003-08-27

- 6 -

An acoustic input signal IN, which has been recorded by a
microphone, for example, is fed to the feature extraction
unit F in which characteristic features are extracted.


For the extraction of features in audio signals, J.M. Kates
in his article titled "Classification of Background Noises
for Hearing-Aid Applications" (1995, Journal of the
Acoustical Society of America 97(1), pp. 461 - 469)
suggested an analysis of time-related sound level
fluctuations and of the sound spectrum. On its part, the
European Patent EP-B1-0 732 036 proposed an analysis of the
amplitude histogram for obtaining the same result. Finally,
the extraction of features has been investigated and
implemented based on an analysis of different modulation
frequencies. In this connection, reference is made to the
two papers by Ostendorf et al. titled "Empirical
classification of different acoustic signals and of speech
by means of a modulation frequency analysis" (1997, DAGA
97, pp. 608 - 609), and "Classification of acoustic signals
based on the analysis of modulation spectra for application
in digital hearing aids" (1998, DAGA 98, pp. 402 - 403). A
similar approach is described in an article by Edwards et
al. titled (Signal-processing algorithms for a new
software-based, digital hearing device" (1998, The Hearing
Journal 51, pp. 44 - 52). Other possible features include
the sound level transmission itself or the zero-crossing
rate as described e.g. in the article by H.L. Hirsch,
titled "Statistical Signal Characterization" (Artech House


CA 02439427 2003-08-27

7 -

1992). The features being used for the analysis of audio
signals so fare are strictly technically-based.
Furthermore, it has been pointed out in the already
mentioned publication of the International Patent
Application WO 01/20 965 that besides the mentioned
technical features the use of auditory-based features is
very advantageous.

According to fig. 1 the features M extracted in the feature
extraction unit F will be fed to the classification unit C
in which one of the known pattern identification methods is
being basically applied for the sound classification.
Particularly suitable pattern recognition systems are the
so-called distance estimators, Bayes' classifiers, fuzzy
logic systems and neuronal networks. Details of the first
two methods mentioned above are contained in the
publication titled "Pattern Classification and Scene
Analysis" by Richard 0. Duda and Peter E. Hart (John Wiley
& Sons, 1973). For information on Neuronal Networks,
reference is made to the standard work by Christopher M.
Bishop, titled "Neural Networks for Pattern Recognition"
(1995, Oxford University Press). Reference is also made to
the following publications: Ostendorf et al.,
"Classification of acoustic signals based on the analysis
of modulation spectra for application in digital hearing
aids" (Zeitschrift fur Audiologie (Journal of Audiology),
pp. 148 -150); F. Feldbusch, "Sound recognition using
neuronal networks" (1998, Journal of Audiology, pp. 30 -


CA 02439427 2003-08-27

8 -

36); European Patent Application with publication No. EP-
A1-0 814 636; and US Patent having publication No. US-5 604
812. Besides the mentioned pattern recognition methods, by
which only the static properties of the interesting sound
classes are being modeled, there are also mentioned other
methods in the already mentioned publication of the
International Patent Application WO 01/20 965 by which
dynamic properties are being considered.

According to fig. 1, class information KI are being
obtained by processing steps implemented in the
classification unit C. The class information KI may be fed,
as the case may be, to a post-processing unit P for the
possible revision of the class affiliation. As a result,
revised class information KI' are obtained.

In fig. 2, a first embodiment of a device according to the
present invention is shown. The device has two processing
stages Si and S2, whereby a feature extraction unit Fl or
F2, respectively, and a classification unit Cl or C2,
respectively, are provided in each stage Si and S2,
respectively. The original input signal IN is fed to both
processing stages Si and S2, respectively, namely to the
feature extraction unit Fl as well as to the feature
extraction unit F2, which are each operatively connected to
the corresponding classification unit Cl and C2,
respectively. It is important to note that the class
information KI1, which are obtained in the first processing
stage Si on the basis of calculations in the classification


CA 02439427 2003-08-27

9 -

unit Cl, has effect on the classification unit C2 of the
second processing stage S2, in fact in such a way that, for
example, one of several possible pattern identification
methods is selected and applied to the sound classification
in the classification unit C2 of the second processing
stage S2.

The embodiment generally represented in fig. 2 of the
present invention will be further described now by way of a
concrete example:

By the feature extraction unit Fl, the features tonality,
spectral center of gravity (CGAV), fluctuation of the
spectral center of gravity (CGFS) and spectral width and
settling time are being extracted and classified in the
classification unit C1, in which a HMM- (Hidden Markov
Model) classifier is being used, whereby the input signal
IN is classified in one of the following classes by the HMM
classifier: "speech", "speech in noise", "noise" or
"music". This result is referred to as class information
KI. The result of the first processing stage Sl is fed to
the classification unit C2 of the processing stage S2 in
which a second set of features is being extracted using the
feature extracting unit F2. Thereby, the additional feature
variance of the harmonic structure (pitch) - also referred
to Pitchvar in the following - is being extracted besides
the features tonality, spectral center of gravity and
fluctuation of the spectral gravity. On the basis of these
features, the result of the first processing stage Si will


CA 02439427 2003-08-27

- 10 -

be verified and, if need be, corrected. The verification is
being done with the aid of a rule-based classifier in the
classification unit C2. The rule-based classifier contains
a few simple heuristic decisions only, which are based on
the four features and which are orientated at the following
reflections:

The feature tonality will be used in each class for the
correction if the value of the feature completely lies
outside of a valid value range of the class information
KI1, which has been determined in the first classification
unit C1 - i.e. by the HMM classifier. It is expected that
the tonality for "music" is high, for "speech" it is in the
middle range, for "speech in noise" it is a little bit
lower and for "noise" it is low. If, for example, an input
signal IN falls into the class "speech" by the
classification unit C1 then it is expected that
corresponding features which have been determined in the
feature extraction unit Fl have indicated to the
classification unit C1 that the relevant signal part in the
input signal IN is strongly fluctuating. If, on the other
side, the tonality for this input signal IN is very low,
the correct class information will not be "speech" with
high probability but "speech in noise". Similar
considerations can be carried out for the other three
features, namely for the variance of the harmonic structure
(Pitchvar), the spectral center of gravity (CGAV) and for
the fluctuation of the spectral gravity (CGFS).
Accordingly, the rules for the rule-based classifier which


CA 02439427 2003-08-27

- 11 -

is implemented in the classification unit C2 can be
formulated as follows:

Class information: Condition: Class information
KIl: KI2:
"speech" if tonality low "speech in noise"

if CGFS high

and CGAV high "music"
otherwise "noise"
"speech in noise" If tonality high "speech"
If tonality low or
CGAV high "noise"
"noise" if tonality high "music"
"music" If tonality low
or Pitchvar low
"noise"
or CGAV high

For this embodiment of the present invention the
recognition has even emerged as a surprise, namely that
almost the same features are used in the second processing
stage S2 as have been used in the first processing stage
Si. Furthermore, it can be noted that the feature tonality
is best suitable in order to correct an error which has
been generated by the classification unit Cl. After all, it
can be noted that the tonality is most important for the
use of the rule-based classifier.


CA 02439427 2003-08-27

- 12 -

A test of the afore described embodiment has revealed that
for the simple process having two stages the hit rate
improved by at least 3% compared to the single-stage
process. In several cases it has been possible to improve
the hit rate by 91%.

In fig. 3, a further embodiment of the present invention is
shown in a general representation in which a process is
shown with n stages. Each of the processing stages Si to Sn
comprises, as a consequence of the aforementioned
considerations, a feature extraction unit Fl, ..., Fn
followed by a classification unit Cl, ..., Cn for the
generation of the corresponding class information KI1, ...,
KIn. As the case may be, a post-processing unit P1, ..., Pn
for the generation of revised class information KI1', ...,
KIn' is provided in each or in a single or in several
processing stages S1, ..., Sn.

In continuation of the embodiment according to fig. 2, the
embodiment according to fig. 3 is particularly suited to a
so-called coarse-fine classification. In a coarse-fine
classification, a result obtained in the processing stage i
will be refined in a following processing stage i+1. In
other words, a coarse classification is provided in a
superior processing stage, whereby, on the basis of the
coarse classification, a fine classification based on more
specific feature extractions and/or classification methods
is implemented in an inferior processing stage. This
process can also be seen as a generation of hypothesis in a


CA 02439427 2003-08-27

- 13 -

superior processing stage which hypothesis is reviewed in a
following, i.e. inferior processing stage, in other words,
the hypothesis is confirmed or rejected in this inferior
processing stage. At this point, it is emphasized that the
hypothesis which is generated in a superior processing
stage (coarse classification) can be provided by other
sources, particularly by manual means, as e.g. by a remote
control or by a switch. In fig. 3, this is indicated,
representatively in the first processing stage Sl, by a
controlled variable ST by which, for example, the
calculation in the classification unit Cl can be overruled.
As a matter of course, the control variable ST can also be
fed to a classification unit C2 to Cn or to a post-
processing unit Pl to Pn of another processing stage Si to
Sn.

In a classification system according to the present
invention having several processing stages S1 to Sn, a task
can be assigned to each of the processing stages Si to Sn,
although it is not mandatory, as for example: a coarse
classification, a fine classification, a localization of a
sound source, a verification whether a certain sound
source, e.g. in-the-car noise, exists, or an extraction of
certain signal parts of an input signal, e.g. the
elimination of echo as a result of certain room
characteristics. Each of the processing stages Si to Sn are
therefore individual in the sense that, for each stage,
different features are extracted and different
classification methods are being used.


CA 02439427 2003-08-27

- 14 -

In a further embodiment of the present invention, it is
provided to locate an individual signal in a mixture of
different signal parts in a first processing stage Si, to
implement a coarse classification of the located signal
source in a second processing stage S2, and to implement a
fine classification of the coarse classification obtained
in the second processing stage S2.

Furthermore, a direction filtering can follow the
localization of a sound source performed in the first
processing stage, e.g. by using the Multi-Microphone
Technology.

Naturally, a feature extraction unit Fl to Fn can be
subdivided into several classification units C1 to On, i.e.
the results of a feature extraction unit Fl to Fn can be
used by several classification units C1 to On. Furthermore,
it is feasible that a classification unit Cl to On can be
used in several processing stages Si to Sn. Finally, it is
possible that the class information KI1 to KIn or the
revised class information KIl' to KIn' obtained in the
different processing stages S1 to Sn are weighted
differently in order to obtain a final classification.

In fig. 4, a further embodiment according to the invention
is represented for which several processing stages S1 to Sn
are again being used. Apart from the embodiment according


CA 02439427 2003-08-27

- 15 -

to fig. 3, the class information KI1 to KIn will not only
be used in the immediately following processing stage but,
as the case may be, in all inferior processing stages. In
analog manner, the results of the superior processing stage
Si to Sn may also have their impact on the inferior feature
extraction units F1 to Fn or on the features to be
extracted, respectively.

The processing units P1 to Pn may also be implemented in
the embodiment according to fig. 4, in which post-
processing units P1 to Pn intermediate results of the
classification are obtained, and in which post-processing
units PI to Pn revised class information KI1' to KIn' are
generated.


In fig. 5, a further embodiment of the present invention is
shown having a multistage device for identifying the
acoustic scene, again in general form. As for the
embodiments according to figs. 3 and 4 several processing
stages Si to Sn are shown with feature extraction units F1
to Fn and classification units Cl to Cn. The class
information KI1 to KIn obtained in each processing stage Si
to Sn are fed to a decision unit FD in which the final
classification is obtained by generating the class
information KI. In the decision unit FD it is provided, if
need be, to generate feedback signals which are fed to the
feature extraction units Fl to Fl and/or to the
classification units Cl to Cn in order to adjust, for
example, one or several parameters in the processing units,


CA 02439427 2003-08-27

- 16 -

or in order to exchange a whole classification unit Cl to
Cn.

It has to be noted that the feedback signals and
connections of the processing units of the embodiments
according to figs. 3 to 5 are not limited to the
represented embodiments. It is conceivable that some of the
feedback signals or some of the connections are omitted. In
general, any combination of processing units is possible to
obtain any possible structure.

Furthermore, it is feasible that - applying the present
invention for hearing devices - the several processing
stages are distributed between two hearing devices, i.e.
one hearing device located at the right ear, the other
hearing device located at the left ear. For this
embodiment, the information exchange is provided by a wired
or a wireless transmission link.

In fig. 6, a simplified embodiment of the present invention
is again represented to illustrate the above mentioned
general explanations to the possible structures and
combinations of processing units. Although only one feature
extraction unit Fl is represented, two processing stages Si
and S2 are provided. The first processing stage Si
comprises a feature extraction unit F1 and a classification
unit C1. In the second processing stage S2, the same
features are used as in the first processing stage Si. A


CA 02439427 2003-08-27

- 17 -

recalculation of the features in the second processing
stage S2 is therefore not necessary, and it is possible to
use the results of the feature extraction unit F1 of the
first processing stage Si in the second processing stage
S2. In the second processing stage S2 the classification
method is therefore adjusted only, in fact in dependence of
the class information KIl of the first processing stage Sl.
Fig. 7 shows the use of the invention in a hearing device
which essentially comprises a transfer unit 200. By the
reference sign 100 a multistage processing unit is
identified which is realized according to one of the
embodiments represented in figs. 2 to 6. The input signal
IN is fed to the multistage processing unit as well as to
the transfer unit 200 in which the acoustic input signal IN
is processed with the aid of the class information KIl to
KIn or the revised class information KI1' to KIn',
respectively, generated in the multistage processing unit
100. Thereby, it is envisioned to select a suitable hearing
program according to the acoustic scene which has been
identified as has been described above and in the
International Patent Application WO 01/20 965.

By the reference sign 300, a manual input unit is
identified by which - for example over a wireless link as
schematically represented in fig. 7 - the multistage
processing unit 100, as described above, or the transfer
unit 200 are affected, if need be. In the case of the


CA 02439427 2010-11-10
18

hearing device 200 reference is made to WO 01/20965.

As possible classification method, one of the following
methods can be used for all described embodiments of the
present invention:

-Hidden Markov Models;
- Fuzzy Logic;

-Bayes' Classifier;
-Rule-based Classifier
-Neuronal Networks;

- Minimal Distance.

Finally, it has to be noted that technical- and/or
auditory-based features can be extracted in the feature
extraction units F1 to Fn (figs. 2 to 7). Extensive
explanations can again be found in the International Patent
Application WO 01/20965 in which technical features as well
as auditory-based features are defined.

The preferred use of the present invention for identifying
the acoustic scene is the selection of a hearing program in
a hearing device. It is also conceivable to use the present
invention for speech detection and speech analysis,
respectively.

Representative Drawing
A single figure which represents the drawing illustrating the invention.
Administrative Status

For a clearer understanding of the status of the application/patent presented on this page, the site Disclaimer , as well as the definitions for Patent , Administrative Status , Maintenance Fee  and Payment History  should be consulted.

Administrative Status

Title Date
Forecasted Issue Date 2011-03-29
(86) PCT Filing Date 2002-01-28
(87) PCT Publication Date 2002-04-25
(85) National Entry 2003-08-27
Examination Requested 2006-12-28
(45) Issued 2011-03-29
Expired 2022-01-28

Abandonment History

There is no abandonment history.

Payment History

Fee Type Anniversary Year Due Date Amount Paid Paid Date
Application Fee $300.00 2003-08-27
Registration of a document - section 124 $100.00 2003-12-03
Maintenance Fee - Application - New Act 2 2004-01-28 $100.00 2003-12-09
Maintenance Fee - Application - New Act 3 2005-01-28 $100.00 2004-12-13
Maintenance Fee - Application - New Act 4 2006-01-30 $100.00 2005-12-28
Request for Examination $800.00 2006-12-28
Maintenance Fee - Application - New Act 5 2007-01-29 $200.00 2007-01-04
Maintenance Fee - Application - New Act 6 2008-01-28 $200.00 2007-12-27
Maintenance Fee - Application - New Act 7 2009-01-28 $200.00 2008-12-04
Maintenance Fee - Application - New Act 8 2010-01-28 $200.00 2009-11-18
Maintenance Fee - Application - New Act 9 2011-01-28 $200.00 2010-12-15
Final Fee $300.00 2011-01-13
Maintenance Fee - Patent - New Act 10 2012-01-30 $250.00 2012-01-05
Maintenance Fee - Patent - New Act 11 2013-01-28 $250.00 2012-12-13
Maintenance Fee - Patent - New Act 12 2014-01-28 $250.00 2013-12-30
Maintenance Fee - Patent - New Act 13 2015-01-28 $250.00 2015-01-26
Maintenance Fee - Patent - New Act 14 2016-01-28 $250.00 2016-01-25
Maintenance Fee - Patent - New Act 15 2017-01-30 $450.00 2017-01-23
Maintenance Fee - Patent - New Act 16 2018-01-29 $450.00 2018-01-22
Maintenance Fee - Patent - New Act 17 2019-01-28 $450.00 2019-01-21
Maintenance Fee - Patent - New Act 18 2020-01-28 $450.00 2020-01-24
Owners on Record

Note: Records showing the ownership history in alphabetical order.

Current Owners on Record
PHONAK AG
Past Owners on Record
ALLEGRO, SILVIA
Past Owners that do not appear in the "Owners on Record" listing will appear in other documentation within the application.
Documents

To view selected files, please enter reCAPTCHA code :



To view images, click a link in the Document Description column. To download the documents, select one or more checkboxes in the first column and then click the "Download Selected in PDF format (Zip Archive)" or the "Download Selected as Single PDF" button.

List of published and non-published patent-specific documents on the CPD .

If you have any difficulty accessing content, you can call the Client Service Centre at 1-866-997-1936 or send them an e-mail at CIPO Client Service Centre.


Document
Description 
Date
(yyyy-mm-dd) 
Number of pages   Size of Image (KB) 
Abstract 2003-08-27 2 94
Claims 2003-08-27 6 164
Drawings 2003-08-27 5 57
Description 2003-08-27 18 651
Representative Drawing 2003-08-27 1 9
Cover Page 2003-10-29 2 45
Representative Drawing 2011-02-25 1 14
Cover Page 2011-02-25 1 48
Description 2010-11-10 21 749
Claims 2010-11-10 7 270
Drawings 2010-11-10 5 77
Fees 2007-01-04 1 43
PCT 2003-08-27 2 64
Assignment 2003-08-27 5 125
Correspondence 2003-10-27 1 26
Fees 2003-12-09 1 27
Assignment 2003-12-03 2 54
Fees 2004-12-13 1 28
Fees 2005-12-28 1 31
Prosecution-Amendment 2006-12-28 1 43
Fees 2007-12-27 1 44
Fees 2008-12-04 1 55
Fees 2009-11-18 1 52
Prosecution-Amendment 2010-05-10 3 89
Correspondence 2010-08-10 1 45
Prosecution-Amendment 2010-11-10 23 671
Correspondence 2010-12-03 1 80
Correspondence 2011-01-13 2 54