Sélection de la langue

Search

Sommaire du brevet 2235275 

Énoncé de désistement de responsabilité concernant l'information provenant de tiers

Une partie des informations de ce site Web a été fournie par des sources externes. Le gouvernement du Canada n'assume aucune responsabilité concernant la précision, l'actualité ou la fiabilité des informations fournies par les sources externes. Les utilisateurs qui désirent employer cette information devraient consulter directement la source des informations. Le contenu fourni par les sources externes n'est pas assujetti aux exigences sur les langues officielles, la protection des renseignements personnels et l'accessibilité.

Disponibilité de l'Abrégé et des Revendications

L'apparition de différences dans le texte et l'image des Revendications et de l'Abrégé dépend du moment auquel le document est publié. Les textes des Revendications et de l'Abrégé sont affichés :

  • lorsque la demande peut être examinée par le public;
  • lorsque le brevet est émis (délivrance).
(12) Brevet: (11) CA 2235275
(54) Titre français: SYSTEME DE COMPRESSION POUR SONS REPETITIFS
(54) Titre anglais: REPETITIVE SOUND COMPRESSION SYSTEM
Statut: Périmé
Données bibliographiques
(51) Classification internationale des brevets (CIB):
  • G10L 19/00 (2006.01)
  • G10L 19/04 (2006.01)
  • G10L 21/00 (2006.01)
(72) Inventeurs :
  • YU, ALFRED (Etats-Unis d'Amérique)
(73) Titulaires :
  • FACEBOOK, INC. (Etats-Unis d'Amérique)
(71) Demandeurs :
  • AMERICA ONLINE, INC. (Etats-Unis d'Amérique)
(74) Agent: LAVERY, DE BILLY, LLP
(74) Co-agent:
(45) Délivré: 2005-12-20
(86) Date de dépôt PCT: 1996-10-21
(87) Mise à la disponibilité du public: 1997-04-24
Requête d'examen: 2001-10-02
Licence disponible: S.O.
(25) Langue des documents déposés: Anglais

Traité de coopération en matière de brevets (PCT): Oui
(86) Numéro de la demande PCT: PCT/US1996/016693
(87) Numéro de publication internationale PCT: WO1997/015046
(85) Entrée nationale: 1998-04-17

(30) Données de priorité de la demande:
Numéro de la demande Pays / territoire Date
08/545,487 Etats-Unis d'Amérique 1995-10-20

Abrégés

Abrégé français

Un système permettant de comprimer les sons fait appel à trois codes distincts utilisant la technique de codage (110) pour générer des codes ou des symboles (120) correspondant à la parole compressée.


Abrégé anglais





A sound compressing system uses
three separate codebooks using the coding
process (110), to output codes or symbols
(120) indicative of the compressed
speech.

Revendications

Note : Les revendications sont présentées dans la langue officielle dans laquelle elles ont été soumises.





-14-


CLAIMS


1. A sound compression system for generating a
compressed sound output, comprising:
a first processing element characterizing a
first sound representation and generating a first
characterization result;
a first comparison element generating a first
comparison result by at least comparing a first comparison
input related to the first sound representation with a
second comparison input related to the first
characterization result, and determining whether further
processing is desirable based on whether the first
comparison result satisfies a first predetermined
threshold criteria: and
an output element generating a compressed sound
output based on at least the first comparison result.

2. The system as in claim 1, further comprising a
second processing element characterizing a second sound
representation and generating a second characterization
result only if the first comparison result satisfies the
first predetermined threshold criteria.

3. The system as in claim 2, wherein the compressed
sound output includes the second characterization result
and excludes the first characterization result when the
first comparison result satisfies the first predetermined
output.

4. The system as in claim 2, wherein said first
processing element comprises a first codebook that




-15-

includes first codes for characterizing the first sound
representation, and said second processing element
comprises a second codebook that includes second codes for
characterizing the second sound representation.

5. The system as in claim 4, wherein said second
codebook includes at least one code that differs from the
first codes of said first codebook.

6. The system as in claim 4, wherein the first and
second characterization results each comprise an
indication of a closest matching code and a residual.

7.The system as in claim 2, wherein the first
sound representation and the second sound representation
comprise perceptually weighted error values.

8. The system as in claim 7, wherein the first
comparison input and the second comparison input used for
comparison by the first comparison element comprise
perceptually weighted error values.

9. The system as in claim 7, wherein the first
comparison input and the second comparison input used for
comparison by the first comparison element comprise non-
perceptually weighted error values.

10. The system as in claim 2, further comprising a
second comparison element generating a second comparison
result by at least comparing a third comparison input
related to the second sound representation with a fourth
comparison input related to the second characterization




-16-

result, and determining whether further processing is
desirable based on whether the second comparison result
satisfies a second predetermined threshold criteria.

11. The system as in claim 10, further comprising a
third processing element characterizing a third sound
representation and generating a third characterization
result only if the second comparison result satisfies the
second predetermined threshold criteria.

12. The system as in claim 11, wherein the
compressed sound output includes the third
characterization result and excludes the first
characterization result and the second characterization
result if the second comparison result satisfies the
second predetermined threshold.

13. The system as in claim 11, wherein:
said first processing element comprises an
adaptive vector quantization codebook;
said second processing element comprises a real
pitch vector quantization codebook that includes a
plurality of pitches indicative of voices; and
said third processing element comprises a noise
vector quantization codebook that includes a plurality of
noise vectors.

14. The system as in claim 11, wherein the first
sound representation characterized by the first processing
element comprises a perceptually weighted difference
between a first received value indicative of a previous




-17-

sound and a second received value indicative of a new
sound.
15. The system as in claim 14, wherein the second
sound representation characterized by the second
processing element comprises a perceptually weighted
residual of the first processing element; and wherein the
third sound representation characterized by the third
processing element comprises a perceptually weighted
residual of the second processing element.

16. The system as in claim 10, wherein the second
comparison element compares the third comparison input and
the fourth comparison input only if the first comparison
satisfies the first predetermined threshold.

17. The system as in claim 1, wherein said first
comparison element performs a correlation function upon
the first comparison input and the second comparison
input, and the first comparison result is a correlation
metric value.

18. The system as in claim 1, wherein the first
sound representation comprises a difference between a
first received value indicative of a previous sound, and a
second received value indicative of a new sound.

19. A method of compressing sound, comprising the
steps of:
characterizing a first sound representation to
produce a first characterization result that includes at
least a first processing element residual;




-18-

generating a first comparison result by at least
correlating a first comparison input related to the first
sound representation with a second comparison input
related to the first characterization result;
comparing the first comparison result to a first
predetermined threshold criteria;
determining whether further processing is
desirable based on whether the first comparison result
satisfies the first predetermined threshold criteria; and
generating a compressed sound output based on
the first comparison result.

20. The method as in claim 19, comprising using a
second processing element to characterize a second sound
representation and to generate a second characterization
result only if the first comparison result satisfies the
first predetermined threshold criteria.

21. The method as in claim 19, wherein the
compressed sound output includes the second
characterization result and excludes the first
characterization result if the first comparison result
satisfies the first predetermined threshold.

22. A sound compression system for generating a
compressed sound output, comprising:
a first processing element characterizing a
first sound representation and generating a first
characterization result;
a second processing element characterizing a
second sound representation and generating a second
characterization result;




-19-

a first comparison element generating a first
comparison result by at least comparing a first comparison
input related to the first sound representation with a
second comparison input related to the second sound
representation and determining contents of the compressed
sound output based on whether the first comparison result
satisfies a first predetermined threshold criteria; and
an output element generating a compressed sound
output based on at least the first comparison result;
wherein the compressed sound output is related
to the first characterization result and the second
characterization result only if the first comparison
result satisfies the first predetermined threshold
criteria.

23. The system as in claim 22, further comprising:
a third processing element characterizing a
third sound representation and generating a third
characterization result; and
a second comparison element generating a second
comparison result by at least comparing the second
comparison input related to the second sound
representation and a third comparison input related to the
third sound representation and determining the contents of
the compressed sound output based on whether a second
comparison result satisfies a second predetermined
threshold criteria;
wherein the output generates the compressed
sound output based on at least the first comparison result
and the second comparison result, and the compressed sound
output is generated based on the first characterization
result, the second characterization result, and the third




-20-

characterization result only if the second comparison
result satisfies the second predetermined threshold
criteria.

24. A sound compression system for generating a
compressed sound output, comprising:
a first processing element characterizing a
first sound representation and generating a first
characterization result;
a second processing element characterizing a
second sound representation and generating a second
characterization result;
a first comparison element generating a first
comparison result by at least comparing a first comparison
input related to the first sound representation with a
second comparison input related to the first
characterization result and determining whether further
processing is desirable based on whether the first
comparison result satisfies a first predetermined
threshold criteria; and
an output element generating a compressed sound
output based on at least the first comparison result;
wherein said second processing element further
characterizes the second sound representation and
generates the second characterization result only after
the first comparison result satisfies the first
predetermined threshold criteria.

25. The system as in claim 24, further comprising:
a third processing element characterizing a
third sound representation and generating a third
characterization result; and




-21-

a second comparison element generating a second
comparison result by at least comparing a third comparison
input related to the second sound representation and a
fourth comparison input related to the second
characterization result and determining whether further
processing is desirable based on whether the second
comparison result satisfies a second predetermined
threshold criteria;
wherein the output element generates the
compressed sound output based on at least the first
comparison result and the second comparison result.

Description

Note : Les descriptions sont présentées dans la langue officielle dans laquelle elles ont été soumises.



CA 02235275 1998-04-17
r
WO 97/I5046 PCT/US96116693
- 1 -
REPETITIVE.SOUND COMPRESSION SYSTEM
' Field of the Invention
The present invention teaches a system for
' S compressing quasi-periodic sound by comparing it to
presampled portions in a codebook.
Background and Summar
Many sound compression schemes take advantage of
the repetitive nature of everyday sounds. For example,
the standard human voice coding device or "vocoder", is
often used for compressing and encoding human voice
sounds. A vocoder is a class of voice coder/decoders
that models the human vocal tract.
A typical vocoder models the input sound as two
parts: the voice sound known as V, and the unvoice sound
known as U. The channel through which these signals are
conducted is modelled as a lossless cylinder. The output
speech is compressed based on this model.
Strictly speaking, speech is not periodic.
However, the voice part of speech is often labeled as
quasi-periodic due to its pitch frequency. The sounds
produced during the un-voiced region, are highly random.
Speech is always referred to as non-stationary and
stochastic. Certain parts of speech may have redundancy
and perhaps correlated to some prior portion of speech to
some extent, but they are not simply repeated.
The main intent of using a vocoder is to find ways
to compress the source, as opposed to performing
compression of the result. The source in this case is
the excitation formed by glottal pulses. The result is
the human speech we hear. However, there are many ways
that the human vocal tract can modulate the glottal
pulses to form human voice. Estimations of the glottal
pulses are predicted and then coded. Such a model


CA 02235275 1998-04-17
WO 97/15046 ~ PCT/U896/16693
_ 2 _
reduces the dynamic range of the resulting speech, hence
rendering the speech more compressible.
More generally, the special kind of speech '
filtering can remove speech portions that are not
perceived by the human ear. With the vocoder model in '
place, a residue portion of the speech can be made
compressible due to its lower dynamic range.
The term "residue" has multiple meanings. It
generally refers to the output of the analysis filter,
ZO the inverse of the synthesis filter which models the
vocal tract. In the present situation, residue takes on
multiple meanings at different stages: at stage 1- after
the inverse filter (all zero filter), stage 2: after the
long term pitch predictor or the so-called adaptive pitch
VQ, stage 3: after the pitch codebook, and at stage 4:
after the noise codebook. The term "residue" as used
herein literally refers to the remaining portion of the
speech by-product which results from previous processing
stages.
The preprocessed speech is then encoded. A
typical vocoder uses an 8 kHz sampling rate at Z6 bits
per sample. These numbers are nothing magic, however -
they are based on the bandwidth of telephone lines.
The sampled information is further processed by a
speech codec which outputs an 8 kHz signal. That signal
may be post-processed, which may be the opposite of the
input processing. Other further processing that is
designed to further enhance the quality and character of
the signal may be used.
The suppression of noise also models the way that
humans perceives sound. Different weights are used at
different times according to the strength of speech both
in the frequency and time domain. The masking properties
of human hearing cause loud signals at different
frequencies to mask the effect of low level signals


CA 02235275 1998-04-17
WO 97/15046 PC~'/TJS96/i6693
- 3 -
around those frequencies. This is also true in the time
domain. The result is that more noise can be tolerated
during that portion of time and frequency. This allows
us to pay more attention elsewhere. This is called
' 5 "perceptual weighting" - it allows us to pick vectors
which are perpectually more effective.
The human vocal tract can be (and is) modeled by a
set of lossless cylinders with varying diameters.
Typically, it a.s modeled by an 8 to 12th order all-pole
filter 1/A(Z). Its inverse counterpart A(Z) is an all-
zero filter with the same order. Output speech is
reproduced by exciting the synthesis filter 1/A(z) with
the excitation. The excitation, or glottal pulses is
estimated by inverse filtering the speech signal with the
inverse filter A(z). A digital signal processor often
models the synthesis filter as the transfer function H(V)
- 1/A(z). This means that this model is an all-pole
process. Ideally, the model is more complicated, and
includes both poles and zeros.
Much of the compressibility of speech comes from
its quasi-periodicity. Speech is quasi-periodic due to
its pitch frequency around voice sound. Male speech
usually has a pitch between 50 and 100 Hz. Female speech
usually has a pitch above 100 Hz.
While the above describes compression systems for
voice coding, the same general principles are used to
code and compress other similar kinds of sound.
Various techniques are known for improving the
model. Each of these techniques, however, increases the
necessary bandwidth to carry the signal. This produces a
tradeoff between bandwidth of the compressed signal and
quality of the non=steady-state sound.
These problems are overcome according to the
present invention by new features.


CA 02235275 1998-04-17
WO 97/15046 PCT/US96/16693
- 4 -
A first aspect of the present invention includes a
new architecture for coding which allows various coding
and monitoring advantages. The disclosed system of the '
present invention includes new kinds of codebooks for
coding. These new codebooks allow faster consequence to '
changes in the input sound stream. Essentially, these
new codebooks use the same software routine many times
over, to improve coding efficiency.
Brief Description of the Drawings
These and other aspects of the present invention
will now be described with reference to the accompanying
drawings in which:,
Figure 1 shows a block diagram of the basic
vocoder of the present invention; and
Figure 2 the advanced codebook technique of the
present invention.
~escrit~tion of the Preferred Embodiments
Figure 1 shows the advanced vocoder of the present
invention. The current speech codec uses a special class
of vocoder which operates based on LPC (linear
predictive coding). All future samples are being
predicted by a linear combination of previous samples and
the difference between predicted samples and actual
samples. As described above, this is modeled after a
lossless tube also known as an all-pole model. The model
presents a relative reasonable short term prediction of
speech.
The above diagram depicts such a model, where the
input to the lossless tube is defined as an excitation
which is further modeled as a combination of periodic
pulses and random noise.
A drawback of the above model is that the vocal
tract does not behave exactly as a cylinder and is not

CA 02235275 1998-04-17
WO 97/15046 PCTlUS96/16693
- 5 -
lossless. The human vocal tract also has side passages
such as the nose.
' Speech to be coded 100 is input to an analysis
block 102 which analyzes the content of the speech as
described herein. The analysis block produces a short
term residual alone with other parameters.
Analysis in this case refers as LPC analysis as
depicted above in our lossless tube model, that includes,
for example, computation of windowing, autocorrelation,
Durbin's recursion, and computation of predictive
coefficients are performed. In addition, filtering
incoming speech with the analysis filter based on the
computed predictive coefficients will generate the
residue, the short term residue STA res 104.
This short term residual 104 is further coded by
the coding process 110, to output codes or symbols 120
indicative of the compressed speech. Coding of this
preferred embodiment involves performing three codebook
searches, to minimize the perceptually-weighted error
signal. This process is done in a cascaded manner such
that codebook searches are done one after another.
The current codebooks used are all shape gain VQ
codebooks. The perceptually-weighted filter is generated
adaptively using the predictive coefficients from the
current sub-frame. The filter input is the difference
between the residue from previous stage versus the shape
gain vector from the current stage, also called the
residue, is used for next stage. The output of this
filter is the perceptually weighted error signal. This
operation is shown_and explained in more detail with
reference to Figure 2. Perceptually-weighted error from
each stage is used as a target for the searching in next
stage.
' The compressed speech or a sample thereof 122 is
also fed back to a synthesizer 124, which reconstitutes a


CA 02235275 1998-04-17
WO 97/15046 PCT/LTS96/16693
- 6 -
reconstituted original block 126. The synthesis stage
decodes the linear combination of the vectors to form a
reconstruction residue, the result is used to initialize "
the state of the next search in next sub-frame.
Comparison of the original versus the
reconstructed sound results in an error signal which will
drive subsequent codebook searches to further minimize
such perceptually-weighted error. The objective of the
subsequent coder is to code this residue very
efficiently.
The reconstituted block 126 indicates what would
be received at the~receiving end. The difference between
the input speech 100 and the reconstituted speech 126
hence represents an error signal 132.
This error signal is perceptually weighted by
weighting block 134. The perceptual weighting according
to the present invention weights the signal using a model
of what would be heard by the human ear. The
perceptually-weighted signal 136 is then heuristically
processed by heuristic processor 140 as described herein.
Heuristic searching techniques are used which take
advantage of the fact that some codebooks searches are
unnecessary and as a result can be eliminated. The
eliminated codebooks are typically codebooks down the
search chain. The unique process of dynamically and
adaptively performing such elimination is described
herein.
The selection criterion chosen is primarily based
on the correlation_between the residue from a prior stage
versus that of the current one. If they are correlated
very well, that means the shape-gain vQ contributes very
little to the process and hence can be eliminated. On
the other hand, if it does not correlate very well the
contribution from the codebook is important hence the "
index shall be kept and used.


CA 02235275 1998-04-17
WO 97/15046 PC'i'/US96/16693
_ 7 _
Other techniques such as stopping the search when
an adaptively predetermined error threshold has been
' reached, and asymptotic searches are means of speeding up
the search process and settling with a sub-optimal
result. The heuristically-processed signal 138 is used
as a control for the coding process 110 to further
improve the coding technique.
This general kind of filtering processing is well
known in the art, and it should be understood that the
present invention includes improvements on the well known
filtering systems in the art.
The coding according to the present invention uses
the codebook types and architecture shown in Figure 2.
This coding includes three separate codebooks: adaptive
vector quantatization (VQ) codebook 200, real pitch
codebook 202, and noise codebook 204. The new
information, or residual 104, is used as a residual to
subtract from the code vector of the subsequent block.
ZSR (Zero state response) is a response with zero input.
The ZSR is a response produced when the code vector is
all zeros. Since the speech filter and other associated
filters are IIR (infinite impulse response) filters, even
when there is no input, the system will still generate
output continuously. Thus, a reasonable first step for
codebook searching is to determine whether it is
necessary to perform any more searches, or perhaps no
code vector is needed for this subframe.
To clarify this point, any prior event will have a
residual effect. Although that effect will diminish as
time passes, the effect is still present well into the
next adjacent sub-frames or even frames. Therefore, the
' speech model must take these into consideration. If the
speech signal present in the current frame is just a
residual effect from a previous frame, then the
perceptually-weighted error signal Eo will be very low or


CA 02235275 1998-04-17
WO 97/I5046 PCT/US96/16693
_ g _
even be zero. Note that, because of noise or other
system issues, all-zero error conditions will almost
never occur.
eo = STA res - ~. The reason ~ vector is used is '
for completeness to indicate zero state response. This
is a set-up condition for searches to be taken place. If
E~ is zero, or approaches zero, then no new vectors are
necessary.
EO is used-to drive the next stage as the ~~target~~
of matching for the next stage. The objective is to find
a vector such that E1 is very close to or equal to zero,
where E1 is the perceptually weighted error from el, and
e1 is the difference between e0-vector(i). This process
goes on and on through the various stages.
The preferr-ed mode of the present invention uses a
preferred system with 240 samples per frame. There are
four subframes per frame, meaning that each subframe has
60 samples.
A VQ search for each subframe is done. This VQ
search involves matching the 60-part vector with vectors
in a codebook using a conventional vector matching
system.
Each of these vectors must be defined according to
an equation. The basic equation used is of the form that
2 5 G$Ai + GbBj + G~Ck.
Since the objective is to come up with a minimum
perceptually weighted error signal E3 by selecting
vectors Ai, Bj, and Ck along with the corresponding gain
Ga, Gb, and Gc. This does NOT imply the vector sum of
3 0 Ga"Ai + GbBj + G~CE = STA_res .
In fact, it is almost never true with the
exception of silence.
The error value Eo is preferably matched to the
values in the AVQ codebook 200. This is a conventional


CA 02235275 1998-04-17
WO 97/15046 PCT/US961I6693
_ g _
kind of codebook where samples of previous reconstructed
speech, e.g., the last 20 ms, is stored. A closest match
' is found. The value el (error signal number 1? represents
the leftover between the matching of Eo with AVQ 200.
According to the present invention, the adaptive
vector quantizer stores a 20 ms history of the
reconstructed speech. This history is mostly for pitch
prediction during voice frame. The pitch of a sound
signal does not change quickly. ,The new signal will be
20 closer to those values in the AVQ than they will to other
things. Therefore, a close match is usually expected.
Changes in voice, however, or new users entering a
conversation, will degrade the quality of the matching.
According to the present invention, this degraded
matching is compensated using other codebooks.
The second codebook used according to the present
invention is a real pitch codebook 202. This real pitch
codebook includes code entries for the most usual
pitches. The new pitches represent most possible pitches
of human voices, preferably from 200 Hz down. The
purpose of this second codebook is to match to a new
speaker and for startup/voice attack purposes. The pitch
codebook is intended for fast attack when voice starts or
when a new person entering the room with new pitch
information not found in the adaptive codebook or the so-
called history codebook. Such a fast attack method
allows the shape of speech to converge more quickly and
allows matches more closely to that of the original
waveform during the voice region.
Usually when a new speaker enters the sound field,
AVQ will have a hard time performing the matching.
Hence, El is still very high. During this initial time,
therefore, there are large residuals, because the
matching in the codebook is very poor. The residual El
represents the new speaker's pitch weighted error. This


CA 02235275 1998-04-17
WO 97/15046 PCT/US96/16693
- 10 -
residual is matched to the pitch in the real pitch
codebook 202.
The conventional method uses some form of random '
pulse codebook which is slowly shaped via the adaptive
process in 200 to match that of the original speech. °
This method takes too long to converge. Typically it
takes about 6 sub-frames and causes major distortion
around the voice attack region and hence suffers quality
loss.
The inventors have found that this matching to-the
pitch codebook 202 causes an almost immediate re-locking
of the signal. For example, the signal might be re-
locked in a single period, where one sub-frame period =
60 samples = 60/8000 = 7.5ms. This allows accurate
representation of the new voice during the transitional
period in the early part of the time while the new
speaker is talking.
The noise codebook 204 is used to pick up the
slack and also help shape speech during the unvoiced
period.
As described above, the G's represent amplitude
adjustment characteristics, and A, B and C are vectors.
The codebook for the AVQ preferably includes 256
entries. The codebooks for the pitch and noise each
include 512 entries.
The system of the present invention uses three
codebooks. However, it should be understood that either
the real pitch codebook or the noise codebook could be
used without the other.
Additional processing is carried out according to
the present invention under the characteristic called
heuristics. As described above, the three-part codebook
of the present invention improves the efficiency of
matching. However, this of course is only done at the
expense of more transmitted information and hence less


CA 02235275 1998-04-17
WO 97/15046 PCTl~JS96/16693
- 11 -
compression efficiency. Moreover, the advantageous
architecture of the present invention allows viewing and
processing each of the error values eo-e3 and Eo-E3. These
error values tell us various things about the signals,
including the degree of matching. For example, the error
value Eo being 0 tells us that no additional processing is
necessary. Similar information can be obtained from
errors Eo-E3. According to the present invention, the
system determines the degree of mismatching to the
codebook, to obtain an indication of whether the real
pitch and noise codebooks are necessary. Real pitch and
noise codebooks are not always used. These codebooks are
only used when some new kind or character of sound enters
the field.
I5 The codebooks are adaptively switched in and out
based on a calculation carried out with the output of the
codebook.
The preferred technique compares Eo to E1. Since
the values are vectors, the comparison requires
correlating the two vectors. Correlating two vectors
ascertains the degree of closeness therebetween. The
result of the correlation is a scalar value that
indicates how good the match is. If the correlation
value is low, it indicates that these vectors are very
different. This implies the contribution from this
codebook is significant, therefore, no additional
codebook searching steps are necessary. On the contrary,
if the correlation value is high, the contribution from
this codebook is not needed, then further processings are
required. Accordingly, this aspect of the invention
compares the two error values to determine if additional
codebook compensation is necessary. If not, the
additional codebook compensation is turned off to
increase the compression.


CA 02235275 1998-04-17
WO 97/15046 PCT/US96/16693
- 12 -
A similar operation can be carried out between El
and EZ to determine if the noise codebook is necessary.
Moreover, those having ordinary skill in the art
will understand that this can be modified other ways
using the general technique that a determination of
whether the coding is sufficient is obtained, and the
codebooks are adaptively switched in or out to further
improve the compression rate and/or matching.
Additional heuristics are also used according to
the present invention to speed up the search. Additional
heuristics to speed up codebook searches are:
a) a subset of codebooks is searched and a partial
perceptually weighted error Ex is determined. If Ex is
within a certain predetermined threshold, matching is
stopped and decided to be good enough. Otherwise we
search through the end. Partial selection can be done
randomly, or through decimated sets.
b) An asymptotic way of computing the perceptually
weighted error is used whereby computation is simplified.
c) Totally skip the perceptually weighted error
criteria and minimize "e" instead. In such case, an
early-out algorithm is available to further speed up the
computation.
Another heuristic is the voice or unvoice
detection and its appropriate processing. The
voice/unvoice can be determined during preprocessing.
Detection is done, for example, based on zero crossings
and energy determinations. The processing of these
sounds is done differently depending on whether the input
sound is voice or unvoice. For example, codebooks can be
switched in depending on which codebook is effective.
Different codebooks can be used for different
purposes, including but not limited to the well known
technique of shape gain vector quantatization and join
optimization. An increase in the overall compression


CA 02235275 1998-04-17
WO 97/I5046 PCT/US96/16693
- 13 -
rate is obtainable based on preprocessing and switching
in and out the codebooks.
Although only a few embodiments have been
described in detail above, those having ordinary skill in
the art will certainly understand that many modifications
are possible in the preferred embodiment without
departing from the teachings thereof.
All such modifications are intended to be
encompassed within the following claims.

Dessin représentatif
Une figure unique qui représente un dessin illustrant l'invention.
États administratifs

Pour une meilleure compréhension de l'état de la demande ou brevet qui figure sur cette page, la rubrique Mise en garde , et les descriptions de Brevet , États administratifs , Taxes périodiques et Historique des paiements devraient être consultées.

États administratifs

Titre Date
Date de délivrance prévu 2005-12-20
(86) Date de dépôt PCT 1996-10-21
(87) Date de publication PCT 1997-04-24
(85) Entrée nationale 1998-04-17
Requête d'examen 2001-10-02
(45) Délivré 2005-12-20
Expiré 2016-10-21

Historique d'abandonnement

Date d'abandonnement Raison Reinstatement Date
2002-10-21 Taxe périodique sur la demande impayée 2003-01-20

Historique des paiements

Type de taxes Anniversaire Échéance Montant payé Date payée
Le dépôt d'une demande de brevet 300,00 $ 1998-04-17
Taxe de maintien en état - Demande - nouvelle loi 2 1998-10-21 100,00 $ 1998-09-30
Enregistrement de documents 100,00 $ 1999-07-07
Enregistrement de documents 100,00 $ 1999-07-07
Taxe de maintien en état - Demande - nouvelle loi 3 1999-10-21 100,00 $ 1999-10-04
Taxe de maintien en état - Demande - nouvelle loi 4 2000-10-23 100,00 $ 2000-10-04
Taxe de maintien en état - Demande - nouvelle loi 5 2001-10-22 150,00 $ 2001-09-20
Requête d'examen 400,00 $ 2001-10-02
Rétablissement: taxe de maintien en état non-payées pour la demande 200,00 $ 2003-01-20
Taxe de maintien en état - Demande - nouvelle loi 6 2002-10-21 150,00 $ 2003-01-20
Taxe de maintien en état - Demande - nouvelle loi 7 2003-10-21 150,00 $ 2003-10-09
Taxe de maintien en état - Demande - nouvelle loi 8 2004-10-21 200,00 $ 2004-10-04
Taxe finale 300,00 $ 2005-09-07
Taxe de maintien en état - Demande - nouvelle loi 9 2005-10-21 200,00 $ 2005-10-03
Taxe de maintien en état - brevet - nouvelle loi 10 2006-10-23 250,00 $ 2006-10-02
Taxe de maintien en état - brevet - nouvelle loi 11 2007-10-22 250,00 $ 2007-10-01
Taxe de maintien en état - brevet - nouvelle loi 12 2008-10-21 250,00 $ 2008-09-30
Taxe de maintien en état - brevet - nouvelle loi 13 2009-10-21 250,00 $ 2009-10-01
Taxe de maintien en état - brevet - nouvelle loi 14 2010-10-21 250,00 $ 2010-09-16
Taxe de maintien en état - brevet - nouvelle loi 15 2011-10-21 450,00 $ 2011-09-20
Taxe de maintien en état - brevet - nouvelle loi 16 2012-10-22 650,00 $ 2012-11-27
Enregistrement de documents 100,00 $ 2013-05-08
Enregistrement de documents 100,00 $ 2013-05-08
Enregistrement de documents 100,00 $ 2013-05-08
Taxe de maintien en état - brevet - nouvelle loi 17 2013-10-21 450,00 $ 2013-09-13
Taxe de maintien en état - brevet - nouvelle loi 18 2014-10-21 450,00 $ 2014-10-01
Taxe de maintien en état - brevet - nouvelle loi 19 2015-10-21 450,00 $ 2015-09-30
Titulaires au dossier

Les titulaires actuels et antérieures au dossier sont affichés en ordre alphabétique.

Titulaires actuels au dossier
FACEBOOK, INC.
Titulaires antérieures au dossier
AMERICA ONLINE, INC.
AOL INC.
AOL LLC
JOHNSON-GRACE COMPANY
YU, ALFRED
Les propriétaires antérieurs qui ne figurent pas dans la liste des « Propriétaires au dossier » apparaîtront dans d'autres documents au dossier.
Documents

Pour visionner les fichiers sélectionnés, entrer le code reCAPTCHA :



Pour visualiser une image, cliquer sur un lien dans la colonne description du document. Pour télécharger l'image (les images), cliquer l'une ou plusieurs cases à cocher dans la première colonne et ensuite cliquer sur le bouton "Télécharger sélection en format PDF (archive Zip)" ou le bouton "Télécharger sélection (en un fichier PDF fusionné)".

Liste des documents de brevet publiés et non publiés sur la BDBC .

Si vous avez des difficultés à accéder au contenu, veuillez communiquer avec le Centre de services à la clientèle au 1-866-997-1936, ou envoyer un courriel au Centre de service à la clientèle de l'OPIC.


Description du
Document 
Date
(yyyy-mm-dd) 
Nombre de pages   Taille de l'image (Ko) 
Revendications 2004-08-31 8 270
Revendications 1998-04-17 2 46
Dessins 1998-04-17 2 24
Page couverture 1998-07-28 1 25
Dessins représentatifs 1998-07-28 1 5
Revendications 2001-10-03 5 241
Abrégé 1998-04-17 1 35
Description 1998-04-17 13 584
Page couverture 2005-11-22 1 29
Dessins représentatifs 2005-11-24 1 6
Taxes 2005-10-03 1 33
Poursuite-Amendment 2004-08-31 10 318
Taxes 1998-09-30 1 47
Cession 1998-04-17 4 105
PCT 1998-04-17 8 281
Correspondance 1998-06-30 1 30
Cession 1999-07-07 5 212
Correspondance 1999-09-03 1 2
Cession 1999-09-15 1 35
Poursuite-Amendment 2001-10-03 6 269
Poursuite-Amendment 2001-10-02 1 23
Poursuite-Amendment 2002-03-27 2 33
Taxes 2003-01-20 1 43
Taxes 2003-10-09 1 39
Taxes 2000-10-04 1 41
Taxes 2001-09-20 1 39
Taxes 1999-10-04 1 44
Poursuite-Amendment 2004-03-18 3 90
Taxes 2004-10-04 1 37
Correspondance 2005-09-07 1 36
Cession 2013-05-08 234 10 040