Sélection de la langue

Search

Sommaire du brevet 2463162 

Énoncé de désistement de responsabilité concernant l'information provenant de tiers

Une partie des informations de ce site Web a été fournie par des sources externes. Le gouvernement du Canada n'assume aucune responsabilité concernant la précision, l'actualité ou la fiabilité des informations fournies par les sources externes. Les utilisateurs qui désirent employer cette information devraient consulter directement la source des informations. Le contenu fourni par les sources externes n'est pas assujetti aux exigences sur les langues officielles, la protection des renseignements personnels et l'accessibilité.

Disponibilité de l'Abrégé et des Revendications

L'apparition de différences dans le texte et l'image des Revendications et de l'Abrégé dépend du moment auquel le document est publié. Les textes des Revendications et de l'Abrégé sont affichés :

  • lorsque la demande peut être examinée par le public;
  • lorsque le brevet est émis (délivrance).
(12) Brevet: (11) CA 2463162
(54) Titre français: PROCEDE ET DISPOSITIF DE TRAITEMENT D'IMAGES
(54) Titre anglais: METHOD AND APPARATUS FOR PROCESSING IMAGES
Statut: Durée expirée - au-delà du délai suivant l'octroi
Données bibliographiques
(51) Classification internationale des brevets (CIB):
  • G06T 05/50 (2006.01)
(72) Inventeurs :
  • HANNA, KEITH J. (Etats-Unis d'Amérique)
  • BERGEN, JAMES R. (Etats-Unis d'Amérique)
  • KUMAR, RAKESH (Etats-Unis d'Amérique)
  • SAWHNEY, HARPREET (Etats-Unis d'Amérique)
  • LUBIN, JEFFREY (Etats-Unis d'Amérique)
  • ADKINS, SEAN M. (Etats-Unis d'Amérique)
(73) Titulaires :
  • IMAX CORPORATION
(71) Demandeurs :
  • IMAX CORPORATION (Canada)
(74) Agent: GOWLING WLG (CANADA) LLP
(74) Co-agent:
(45) Délivré: 2009-10-27
(22) Date de dépôt: 1999-08-30
(41) Mise à la disponibilité du public: 2000-03-09
Requête d'examen: 2004-08-12
Licence disponible: S.O.
Cédé au domaine public: S.O.
(25) Langue des documents déposés: Anglais

Traité de coopération en matière de brevets (PCT): Non

(30) Données de priorité de la demande:
Numéro de la demande Pays / territoire Date
09/384,118 (Etats-Unis d'Amérique) 1999-08-27
60/098,368 (Etats-Unis d'Amérique) 1998-08-28
60/123,615 (Etats-Unis d'Amérique) 1999-03-10

Abrégés

Abrégé français

Procédé et appareil de traitement d'image et, plus particulièrement, procédé et appareil pour améliorer la qualité d'une image.


Abrégé anglais

The invention relates to an image processing method and apparatus and, more particularly, the invention relates to a method and apparatus for enhancing the quality of an image.

Revendications

Note : Les revendications sont présentées dans la langue officielle dans laquelle elles ont été soumises.


-23-
1. Apparatus for enhancing an image comprising :
a first imaging device for producing images at a first resolution;
a second imaging device for producing images at a second resolution; and
an image processor for aligning said images from said first and second
imaging devices to form a plurality of aligned images in a temporal image
sequence, computing a flow estimation for each of the aligned images, and
using the flow estimation to enhance regions in a first aligned image with
information from at least one other aligned image.
2. The apparatus of claim 1 wherein said image processor comprises:
an image flow generator.
3. The apparatus of claim 2 wherein said image flow generator is a parallax
computer.
4. The apparatus of claim 3 wherein said parallax computer further comprises
one or more
augmentation modules selected from the group consisting of:
a module for dividing the images into tiles, a depth correlator, a module
which performs Just
Noticeable Differences, a correspondence checker, and a blank area avoidance
module.

Description

Note : Les descriptions sont présentées dans la langue officielle dans laquelle elles ont été soumises.


CA 02463162 2006-01-26
WO 00/13142 PCTNS99/19705
-1-
METHOD AND APPARATUS FOR PROCESSING IMAGES
The invention relates to an image processing method and apparatus
and, more particularly, the invention relates to a method and apparatus
for enhancing the quality of an image.
BACKGROUND OF THE DISCLOSURE
For entertainment and other applications; it is useful to obtain high-
resolution stereo imagery of a scene so that viewers can visualize the
scene in three dimensions. To obtain such high-resolution imagery, the
commom practice of the prior art is to use two or more high-resolution
devices or cameras, displaced from each other. The first high-resolution
camera captures an image or image sequence, ' that can be merged with
other high-resolution images taken from a viewpoint different than the
first high-resolution camera, creating a stereo image of the scene.
However, creating stereo imagery with multiple high-resolution
cameras can be difficult and very expensive. The number of high-
resolution cameras used to record a scene can contribute significantly to
the cost of producing the stereo image scene. ' Additionally, high-
resolution cameras are large and unwieldy. As such, the high-resolution
cameras are not easy to move about when filming a scene. Consequently,
some viewpoints may not be able to be accommodated because of the size of
the high-resolution cameras, thus limiting the viewpoints available for
creating the stereo image.

CA 02463162 2004-03-25
WO 00/13142 PCT/US99/19705
-2-
Similarly, in other applications given a collection of captured digital
imagery, the need is to generate enhanced imagery for monocular or
binocular viewing Examples of such application are resolution
enhancement of video and other digital imagery, quality enhancement in
terms of enhanced focus, depth of field, color and brightness/contrast
enhancement; and creation of synthetic imagery from novel viewpoints
based on captured digital imagery and videos.
All the above applications involve combining multiple co-temporal
digital sensors (camera for example) andlor temporally separated sensors
for the purpose of creation of synthetic digital imagery. The various
applications can be broadly divided along the following lines (but are not
limited to these):
1. Creation of an enhanced digital image by processing one or more
fraanes of imagery from cameras and or other sensors which have
captured the imagery at the same time instant. The synthesized frame
represents the view of an enhanced synthetic camera located at the
position of one of the real sensors.
2. Creation of enhainced digital imagery by processing frames that
have been captured over time and space (multiple cameras/sensors
capturirig video imagery over time). The synthesized frames represent
enhanced synthetic cameras. located at the position of one or more of the
real sensors.
3. Creation of enhanced digital imagery by processing frames that
have been captured over time and space (multiple cameras/sensors
capturing video imagery over time). The synthesized frames represent
enhanced synthetic caineras that are located at positions other than those
of the real sensors.
Therefore, a need exists in the art for a method and apparatus for
creating a synthetic high-resolution image and / enhancing images using
only one high-resolution camera.

CA 02463162 2004-03-25
WO 00/13142 PCT/[7S99/29705
-3-
SUMNiARY OF THE INVENTION
The disadvantages associated with the prior art are overcome by the
present invention for a method and apparatus for accurately computing
image flow information as captured by imagery of a scene. The invention
computes the image flow information of each point in an image by
computing the image flow within windows that are offset with respect to
the point for which the image flow is being computed. Additionally, image
flow computations are performed over multiple frames of imagery to
ensure accuracy of the image flow computation and to facilitate correction
of occluded imagery.
In one illustrative embodiment of the invention, the irnage flow
computation is constrained to compute parallax information. The
imagery and parallax (or flow) information can be used to enharice
various image processing techniques such as image resolution
enhancement, enhancement of focus, depth of field, color, and brightness.
The parallax (or flow) information can also be used to generate a synthetic
high-resolution image that can be used in combination with the original
image to form a stereo image. Specifically, the apparatus comprises an
imaging device for producing images (e.g., video frame sequences) and a
scene sensing device for producing information regarding the imaged
scene. An image processor uses the information f'rom the scene sensing
device to process the images produced by the imaging device. This
processing produces parallax information regarding the imaged. scene.
The imagery from the imaging device and the parallax information can be
used to enhance ariy one of the above-mentioned image processing
applications.
The invention includes a method that is embodied in a software
routine, or a combination of software and hardware. The inventive
method comprises the steps of supplying image data having a first
resolution and supplying image information regarding the scene
represented by the image data. The image data and information are
processed by, for example, warping the first image data to form a
synthetic image having a synthetic view, where the viewpoint of the

CA 02463162 2004-03-25
WO 00/13142 PCT/tJS99/19705

synthetic image is different from the viewpoint represented in the image
data. The synthetic image and the original image can be used to compute
parallax information regarding the scene. By using multiple frames from
the original imagery and the synthetic view imagery, the inventive process
improves the accuracy of the parallax computation.
Alternate embodiinents of the invention include but are not limited
to, utilizing multiple sensors in addition to the scene sensing device to
provide greater amounts of scene data for use in enhancing the synthetic
image, using a displacement device in conjunction with the second
imaging device to create a viewpoint for the warped image that is at the
location of the displacement device, and using a range finding device as
the second imaging device to provide image depth information.
BRIEF DESCRIPTION OF THE D;?LWINGS
The teachings of the present invention can be readily understood by
considering the following detailed description in conjunction with the
. accompanying drawings, in which:
Fig. 1 depicts a block diagram of an imaging apparatus
incorporating the image analysis method and apparatus of the invention;
Fig. 2 depicts a block schematic of an imaging apparatus and an
image analysis method used to produce one embodiment of the subject
invention;
Fig. 3 is a flow chart of the parallax computation method;
Fig. 4 is a flow chart of the image warping method;
Fig. 5 depicts a block diagram of an imaging apparatus and an
image analysis method used to produce a second embodiment of the
subject invention;
Fig. 6 depicts a block diagram of an imaging apparatus and an
image analysis method used to produce a third embodiment of the subject
invention;
Fig. 7 depicts a schematic view of multiple offset windows as used to
compute parallax at points within an image; and

CA 02463162 2004-03-25
WO 00/13142 PCTNS99119705
-5-
Fig. 8 depicts an illustration for a process to compute a quality
measure for parallax computation accuracy,
To facilitate understanding, identical reference numerals have been
used, where possible, to designate identical elements that are common to
the figures.
DETAILED DESCRIPTION
FIG. 1 depicts a high-resolution synthetic image generation
apparatus 100 of the present invention. An input video sequence 112 is
supplied to a computer 102. The computer 102 comprises a central
processing unit (CPU) 104, support circuits 106, and memory 108.
Residing within the memory 108 is a high-resolution synthetic image
generation routine 110. The high-resolution synthetic image generation
routine 110 may alternately be readable froin another source such as a
floppy disk, CD, remote memory source or via a network. The computer
additionally is coupled to input/output accessories 118. As a brief
description of operation, an input video sequence 112 is supplied to the
computer 102, which after operation of the high-resolution synthetic
image generation routine 110, outputs a synthetic high-resolution image
114.
The high-resolution synthetic image generation. routine. 110
hereinafter referred to as the routine 110, can be understood in greater
detail by referencing Fig. 2. Although the process of the present invention
is discussed as being implemented as a software routine 110, some of the
method steps that are disclosed therein may be performed in hardware as
well as by the software coritroller. As such, the invention may be
implemented in software as executed upon a computer system, in
hardware as an application specific integrated circuit or other type of
hardware implementation, or a combination of software and hardware.
Thus, the reader should note that each step of the routine 110 should also
be construed as having an equivalent application specific hardware device
(module), or hardware device used in combination with software.

CA 02463162 2004-03-25
WO 00/13142 PCT/US99/19705
-6-
The high-resolution synthetic image generation routine 110 of one
illustrative embodiment of the invention receives the input 112 from a first
image acquisition device 206 and a second image acquisition device 208.
The first image acquisition device 206 views a scene 200 from a first
viewpoirit 216 while the second image acquisition device 208 views the
scene 200 from a second viewpoint 218. The second viewpoint 218 may
include the first viewpoint 216 (i.e., the first and second image acquisition
devices 206 and 208 may view the scene 200 from the same position).
Alternately, a displacement mechanism 232 (e.g., a mirror) positioned in
a remote location 234 may be used to make the data captured by the second
image acquisition device 208 appear as if the second image acquisition
device 208 is positioned at the remote location 234. As such, the scene
would be imaged by device 208 from the mirror 232 rather than directly.
The first image acquisition device 206 has an image resolution higher
than that of'the second image acquisition device 208. The first image
acquisition device 206 may comprise a number of different devices having
a number of different data output formats, as one skilled in the art will
readily be able to adapt the process described by the teachings herein to
any number of devices and data formats and/or protocols. In one
embodiment, the first image acquisition device 246 is a high-definition
camera, i.e., a camera with a resolution of at least 8000 by 6000 pixels/cm'.
Similarly, the second image acquisition device 208 may also comprise a
varied number of devices, since one skilled in the art can readily adapt the
routine 110 to various devices as discussed above. In one embodiment, the
second image acquisition device 206 is a camera having a resolution lower
than the resolution of the higfi-resolution device, i.e., a standard
definition
video camera. For example,'the high resolution imagery may have 8000 by
6000 pixels/cm' and the lower resolution image may have 1000 by 1000
- pixels/cm'.
The routine 110 receives input data from the first image acquisition
device 206 and corrects the spatial, intensity and chroma distortions in
step 202. The chroma distortions are caused by, for example, lens
distortion. This correction is desired in order to improve the accuracy of
subsequent steps executed in the routine 110. Methods are known in the

CA 02463162 2006-01-26
WO 00/13142 PCT/US99/19705
-7-
art for computing a parametric function that describes the lens distortion
function. For example, the parameters are recovered in step 202 using a
calibration procedure as described in H. S. Sawhney and R. Kumar, True
Multi-Image Alignment and its Application to Mosaicing and Lens
5-Distortion, Computer Vision and Pattern Recognition Conference
proceedings, pages 450-456, 1997.
Additionally, step 202 also performs chromanence (chroma) and
intensity corrections. This is necessary since image data from the second
image acquisition device 208 is merged with data from the first image
acquisition device 206, and any differences in the device response to scene
color and intensity or due to lens vignetting, for example, results in image
artifacts in the synthesized image 114. The correction is performed by pre-
calibrating the devices (i.e., the first image acquisition device 206 and the
second image acquisition device 208) such that the mapping of chroma
and intensity from one device to the next is known. The measured chroma
and intensity from each device is stored as look-up table or a parametric
function. The look up table or parametric equation are then accessed to
perform the chroma and intensity corrections in order to match the
chroma and intensity of the other device.
Input data from the second image acquisition device 208 is also
corrected for spatial, intensity and chroma distortions in step 204. The
process for correcting the low-resolution distortions in step 204 follow the
same process as the corrections performed in step 202.
To clarify, the chroma and intensity correction between the high
resolution and low resolutioii imaging devices, or between multiple same
resolution imaging devicc's, may also be performed by automatically
aligning images based on parallax or temporal optical flow computation
either in a pre-calibration step using fixed patterns or through an- online
computation as a part of the frame synthesis process. After aligning
corresponding frames using the methods described below, regions of
alignment and misalignment are labeled using a quality of alignment
metric. By using pixels between two or more cameras that have aligned
well, parametric transformations are computed that represent color and

CA 02463162 2006-01-26
WO 00/13142 PCTNS99/19705
-8-
intensity transformations between the cameras. . With the knowledge of
each parametric transformation, the source color pixels can be
transformed into the destination color pixels that completely match the
original destination pixels.
The corrected high-resolution data from step 202 is subsequently
filtered and subsampled in step 210 . The purpose of step 210 is to reduce
the resolution of the high-resolution imagery such that it matches the
resolution of the low-resolution image. Step 210 is necessary since
features that appear in the high-resolution imagery may not be present in
the low-resolution imagery, and cause errors in a depth recovery process
(step 306 detailed in Fig. 3 below). Specifically, these errors are caused
since the depth recovery process 306 attempts to determine the
correspondence between the high-resolution imagery and the low-
resolution imagery, and if features are present in one image and not the
other, then the correspondence process is inherently error-prone.
The step 210 is performed by first calculating the difference in
spatial resolution between the high-resolution and low-resolution devices.
Froin the difference in spatial resolution, a convolution kernel can be
computed that reduces the high-frequency components in the high-
resolution imagery such that the remaining frequency components- match
those components in the low-resolution imager. This can be performed
using standard, sampling theory (e.g., see P_ J. Burt and E. H. Adelson,
The Laplacian Pyramid as a Compact Image Code, IEEE Transactions on
Communication, Vol. 31, pages 532-540, 1983;.
t+'or example, if the high-resolution and low-resolution imagery
were different in spatial resolution by a factor of 2 vertically and
horizontally, then an appropriate filter kernel is [1,4,6,4,11/16. This filter
is applied first vertically, then horizontally. The high-resolution image
can then be sub-sampled by a factor of 2 so that the spatial sampling of the
image data derived from the high-resolution imager matches that of the
low-resolution imager.
Once the high-resolution image data has been filtered and
subsampled in step 210, the parallax is computed in step 212 at each frame

CA 02463162 2004-03-25
WO 00/13142 PCT/US99/19705
-9-
time to determine the relationship between viewpoint 216 and viewpoint
218 in the high-resolution and low-resolution data sets. More specifically,
the parallax computation of step 212 computes the displacement of image
pixels between the images taken from view point 216 and viewpoint 218 due
to their difference in viewpoint of the scene 200.
The pair of images can be left and right images (images from
viewpoints 216 and 218) to form a stereo pair captured at the same time
instant, or a pair of images captured at two closely spaced time intervals,
or two images at different time instants during which no substantial
independent object motion has taken place. In any of these cases the
parallax processing is accomplished using at least two images and, for
more accurate results, uses many images, e.g., five.
Because this parallax information depends on the relationship
between the at least two input images having different viewpoints (216 and
218, respectively) of a scene 200, it is initially computed at the spatial
resolution of the lower resolution image. This is accomplished by
resampling the high-resolution input image using an appropriate
filtering and sub-sampling process, as described above in step 210.
Generally speaking, the resolution of the input images may be the
same. This is a special case of the more general variable resolution case.
The parallax computation techniques are identical for both the cases once
the high resolution image has been filtered and subsampled to be
represented at the resolution of the low resolution image.
The computation of step 212 is performed using more or less
constrained : algorithms depending on the assumptions made about the
availability and accuracy of calibration information. In the uncalibrated
extreme case, a two-dimens'ional flow vector is computed for each pixel in
the to which alignment is being performed. If it is known that the
epipolar geometry is stable and accurately known, then the computation
reduces to a single value for each image point. The computation used to
produce image flow information can be constrained to produce parallax
information. The techniques described below can be applied to either the
flow information or parallax information.

CA 02463162 2008-04-02
-10-
In many situations, particularly those in which parallax
magnitudes are large, it is advantageous in step 212 to compute parallax
with respect to some local parametxic surface. This is method of
computation is known as "plane plus parallax". The plane plus parallax
representation can be used to reduce the size of per-pixel quantities that
need to be estimated. For example, in the case where scene 200 comprises
an urban scene with a lot of approacimately planar facets, parallax may be
computed in step 212 as a combination of planar layers with additional
out-of-plane component of structure. The procedure for performing the
plane plus parallax method is detailed in United State Patent Application
No. 08/493,632, fiied June 22, 1995; R. Kumar et aI., Direct Recovery of
Shape From Multiple Views: A Parallax Based Approach, 12' ICPR,
1994; Harpreet Sawhney, 3D Geometry From Planar Parallax, CVPR 94;
June 1994; and A. Shashua and N. Navab, Relative Affine Structure,
Theory and Application to 3D Construction From 2D Views, IEEE
Conference on Computer Vision and Pattern Recognition, June 1994.
Other algorithms are available that can perform paraIIax analysis
in-lieu of theplane plus parallax method. These algorithms generally use
a coarse-fine recursive estimation process using multiresolution image
pyramid representations. These algorithms begin estimation of image
displacements at reduced resolution and then refine these estimates
through repeated warping and residual displacement estimation at
successively finer resolution levels. The key advantage of these methods. is
that they provide very efficient computation even when large
displacements are present but also provide sub-pixel accuracy in
displacement. estimates. A number of published papers describe the
underlying techniques employed in the pa'rallax computation of step 212.
Details of such techniques can be found in US patent number 5,259,040,
issued November 2, 1993; J. R. Bergen et al., Hierarchical Model-Based
Motion Estimation, 2d European Conference on Computer Vision, pages
237-252, 1992; K. J. Hanna, Direct Multi-Resolution Estimation of Ego-
Motion and Structure From Motion, IEEE Workshop on Visual Motion,
pages 156-162; 1991; K J. Hanna. and Neil E. Okamoto, Combining Stereo

CA 02463162 2006-01-26
WO 00/13142 PCT/US99/19705
-11-
and Motion Analysis for Direct Estimation of Scene Structure,
International Conference on Computer Vision, pages 357-356, 1993; R.
Kumar et al., Direct Recovery of Shape from Multiple Views: A Parallax
Based Approach, ICPR, pages 685-688, 1994; and S. Ayer and J. S.
Sawhney, Layered Representation of Motion Video Using Robust
Maximum-Likelihood Estimation of Mixture Models and MDL Encoding,
International Conference on Computer Vision, pages 777-784, 1995.
Although the step 212 can be satisfied by simply computing parallax
using the plane plus parallax method described above, there are a number
of techniques that can be used to make the basic two-frame stereo parallax
computation of step 212 more robust and reliable. These techniques may
be performed singularly or in combination to improve the accuracy of step
212. The techniques are depicted in the block diagram of Fig. 3 and
comprise of augmentation routines 302, sharpening 304, routines that
compute residual parallax 306, occlusion detection 308, and motion
analysis 310.
The augmentation routines 302 make the basic two-frame stereo
parallax computation robust and reliable. One approach divides the
images into tiles and, within each tile, the parameterization is of a
dominant plane and parallax.. In particular, the dominant plane could be
a frontal plane. The planar parameterization for each tile is constrained
through a global rotation and translation (which is either known through
pre-calibration of the stereo set up or can be solved for using a direct
method).
Another augmentatibn routine 302 handles occlusions and
textureless areas that may induce errors into the parallax computation.
To process occlusions and textureless areas, depth matching across two
frames is done using varying window sizes, and from coarse to fine
spatial frequencies. A "window" is a region of the image that is being
processed to compute parallax information for a point or pixel within the
window. Multiple window sizes are used at any given resolution level to
test for consistency of depth estimate and the quality of the correlation.
Depth estimate is considered reliable only if at least two window sizes

CA 02463162 2004-03-25
WO 00/13142 PCT/US99/19705
- 12-
produce acceptable correlation levels with consistent depth estimates.
Otherwise, the depth at the level which produces unacceptable results is
not updated. If the window under consideration does not have sufficient
texture, the depth estimate is ignored and a consistent 'depth estimate
from a larger window size is preferred if available. Areas in which the
depth remains undefined are labeled as such as to that they can be filled in
either using preprocessing, i.e., data from the previous synthetic frame or
through temporal predictions using the low-resolution data, i.e., up-
sampling low-resolution data to fill in the labeled area in the synthetic
image 114.
Multiple windows are defined in terms of their sizes as well as
relative location with respect to the pixel/region for which depth/parallax
estimation is performed. Windows are defined both as centered on the
pixel for which depthlparallax is desired as well as off-centered windows.
Along with selection of windows based on a consistent depth estimate, the
selection is also accomplished on the basis of error in alignment;
specifically windows that are used to compute parallax information that
leads to a minimum alignment error and consistent depth estimates are
selected as the parallax information for the point in the image. An
illustration of the multi-window concept is shown in Fig. 7. Fig. 7 depicts
an overall image region 702 that is being processed and a plurality of
windows 700A,700B, 700C, 700D, 700E used to process the image region.
Each window 700 A-E contains the image point r104 for which the parallax
information is being generated. Window 700E is centered on the point 704,
while windows 700A-D are not centered on the point 704A (i.e., the
windows are offset fr6m the point 704), Parallax information is computed
for each window 700A-E arsd the parallax information corresponding to
the window having a minimum alignment error and consistent depth
estimates is selected as the parallax information for the image point 704.
The size and shape of the windows 700A-E are for illustrative purposes
and do not cover all the possible window conf gurations that could be used
to process the imagery. For example, windows not aligned with the
coordinate axes (vertical and horizontal) are also used. In particular,
these may be diagonal shaped windows.

CA 02463162 2004-03-25
WO 00/13142 PCTI[7S99/19705
-13-
An additional approach for employing an augmentation routine 302
is to use Just Noticeable Difference Models (JND) models in the
optimization for depth estimation. For example, typically image
measures such as intensity difference are used to quantify the error in the
depth representation. However, these measures can be supplemented
with JND measures that attempt to measure errors that are most visible to
a human observer. The approach for employing JND methods are
discussed in greater detail below.
An additional augmentation routine 302 provides an algorithm for
computing image location correspondences. First, all potential
correspondences at image locations are defined by a given camera rotation
and translation at the furthest possible range, and then correspondences
are continuously checked at point locations corresponding to successively
closer ranges. Consistency between correspondences recovered between
adjacent ranges gives a measure of the accuracy of the correspondence.
Another augmentation routine 302 avoids blank areas around the
perimeter of the synthesized image. Since the high-resolution imagery is
being warped such that it appears at a different location, the image
borders of the synthesized image may not have a correspondence in the
original synthesized image. Such areas may potentially be left blank.
This problem is solved using three approaches. The first approach is to
display only a central window of the original and high-resolution imagery,
such that the problem area is not displayed. The second approach is to use
data from previous synthesized frames to fill in the region at the
boundary. The third approach is to filter and up-sample the data from the
low-resolution device,- and insert that data at the image boundary.
An additional augmentation routine 302 provides an algorithm that
imposes -global 3D and local (multi-) plane constraints Specifically, the
approach is to represent flow between frame pairs as tiled parametric
(with soft constraints across tiles) and smooth residual flow. In addition;
even the tiles can be represented in terms of a small number of parametric
layers per tile. In the case when there is a global 3D constraint across the
two frames (stereo), then the tiles are represented as planar layers where
within a patch more than one plane may exist.

CA 02463162 2004-03-25
Y tI ' WO 00/13142 PCT/US99/1970r
-14-
Another method for improving the quality of the parallax
computation of step 212 is to employ a sharpening routine 304. For
example, in the neighborhood of range discontinuities or other rapid
transitions, there is typically a region of intermediate estimated parallax
due to the finite spatial support used in the computation process 212.
Explicit detection of such transitions and subsequent "sharpening" of the
parallax field minimize these errors. As an extension to this basic
process, information from earlier (and potentially later) portions of the
image sequence is used to improve synthesis of the high-resolution image
114. For example, image detail in occluded areas may be visible from the
high-resolution device in preceding or subsequent frames. Use of this
information requires computation of motion information from frame to
frame as well as computation of parallax. However, this additional
computation is performed as needed to correct errors rather than on a
continual basis during the processing of the entire sequence.
Additionally, the parallax computation of step 212 can be improved
by computing the residual parallax (depth) using a method described as
follows or an equivalent method that computes residual parallax 306. One
method monitors the depth consistency over time to further constrain
depth/disparity computation when a motion stereo sequence is available 'as
is the case, for example, with a hi-resolution still image. Within two
images captured at the same time instant, a rigidity constraint is valid
and is exploited in the two-frame computation of depth outlined above. For
multiple stereo frames, optical flow is computed between the
corresponding frames over time. The optical flow serves as a predictor of
depth in the new frames. Within the new frames, depth computation is
accomplished between the= pair while being constrained with soft
constraints coming from the predicted depth estimate. This can be
performed forward and backwards in time. Therefore, any areas for
which estimates are available at one time instant but not at another can be
filled in for both the time instants.
Another method of computing residual parallax 306 is to use the
optical flow constraint along with a rigidity constraint for simultaneous
depth/disparity computation over multiple stereo pairs, i.e., pairs of

CA 02463162 2004-03-25
r ti .
WO 00/13142 PCT/US99/19705
-15-
images over time. In particular, if large parts of the scene 200 are rigid,
then the temporal rigidity constraint is parameterized in the depth
computation in exactly the same manner as the rigidity constraint
between the two frames at the same time instant. When there may be
independently moving components in the scene 200, the optical flow
constraint over time may be employed as a soft constraint as a part of the
multi-time instant depth computation.
Another method of computing residual parallax 306 is to constrain
depth as consistent over time to improve alignment and maintain
consistency across the temporal sequence. For example, once depth is
recovered at one time instant, the depth at the next frame time can be
predicted by shifting the depth by the camera rotation and translation
recovered between the old and new frames. This approach can also be
extended by propagating the location of identified contours or occlusion
boundaries in time to improve parallax or flow computation.
In order to compute a consistent depth map in a given reference
frame, multiple frames over time can be used. Regions of the scene that
are occluded in one pair (with respect to the reference frame) are
generally visible in another image pair taken at some other instant of
time. Therefore, in the coordinate system of a reference frame, matching
regions from multiple frames can be used to derive a consistent
depth/parallax map.
An additional approach for computing residual parallax 306 is to
directly solve for temporally smooth stereo, rather than solve for
instantaneous depth, aind impose subsequent constraints to smooth the
result. This can be implemented using a combined epipolar and flow
constraint. For example, assuming that previous synthesized frames are
available, the condition imposed on the newly synthesized frame is that it
is consistent with the instantaneous parallax computation and that it is
smooth in time with respect to the previously generated frames. This
latter condition can be imposed by making a flow-based prediction based
on the previous frames and making the difference from that prediction
part of the error term. Similarly, if a sequence has already been
generated, then the parallax-based frame (i.e., the warped high-resolution

CA 02463162 2004-03-25
WO 00/13142 1'CT/US99/19705
-16-
image) can be compared with the flow based temporally interpolated
frame. This comparison can be used either to detect problem areas or to
refine the parallax computation. This approach can be used without
making rigidity assumptions or in conjunction with a structure/power
constraint. In this latter case, the flow-based computation can operate
with respect to the residual motion after the rigid part has been
compensated. An extension of this technique is to apply the -planar
constraint across frames along with the global rigid motion constraint
across all the files in one frame.
An additional approach is to enhance the quality of imagery using
multiple frames in order to improve parallax estimates, as well as to
produce imagery that has higher visual quality. The approach is as
follows
* perform alignment over time using a batch of frames (11 is an
example number of frames) using the optical flow approaches
described above so that images are in the same coordinate system
* sort the intensities for the batch of frames
* Perform a SELECTION process. An example is rejecting the top 2
and the lowest 2 intensities in the sorted list at each pixel.
* Perform a COMBINATION process. An example is averaging the
remaining pixels.
The result is a enhanced image. The approach can be extended so that the
approach is performed on pre-filtered images, and not on the raw
intensity images. An example of a pre-filter is an oriented band-pass
filter, for example, those described in "Two-dimensional signal and image
processing" by Jae Lim, 1990, published by Prentice-Hall, Engelwood
Cliffs, NJ.
Additionally, a method of computing residual parallax 306 which
avoids a potential problem with instability in the synthetic stereo sequence
in three dimensional structure composed using the synthetic image 114 is
to limit and amount of depth change between frames. To reduce 'this
problem, it is important to avoid temporal fluctuations in the extracted
parallax structure using temporal smoothing. A simple form of this
smoothing can be obtained by simply limiting the amount of change

CA 02463162 2004-03-25
WO 00/13142 FCT/US99119705
-17-
introduced when updating a previous estimate. To do this in a systematic
way requires inter-frame motion analysis as well as intra-frame parallax
computation to be performed.
The multi-window approach described above for the parallax
computation is also valid for flow and/or parallax computation over time.
Essentially window selection is accomplished based on criterion involving
consistency of local displacement vector (flow vector over time) and
minimum alignment error between frame pairs as in the case of two-
frame parallax/depth computation.
Occlusion detection 308 is helpful in situations in which an area of
the view to be synthesized is not visible from the position of the high-
resolution camera. In such situations, it is necessary to use a different
source for the image information in that area. Before this can be done, it
is necessary to detect that such a situation has occurred. This can be
accomplished by comparing results obtained when image correspondence
is computed bi-directionally. That is, in areas in which occlusion is not a
problem, the estimated displacements from computing right-left
correspondence and from computing left-right correspondence agree. In
areas of occlusion, they generally do not agree. This leads to a method for
detecting occluded regions. Occlusion conditions can also be predicted
from the structure of the parallax field itself. To the extent that this is
stable over time areas of likely occlusion can be flagged in the previous
frame. The'bi-directional technique can then be used to confirm the
condition.
Areas of occlusion and more generally areas of mismatch between
an original frarne and a parallax/flow-warped frame are detected using a
quality-of-alignment measure applied to the original and warped frames.
One method for generating such a measure is through normalized
correlation between the pair of frames. Areas of low variance in both the
frames are ignored since they do not affect the warped frame. Normalized
correlation is defined over a number of different image representations
some of which are: color, intensity, outputs of oriented and scaled filters.
Motion analysis 310 also improves the parallax computation of step
212. Motion analysis 310 involves analyzing frame-to-frame motion within

CA 02463162 2004-03-25
WO 00/13142 PCT/US99/19705
-1&
the captured sequence. This information can be used to solve occlusion
problems because regions not visible at one point in time may have been
visible (or may become visible) at another point in time. Additionally, the
problem of temporal instability can be reduced by requiring consistent
three-dimensional structure across several frames of the sequence.
Analysis of frame-to-frame motion generally involves parsing the
observed image change into components due to viewpoint change (i.e.,
camera motion), three dimensional structure and object motion. There is
a collection of techniques for performing this decomposition and
estimating the - respective components. These techniques include direct
camera motion estimation, motion parallax estimation, simultaneous
motion and parallax estimation, and layer extraction for representation of
moving objects or multiple depth surfaces. A key component of these
techniques is the "plane plus parallax" representation. In this approach,
parallax structure is represented as the induced motion of a plane (or
other parametric surface) plus a residual per pixel parallax map
representing the variation of induced motion due to local surface
structure. Com.putationally, the parallax estimation techniques referred
to above are essentially special cases of motion analysis techniques for the
case in which camera motion is assumed to be given by the fixed stereo
baseline.
Once the parallax field has been computed in step 212, it is used to
produce the high-resolution synthesized image 114 in a warping step 214.
The reader is encouraged to simultaneously refer to Fig. 2 and Fig. 4 for
the best understanding of the warping step 214.
Conceptually the process of warping involves two steps: parallax
interpolation and image warping. In practice these two steps are usually
combined into one operation as represented by step 214. In either case, for
each pixel in the to-be-synthesized image, the computation of step 214
involves accessing a displacement vector specifying a location in the high-
resolution source image from the first image acquisition device 206 (step
502), accessing the pixels in some neighborhood of the specified location
and computing, based on those pixels (step 504), an interpolated value for
the synthesized pixels that comprise the synthetic image 114 (step 506).

CA 02463162 2004-03-25
WO 00/13142 PCT([3S99/19105
-19-
Step 214 should be performed at the full target image resolution. Also, to
preserve the desired image quality in the synthesized image 114, the
interpolation step 506 should be done using at least a bilinear or bicubic
interpolation function. The resultant synthesized image 114 has an
apparent viewpoint 230. The apparent viewpoint 230 may be chosen by the
user to comprise all viewpoints other than the first viewpoint 216.
Even more effective warping algorithms can make use of motion,
parallax, other information (step 508). For example, the location of depth
dzscontinuities from the depth recovery process can be used to prevent
spatial interpolation in the warping across such discontinuities. Such
interpolation can cause blurring in such regions. In addition, occluded
areas can be filled in with information from previous or following frames
using flow based warping. The technique described above in the
discussion of plane plus parallax is applicable for accomplishing step 508.
Also, temporal scintillation of the synthesized imagery can be
reduced using flow information to impose temporal smoothness (step 510).
This flow information can be both between frames in the synthesize
sequence, as well as between the original and synthesized imagery.
Scintillation can also be reduced by adaptively peaking pyramid-based
appearance descriptors for synthesized regions with the corresponding
regions of the original high resolution frames. These can be smoothed
over time to reduce "texture flicker."
Temporal flicker in the synthesized frames is avoided by creating a
synthesized frame from. a window of original resolution frames rather
than from just one fraine. For example for the high resolution iniage
synthesis application, a window of, for example, five frames is selected,
Between the stereo image pair involving the current low resolution and
high resolution frames, parallax/depth based correspondences are
computed as described above. Furthermore, between the current low
resolution and previous and future high resolution frames within the
window generalized flow and parallax based correspondences are
computed (again as described above). Given the multiple correspondence
maps between the current low resolution frame and the five high
resolution frames within the window, quality of alignment maps are

CA 02463162 2004-03-25
W l R WO O0113142 PCT(US99l19705
-20-
computed for each pair of low resolution/high resolution frames.
Subsequently, a synthetic high resolution frame is synthesized by
compositing the multiple high resolution frames within the window after
warping these with their corresponding correspondence maps. The
compositing process uses weights that are directly proportional to the
quality of alignment at every pixel and the distance of the high resolution
frame in time from the current frame. Further off frames -are given
lesser weight than the closer frames.
~x'~=(P;ta)~,(t,~)õ=
I(p;t)
Y- ,Wr(P;tk)W,(tk)
lj
where w,,(p;tk) is the quality-of-alignment weight between frames t and tk
(this variable is set to zero if the quality measure is below a pre-defined
threshold); and wt(tk) is a weight that decreases as a function of time away
from frame t. Any pixels that are left unfilled by this process are filled
from the original (upsampled) frame as described above. An illustration
of the concept of temporal windows is shown in Fig, 8.
For the video enhancement application, the same method can be
applied to combine frames over time. Correspondences over time are
established using flow estimation as described above. Multiple frames are
then combined by quality weighted averaging as above.
Temporal flicker is also reduced using the constraint that regions of
error are typically consistent over time. For example, an occlusion
boundary between two. frames is typically present in subsequent frames,
albeit in a slightly different 'image location. The quality of alignment
metric can be computed as "described above and this quality metric itself
can be tracked over time in order to locate the movement of problematic
regions such as occlusion boundaries. The flow estimation method
described above can be used to track the quality metric and associated
occlusion boundaries. Once these boundaries have been aligned, the
compositing result computed above can be processed to reduce flicker. For
example the compositing result can be smoothed over time.

CA 02463162 2008-04-02
WO DO/13191 rt, auwas7#+.7.v%#
-21-
The warping step 214 can also be performed using data collected
over an image patch, rather than just a small neighborhood of pixels, For
example, the image can be split up into a number of separate regions, and
the resampling is performed based on the area covered by the region in the
target image (step 512).
The depth recovery may not produce completely precise depth
estimates at each image pixel. This can result in a difference between the
desired intensity or chroma value and the values produced from the
original bigh-resolution imagery. The warping module can then choose
to select. one or more of the following options as a depth recovery technique
(step 514), either separately, or in combination:
= leave the artifact as it is (step 616)
= insert data that has been upsampled . from the' low-resolution
imagery (step 518)
= use data'that has been previously synthesized (step,620)
= aIlow an operator to manually correct the problem (step 522).
A Just Noticeable Difference (JND) technique can be used for
selecting the appropriate combination of choices. The JND measures
performed on the synthesized sequence, and comparing the difference
between a low-resolution form of the synthesized data and data from the
low-resolution camera. Various JND measures are described in United
States Patent Application No.'s 09/055,076, filed April 3, 1989, 08/829,540,
filed March 28, 1997, 08/87,9,516, filed March 28, 1997, and 081828,161, filed
March 28, 1997 and United States Patent No.'s 5,738,430 and 5,694,491.
Additionally, the JND can be performed between the synthesized high-
resolution image data, and the previous synthesized , high-resolution
image after being warped by the flow field computed from the parallax
computation in step 212.
Depicted in Fig. 5 is a second embodiment of the routine 110. The
routine 110 receives the input, 112 from a plurality of image acquisition
devices 503 comprising the first image acquisition device 206, the second
image acquisition device 208 and a third low-resolution image acquisition

CA 02463162 2004-03-25
. .,~ ,. > ro . . WO 00/13142 PCT/US99/19'70s
-22-
device 502. Additional low resolution image acquisition devices may be
added as needed. The first, second and third image acquisition devices,
206, 208 and 502, view the scene 200 respectively from a first viewpoint 216,
a second viewpoint 218 and a third viewpoint 504. The routine 110 receives
processes the input data from the image acquisition devices, 206, 208 and
502 as discussed above with reference to steps 202, 204, 210, 212 and 214.
The additional image(s) received from the at least third image acquisition
device 502 provides data that is used in concert with the data received from
the second image acquisition device 208 during the parallax computation
step 212 and the warping step 214 to enhance the quality of the synthetic
image 114, particularly the ability to place the apparent viewpoint 230 in
locations not containing one of the image acquisition devices (i.e., the
greater number of image acquisitions devices used results in having more
lower-resolution data available to interpolate and fill in occluded or
textureless areas in the synthesized image)..
A third embodiment of the routine 110 can be understood in greater
detail by referencing Fig. 6. The routine 110, receives the input 112 from
the first image acquisition device 206 and the second image acquisition
device 208 wherein the low-resolution. image acquisition device captures
range data, for example, a laser range finder. The first image acquisition
device 206 views the scene 200 from a first viewpoint 216 while the second
image acquisition device 208 views the scene 200 from a second viewpoint
218. The routine 110 receives input data from the first image acquisition
device 206 and corrects the spatial, intensity and chroma distortions in
step 202 as discussed above.
After the high-resolution data has been corrected in step 202, the
warping step 214 creates the synthesized image 114 by using the range
(depth) data acquired from the second image acquisition device 208. The
warping step 214 again is performed as discussed above.
Although the embodiment which incorporate the teachings of the
present invention have been shown and described in detail herein, those
skilled in the art can readily devise many other varied embodiments that
still incorporate these teachings and spirit of the invention.

Dessin représentatif

Désolé, le dessin représentatif concernant le document de brevet no 2463162 est introuvable.

États administratifs

2024-08-01 : Dans le cadre de la transition vers les Brevets de nouvelle génération (BNG), la base de données sur les brevets canadiens (BDBC) contient désormais un Historique d'événement plus détaillé, qui reproduit le Journal des événements de notre nouvelle solution interne.

Veuillez noter que les événements débutant par « Inactive : » se réfèrent à des événements qui ne sont plus utilisés dans notre nouvelle solution interne.

Pour une meilleure compréhension de l'état de la demande ou brevet qui figure sur cette page, la rubrique Mise en garde , et les descriptions de Brevet , Historique d'événement , Taxes périodiques et Historique des paiements devraient être consultées.

Historique d'événement

Description Date
Représentant commun nommé 2019-10-30
Représentant commun nommé 2019-10-30
Inactive : Périmé (brevet - nouvelle loi) 2019-08-30
Requête pour le changement d'adresse ou de mode de correspondance reçue 2018-06-11
Accordé par délivrance 2009-10-27
Inactive : Page couverture publiée 2009-10-26
Préoctroi 2009-07-28
Inactive : Taxe finale reçue 2009-07-28
Lettre envoyée 2009-02-13
Un avis d'acceptation est envoyé 2009-02-13
Un avis d'acceptation est envoyé 2009-02-13
Inactive : Approuvée aux fins d'acceptation (AFA) 2009-02-10
Modification reçue - modification volontaire 2008-04-02
Inactive : Dem. de l'examinateur par.30(2) Règles 2007-10-04
Modification reçue - modification volontaire 2007-03-02
Inactive : Dem. de l'examinateur par.30(2) Règles 2006-09-05
Inactive : Lettre officielle 2006-01-26
Modification reçue - modification volontaire 2006-01-26
Lettre envoyée 2006-01-25
Lettre envoyée 2006-01-25
Inactive : Correspondance - Transfert 2005-12-08
Inactive : Lettre officielle 2005-10-26
Inactive : Transfert individuel 2005-07-27
Inactive : Dem. de l'examinateur par.30(2) Règles 2005-07-26
Inactive : Dem. de l'examinateur art.29 Règles 2005-07-26
Lettre envoyée 2004-08-26
Toutes les exigences pour l'examen - jugée conforme 2004-08-12
Exigences pour une requête d'examen - jugée conforme 2004-08-12
Requête d'examen reçue 2004-08-12
Inactive : Page couverture publiée 2004-06-08
Inactive : Lettre officielle 2004-06-02
Inactive : CIB en 1re position 2004-05-26
Lettre envoyée 2004-05-11
Exigences applicables à une demande divisionnaire - jugée conforme 2004-05-10
Demande reçue - nationale ordinaire 2004-05-07
Demande reçue - divisionnaire 2004-03-25
Demande publiée (accessible au public) 2000-03-09

Historique d'abandonnement

Il n'y a pas d'historique d'abandonnement

Taxes périodiques

Le dernier paiement a été reçu le 2009-08-04

Avis : Si le paiement en totalité n'a pas été reçu au plus tard à la date indiquée, une taxe supplémentaire peut être imposée, soit une des taxes suivantes :

  • taxe de rétablissement ;
  • taxe pour paiement en souffrance ; ou
  • taxe additionnelle pour le renversement d'une péremption réputée.

Les taxes sur les brevets sont ajustées au 1er janvier de chaque année. Les montants ci-dessus sont les montants actuels s'ils sont reçus au plus tard le 31 décembre de l'année en cours.
Veuillez vous référer à la page web des taxes sur les brevets de l'OPIC pour voir tous les montants actuels des taxes.

Titulaires au dossier

Les titulaires actuels et antérieures au dossier sont affichés en ordre alphabétique.

Titulaires actuels au dossier
IMAX CORPORATION
Titulaires antérieures au dossier
HARPREET SAWHNEY
JAMES R. BERGEN
JEFFREY LUBIN
KEITH J. HANNA
RAKESH KUMAR
SEAN M. ADKINS
Les propriétaires antérieurs qui ne figurent pas dans la liste des « Propriétaires au dossier » apparaîtront dans d'autres documents au dossier.
Documents

Pour visionner les fichiers sélectionnés, entrer le code reCAPTCHA :



Pour visualiser une image, cliquer sur un lien dans la colonne description du document (Temporairement non-disponible). Pour télécharger l'image (les images), cliquer l'une ou plusieurs cases à cocher dans la première colonne et ensuite cliquer sur le bouton "Télécharger sélection en format PDF (archive Zip)" ou le bouton "Télécharger sélection (en un fichier PDF fusionné)".

Liste des documents de brevet publiés et non publiés sur la BDBC .

Si vous avez des difficultés à accéder au contenu, veuillez communiquer avec le Centre de services à la clientèle au 1-866-997-1936, ou envoyer un courriel au Centre de service à la clientèle de l'OPIC.

({010=Tous les documents, 020=Au moment du dépôt, 030=Au moment de la mise à la disponibilité du public, 040=À la délivrance, 050=Examen, 060=Correspondance reçue, 070=Divers, 080=Correspondance envoyée, 090=Paiement})


Description du
Document 
Date
(aaaa-mm-jj) 
Nombre de pages   Taille de l'image (Ko) 
Description 2004-03-24 22 1 482
Abrégé 2004-03-24 1 7
Revendications 2004-03-24 1 30
Dessins 2006-01-25 7 161
Revendications 2006-01-25 1 26
Description 2006-01-25 22 1 410
Description 2007-03-01 22 1 403
Description 2008-04-01 22 1 400
Rappel - requête d'examen 2004-05-25 1 116
Accusé de réception de la requête d'examen 2004-08-25 1 185
Courtoisie - Certificat d'enregistrement (document(s) connexe(s)) 2006-01-24 1 104
Courtoisie - Certificat d'enregistrement (document(s) connexe(s)) 2006-01-24 1 104
Avis du commissaire - Demande jugée acceptable 2009-02-12 1 163
Correspondance 2004-05-10 1 45
Correspondance 2004-06-01 1 15
Taxes 2005-08-15 1 32
Correspondance 2005-10-25 1 26
Correspondance 2006-01-25 1 18
Taxes 2008-07-27 1 27
Correspondance 2009-07-27 2 49
Taxes 2009-08-03 1 26