Language selection

Search

Patent 2397151 Summary

Third-party information liability

Some of the information on this Web page has been provided by external sources. The Government of Canada is not responsible for the accuracy, reliability or currency of the information supplied by external sources. Users wishing to rely upon this information should consult directly with the source of the information. Content provided by external sources is not subject to official languages, privacy and accessibility requirements.

Claims and Abstract availability

Any discrepancies in the text and image of the Claims and Abstract are due to differing posting times. Text of the Claims and Abstract are posted:

  • At the time the application is open to public inspection;
  • At the time of issue of the patent (grant).
(12) Patent Application: (11) CA 2397151
(54) English Title: A METHOD AND SYSTEM FOR FORM RECOGNITION AND DIGITIZED IMAGE PROCESSING
(54) French Title: PROCEDE ET SYSTEME DE RECONNAISSANCE DE FORMULAIRE ET DE TRAITEMENT D'IMAGE NUMERISEE
Status: Dead
Bibliographic Data
(51) International Patent Classification (IPC):
  • H04N 1/00 (2006.01)
  • G06K 9/20 (2006.01)
  • G06K 9/22 (2006.01)
(72) Inventors :
  • VANSICKLE, SCOTT (United States of America)
  • LEONE, ARMAND JR. (United States of America)
  • GROSSBERG, DAVID (United States of America)
  • READ, SUSAN B. (United States of America)
  • BERGELSON, ALAN P. (United States of America)
(73) Owners :
  • VANSICKLE, SCOTT (Not Available)
  • LEONE, ARMAND JR. (Not Available)
  • GROSSBERG, DAVID (Not Available)
  • READ, SUSAN B. (Not Available)
  • BERGELSON, ALAN P. (Not Available)
(71) Applicants :
  • WORKONCE WIRELESS CORPORATION (Canada)
(74) Agent: GOWLING LAFLEUR HENDERSON LLP
(74) Associate agent:
(45) Issued:
(86) PCT Filing Date: 2001-01-09
(87) Open to Public Inspection: 2001-07-19
Availability of licence: N/A
(25) Language of filing: English

Patent Cooperation Treaty (PCT): Yes
(86) PCT Filing Number: PCT/US2001/000619
(87) International Publication Number: WO2001/052519
(85) National Entry: 2002-07-11

(30) Application Priority Data:
Application No. Country/Territory Date
09/480,914 United States of America 2000-01-11
09/692,272 United States of America 2000-10-19
09/692,289 United States of America 2000-10-19
09/692,291 United States of America 2000-10-19

Abstracts

English Abstract




A method and system for automated identification of pre-printed forms (45),
based on the location of a mark (230) made manually on the form (45). The
method comprises the steps of receiving digitized ink image data (412),
identifying the existence and location of a mark (230) and determining any
misalignment between the received ink data and a stored image of the form, and
shifting the ink data to correct the misalignment. The system comprises a
digitizer (35) and a pre-printed form (45) comprising one or more pre-printed
indicators unique to the form indicating where a user is to enter one or more
identifying marks (220, 230). The digitizer (35) transmits image data to a
computer processor (25), coupled to a database, including one or more
digitized images of marks made on a form. The processor (25) identifies the
existence and location of one or more identifying marks made by the user, and
identifies the form based on the location of said one or more marks.


French Abstract

L'invention concerne un procédé et un système d'identification automatisée de formulaires (45) pré-imprimés, à partir de l'emplacement d'une marque (230) faite manuellement sur le formulaire (45). Le procédé consiste à recevoir des données d'image d'encre numérisées (412), à identifier l'existence et l'emplacement d'une marque (230) à déterminer tout désalignement entre les données d'encre reçues et une image stockée du formulaire, et à décaler les données d'encre afin de corriger le désalignement. Le système comprend un numériseur (35) et un formulaire (45) pré-imprimé comprenant un ou plusieurs indicateurs pré-imprimés uniques au formulaire indiquant où l'utilisateur doit introduire une ou plusieurs marques (220, 230) d'identification. Le numériseur (35) transmet des données d'image à un processeur (25) informatique, couplé à une base de données, comprenant une ou plusieurs images numérisées de marques faites sur un formulaire. Le processeur (25) identifie l'existence et l'emplacement d'une ou de plusieurs marques d'identification faites par l'utilisateur, et identifie le formulaire à partir de l'emplacement de la ou des marques.

Claims

Note: Claims are shown in the official language in which they were submitted.



CLAIMS

What is claimed is:

1. A method of automated identification of pre-printed forms, comprising the
steps of:
(a) receiving and storing data describing a location of a mark made manually
on a pre-printed form; and
(b) identifying said form based on the location of said mark.

2. The method of claim 1 wherein the data received and stored in step (a) is
digitized
image data that is received by a computer and stored in a computer storage
medium.

3. The method of claim 2 wherein step (b) comprises determining a form
identification
zone that includes said location of said mark, and determining the form that
corresponds to the form identification zone.

4. The method of claim 3 wherein step (b) comprises storing a table that
matches form
identification zones with form IDs.

5. The method of claim 4 wherein step (b) further comprises:
reformatting received digitized image data into a format to be processed by a
recognition engine;
translating said reformatted digitized image data into recognized data by
processing said reformatted image data with said recognition engine;
retrieving recognized data for one or more form identification zones;
searching said retrieved recognized data for the zone containing the mark; and
identifying the form by matching the zone to a form ID.

6. The method of claim 5 further comprising the step of displaying digitized
image data
on a computer monitor to enable a user to manually identify the form on which
the
image was written.

-20-


7. The method of claim 6 further comprising the step of displaying recognized
data on the
user's computer monitor in a format convenient for auditing.

8. The method of claim 7 further comprising the step of displaying digitized
image data
superimposed on an image of the identified form so as to present an on-screen
image of
the form and the images written thereon.

9. The method of claim 8 further comprising the step of auditing said
recognized data by
comparing said recognized data to said displayed digitized image data
superimposed
on an image of the identified form.

10. A system for identification of pre-printed forms, comprising
a) a digitizer;
b) a pre-printed form comprising one or more pre-printed indicators unique to
the form indicating where a user is to enter one or more identifying marks;
and
c) a computer processor, coupled with a database, that
i) receives one or more digitized images of marks made on the form,
ii) identifies the existence and location of one or more identifying marks
made by the user, and
iii) identifies the form based on the location of said one or more marks.

11. Computer software for automated identification of pre-printed forms,
comprising:
(a) software for receiving and storing data describing a location of a mark
made
manually on a pre-printed form; and
(b) software for identifying said form based on the location of said mark.

12. The software of claim 11 further comprising software for storing a table
that matches
form identification zones with form IDs.

13. The software of claim 11 wherein the received and stored data is digitized
image data
that is stored in a computer storage medium.

-21-


14. The software of claim 13 further comprising software for management of
said digitized
image data for subsequent reference.

15. The software of claim 14 wherein the software for identifying comprises
software for
determining a form identification zone that includes said location of said
mark, and
software for determining the form that corresponds to the form identification
zone.

16. The software of claim 15 wherein the software for identifying further
comprises
software for:
reformatting received digitized image data into a format to be processed by a
recognition engine;
translating said reformatted digitized image data into recognized data by
processing said reformatted image data with said recognition engine;
retrieving recognized data for one or more form identification zones;
searching said retrieved recognized data for a zone containing the mark; and
identifying the form by matching the zone to a form ID.

17. The software of claim 16 further comprising software for displaying
digitized image
data on a computer monitor to enable a user to manually identify the form on
which the
image was written.

18. The software of claim 17 further comprising software for displaying
recognized data
on the user's computer monitor in a format convenient for auditing.

19. The software of claim 18 further comprising software for displaying
digitized image
data superimposed on an image of the identified form so as to present an on-
screen
image of the form and the images written thereon.

-22-


20. The software of claim 19 further comprising software for auditing said
recognized data
by comparing said recognized data to said displayed digitized image data
superimposed
on an image of the identified form.

21. The software of claim 20 further comprising software for recognizing
pertinent
information on a form based on non-standard symbols written on the form.

22. A method of processing digitized image data, comprising the steps of:
(a) receiving digitized image data;
(b) determining a location of a first mark comprised in said digitized image
data;
(c) determining a distance and direction from the location of said first mark
to a
reference location; and
(d) performing a shift of said digitized image data, wherein said shift is
based
on said distance and direction.

23. A method as in claim 22, wherein said digitized image data corresponds to
marks made
on a form placed on a digitizer pad.

24. A method as in claim 23, wherein said form is identified by a location of
a zone in
which an identifying mark has been made.

25. A method as in claim 24, wherein said first mark is said identifying mark.

26. A method as in claim 25, wherein said identifying mark is approximately
circular.

27. A method as in claim 25, wherein said location of said first mark is
determined by
locating a center of said first mark.

28. A method as in claim 26, wherein said location of said first mark is
determined by
locating a center of said first mark.

-23-


29. A method of processing digitized image data, comprising the steps of:
(a) receiving digitized image data;
(b) determining a location of a first mark comprised in said digitized image
data:
(c) determining a distance and direction from the location of said first mark
to a
reference location;
(d) performing a shift of said digitized image data, wherein said shift is
based on said
distance and direction;
(e) translating said shifted digitized image data into recognized image data
by
processing said digitized image data with a recognition engine; and
(f) merging said recognized image data with a reference image.

30. A method as in claim 29, wherein said digitized image data corresponds to
marks made
on a form placed on a digitizer pad.

31. A method as in claim 30, wherein said reference image is an image of said
form.

32. A method as in claim 31, wherein said form is identified by a location of
a zone in
which an identifying mark has been made.

33. A method as in claim 32, wherein said first mark is said identifying mark.

34. A method as in claim 33, wherein said identifying mark is approximately
circular.

35. A method as in claim 33, wherein said location of said first mark is
determined by
locating a center of said first mark.

36. A method as in claim 34, wherein said location of said first mark is
determined by
locating a center of said first mark.

37. A method of automated identification of pre-printed forms, comprising the
steps of:

-24-




(a) receiving and storing data describing a location of a mark made manually
on
a pre-printed form, wherein said mark is made by filling in a hole in a
template affixed to a
digitizer pad; and
(b) identifying said form based on the location of said mark.

38. The method of claim 37 wherein the data received and stored in step (a) is
digitized
image data that is received by a computer and stored in a computer storage
medium.

39. The method of claim 38 wherein step (b) comprises determining a form
identification
zone that includes said location of said mark, and determining the form that
corresponds to the form identification zone.

40. The method of claim 39 wherein step (b) comprises storing a table that
matches form
identification zones with form IDs.

41. The method of claim 40 wherein step (b) further comprises:
reformatting received digitized image data into a format to be processed by a
recognition engine;
translating said reformatted digitized image data into recognized data by
processing said reformatted image data with said recognition engine;
retrieving recognized data for one or more form identification zones;
searching said retrieved recognized data for the zone containing the mark; and
identifying the form by matching the zone to a form ID.

42. The method of claim 41 further comprising the step of displaying digitized
image data
on a computer monitor to enable a user to manually identify the form on which
the
image was written.

43. The method of claim 42 further comprising the step of displaying
recognized data on
the user's computer monitor in a format convenient for auditing.



-25-




44. The method of claim 43 further comprising the step of displaying digitized
image data
superimposed on an image of the identified form so as to present an on-screen
image of
the form and the images written thereon.

45. The method of claim 44 further comprising the step of auditing said
recognized data by
comparing said recognized data to said displayed digitized image data
superimposed
on an image of the identified form.

46. A system for identification of pre-printed forms, comprising
a) a digitizer;
b) a pre-printed form;
c) a template affixed to said digitizer; wherein said template comprises one
or
more holes for a user is to enter one or more identifying marks to identify
said form; and
c) a computer processor, coupled with a database, that
i) receives one or more digitized images of marks made on the form,
ii) identifies the existence and location of one or more identifying marks
made by the user, and
iii) identifies the form based on the location of said one or more marks.

47. Computer software for automated identification of pre-printed forms,
comprising:
(a) software for receiving and storing data describing a location of a mark
made
manually on a pre-printed form by filling in one or more holes in a template;
and
(b) software for identifying said form based on the location of said mark.

48. The software of claim 47 further comprising software for storing a table
that matches
form identification zones with form IDs.

49. The software of claim 47 wherein the received and stored data is digitized
image data
that is stored in a computer storage medium.



-26-




50. The software of claim 49 further comprising software for management of
said digitized
image data for subsequent reference.

51. The software of claim 50 wherein the software for identifying comprises
software for
determining a form identification zone that includes said location of said
mark, and
software for determining the form that corresponds to the form identification
zone.

52. The software of claim 51 wherein the software for identifying further
comprises
software for:
reformatting received digitized image data into a format to be processed by a
recognition engine;
translating said reformatted digitized image data into recognized data by
processing said reformatted image data with said recognition engine;
retrieving recognized data for one or more form identification zones;
searching said retrieved recognized data for a zone containing the mark; and
identifying the form by matching the zone to a form ID.

53. The software of claim 52 further comprising software for displaying
digitized image
data on a computer monitor to enable a user to manually identify the form on
which the
image was written.

54. The software of claim 53 further comprising software for displaying
recognized data
on the user's computer monitor in a format convenient for auditing.

55. The software of claim 54 further comprising software for displaying
digitized image
data superimposed on an image of the identified form so as to present an on-
screen
image of the form and the images written thereon.

56. The software of claim 55 further comprising software for auditing said
recognized data
by comparing said recognized data to said displayed digitized image data
superimposed
on an image of the identified form.



-27-




57. The software of claim 56 further comprising software for recognizing
pertinent
information on a form based on non-standard symbols written on the form.



-28-

Description

Note: Descriptions are shown in the official language in which they were submitted.



CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
A METHOD AND SYSTEM FOR FORM RECOGNITION
AND DIGITIZED IMAGE PROCESSING
FIELD OF THE INVENTION
The present invention relates generally to data entry using a digitizer device
and, more
particularly, to a system and method for identifying forms placed on a
digitizer and for
determining and correcting digital image offset for digitizer pads.
BACKGROUND
Systems are known that automatically identify forms placed on a digitizer
device.
(The term "form," as used herein. refers to a printed image on a page, rather
than to an actual
page; that is, if two pages are printed with the same image, they are
considered the same
form.) For example, U.S. Pat. INTO. x,869,789 (Reid-Green) discusses a page
recognition
system that detects a page number pre-coded onto a sheet of paper when the
sheet of paper is
mounted on the digitizer system. The system utilizes a built-in scanner to
detect a pre-coded
page identifier on the back-side of the page. U.5. Pat. No. 5,555,101 (Larson
et al.) discusses
an automated page recognition system for forms received by facsimile and also
uses a pre-
printed pattern to identify a form.
U.S. Pat. No. 5,629,499 (Flickinger et al.) discloses an electronic clipboard
and
discusses two methods for recognizing forms used thereon. In the first method,
a toggle
switch is used to toggle between various form identifiers displayed on the
clipboard; in the
second method, a bar-code reader reads a bar-code pre-printed on the form. The
toggle-
display requires the clipboard to contain a micro-processor that may need to
be re-
programmed every time a new set of forms is used.
Digitizing notepads are known that record notations made on paper forms
mounted
thereon. For example, A.T. Cross Company has marketed a CrossPad 3~ (see FIG.
1), a
portable digital notepad that digitizes and records pen movements using radio
transmission.
The CrossPad transmits the recorded pen movements to a personal computer for
processing
and analysis. The CrossPad, however, is not itself capable of identifying pre-
printed forms.
The information recorded by the CrossPad and transmitted to a computer only
comprises the
pen strokes entered by a user writing on a form. Information pre-printed on
the form cannot


CA 02397151 2002-07-11
WO 01/52519 PCT/USOi/00619
be detected by the Crosspad. The CrossPad does not have a form ID display that
can be
toggled through by a user, nor does it have a bar-code reader.
Because devices such as the CrossPad only detect information that is written
by a
user, such devices generally cannot be used in systems that identify a form
based on pre-
y printed information. Even systems such as those disclosed in Reid-Green are
not capable of
detecting form identification information when multiple forms are placed on
the digitizer.
For example, Reid-Green would only identify the form that is at the bottom of
the stack. The
device discussed in Flickinger et al. might be capable of reading multiple
pages, but such a
device requires additional (and costly) form-reading mechanisms not present in
CrossPad-
type devices. The use of a bar-code reader or other form-reading mechanism in
a digitizer
increases the cost and the weight of the digitizer.
Another drawback of CrossPad-type devices is that the spatial relationship of
written
marks on the paper form is not identical to the spatial relationship of the
corresponding digital
markings on the underlying "digital form." This offset ("digital drift") can
be as much as
several millimeters.
A further drawback of CrossPad-type devices arises from the fact that when a
form is
used, the pre-printed information on the form is not part of the digital ink
file that is
uploaded. In order for the digital ink file to be optimally useful after
upload, an image of the
contextual information present on the paper form at the time of use needs to
be added to the
digital ink image. The two images must be correctly aligned. so that an image
of the original
form, with the information written thereon, can be re-created.
There is therefore a need for a system of automated form identification that
is capable
of recognizing forms used on the CrossPad, and similar devices, without
relying on pre-
printed information on the forms.
2~ There is a further need for a system that does not require a cumbersome LCD
display
(along with a programmable microprocessor) and toggle switch or other input
device to be
integrated into a CrossPad-type of electronic clipboard.
There is a further need for a method and system for correcting digital drift
for
CrossPad-type devices and other digitizers. There is a further need for a
method of
combining and aligning a digital ink image of handwritten information on a
form with an
image of the underlying form.
-2-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
SUMMARY
The present invention comprises a method for automated identification of pre-
printed
forms, comprising the steps of identifying the existence and location of a
mark made
manually on a pre-printed form and identifying the pre-printed form by
comparing the
location of the mark to a database listing forms and corresponding locations
of identifying
marks. The invention further comprises a system for identification of pre-
printed forms,
comprising a digitizer; a pre-printed form comprising one or more pre-printed
indicators
unique to the form indicating where a user is to enter one or more identifying
marks; and a
computer processor, coupled with a database, that receives one or more
digitized images of
marks made on the form, identifies the existence and location of one or more
identifying
marks made by the user, and identifies the form by comparing the location of
said one or
more marks to a database listing forms and corresponding locations of
identifying marks.
The present invention also comprises computer software for automated
identification of pre-
printed fortes, comprising software for receiving and storing data describing
the location of a
mark made manually on a pre-printed form and software for identifying said
form by
comparing the location of said mark to a database listing forms and
corresponding locations
of identifying marks.
The present invention further comprises a method for correcting digital drift
and for
mapping a digital ink file onto a digital form. Typically, a digitizing pad
(e.g., a CrossPad or
similar device) suffers from a certain amount of digital drift -- the spatial
difference between
the relative location of a writing on a paper form and the relative position
of the
corresponding digital mark on the underlying digital form. The amount and
direction of
digital drift is constant for each pad (and anywhere on a given form), but
typically varies from
pad to pad.
Software of a preferred embodiment determines the amount and direction of
digital
drift by reference to a known landmark on a given form. Preferably. such a
landmark is a
form-identification mark, discussed below. In a preferred embodiment, the form
identifier is
a circle. The digital drift is determined by comparing the known position of
the center of the
form identifier circle on the paper form with the center of the corresponding
digital mark on
the digital form underneath, then determining variances in horizontal and
vertical directions.
- J


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
Once the digital drift is determined, it is applied as a correction factor to
all of the digital
markings on the form.
This correction factor is preferably used to facilitate the alignment of a
handwritten
digital ink image with an image of the underlying form. This alignment is
accomplished by
first identifying the form that digital ink data was written on. A pre-loaded
image of the
identified form is then retrieved from a database. The correction factor is
used to align the
ink data with the digital form. The correct alignment of the digital form with
the digital ink
enables replication of the paper form as it appeared when it was filled out.
The correct
alignment further enables the composite image to be processed by optical mark
recognition
(OMR), optical character recognition (OCR), and image snippets storage into a
database.
The present invention further comprises a method and system for creating
enabling
identification of forms that have not had a form-identification mark pre-
printed thereon. In a
preferred embodiment of this aspect of the invention, a metal or plastic guide
with three-
millimeter holes positioned at various selected locations in the guide is
affixed to one side of
a digitizer pad. Preferably, the guide is affixed at a pre-determined location
linked to a
customer. The fixing of the guide to the side of the pad in a pre-determined
location enables
a user to make a mark in a hole, wherein the location of the hole corresponds
to the location
that a pre-printed mark would have. That is, the mark has the same location
that a mark that
filled in a pre-printed form identification circle would have. This allows
clients who have
preprinted forms that were not printed with form identification marks to use
those forms in
the remainder of the disclosed system.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 displays components of a preferred embodiment of a preferred digitizer
system.
FIGS. 2 and 3 depict data collection forms used in a preferred embodiment of
the
present invention.
FIG. 4 is a flow diagram showing the operation of software for processing
digitized
data.
_4_


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
FIGS. 5 and 6 depict a flow diagram showing the operation of form recognition
software.
FIG. 7 illustrates placement of zones used in a preferred embodiment of the
present
invention.
FIG. 8 depicts a graphical user interface used in a preferred embodiment of
the
present invention.
FIG. 9 depicts steps of a main form alignment process used in a preferred
embodiment.
FIG. 10 illustrates coordinate definitions used in a preferred embodiment.
FIG. 11 depicts a locate-ink-bubble-center process of a preferred embodiment.
FIG. 12 depicts a recognize-aligned-image process of a preferred embodiment.
FIG. 13 depicts a merge-aligned-ink-with-reference-image process of a
preferred
embodiment.
FIG. 14 depicts a form identification process of a preferred embodiment.
FIG. 15 depicts a form identifier template of a preferred embodiment.
FIG. 16 depicts a digitizer pad with three side paper guides.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
In a preferred embodiment, the system and method are used in conjunction with
a
digitizer system, shown in FIG. 1, comprising a digitizer 35, such as a
CrossPad, pre-printed
forms 4~, a personal computer 2~, and accompanying software. For ease of
reference, the
term "CrossPad" is used herein to refer to any digitizer pad to which the
described invention
is applicable. Those skilled in the art will recognize that the system and
method are
applicable to any digitizer system in which a form can be placed in a
digitizer field. Further,
although the term "form" is used in this description to refer to forms printed
on paper, those
skilled in the art will recognize that the disclosed system and method is
equally applicable to
forms embodied in other media. For example, the disclosed invention also works
for forms
on plastic laminates or "electronic paper" such as the "Gyricon" product
developed by Xerox
Corporation of Palo Alto, California, or a similar product developed by E Ink
Corporation of
Cambridge, Massachusetts.
-5-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
The general operation of the digitizer system shown in FIG. 1 is as follows. A
pre-
printed data collection form 45 is placed on a CrossPad digitizer 35. A user
places marks on
the form 45 using the special pen that is required by the CrossPad 35. The
marks made on the
form 45 are recorded as digitized images by the CrossPad 35. The user connects
the
CrossPad 35 to a computer 25 and the CrossPad 35 transmits the images it has
stored to the
computer 25. In an alternate embodiment, the CrossPad is connected to a
converter that
converts the Crosspad data into an Internet-compatible format, and the
converted CrossPad
data is transmitted by the converter to a remotely located computer via the
Internet. In either
embodiment, the stored Crosspad images on the computer 25 are then processed
by software
stored on a computer. The forms and software used in a preferred embodiment
are described
below.
FIG. 2 depicts a first data collection form 45 used in a preferred embodiment.
An
area 210 at the left margin of the form is reserved for form identification. A
"Start Here"
instruction 225 reminds a user that an indicator 230 for a form identifier is
to be filled in (the
circle is the indicator; the mark made by filling in the circle is the form
identifier). For each
different form, the form identifier indicator 230 is located in a different
zone in area 210.
The forms used in a preferred embodiment are printed so that they may be used
in one
or more data-entry systems. For example, the cross-hair images 215 at the
lower right of the
form and 220 at the upper left of the form are used to align the image of the
form if it is
scanned in by an optical scanner.
FIG. 3 depicts a second data collection form used in a preferred embodiment.
The
form in FIG. 3 is shown without reference numbers in order to more clearly
show the
appearance of forms used in a preferred embodiment. Note that the form
identifier indicator
on the form in FIG. 3 is located in a different zone of area 210 than the form
identifier
indicator 230 on the form shown in FIG. 2.
Preferably a user fills in a form identifier indicator each time the user
begins to write
on a form. However, to reduce errors, a first user may prefer to fill in a
plurality of indicators
before handing a CrossPad over to a second user. For example, a first user may
use a
template comprising indicators for all forms. If forms 1, 3, and 6 are to be
used, the first user
places the template on the CrossPad and fills in the indicator for form 1,
hits the Next Page
button on the CrossPad, fills in the indicator for form 3, hits the Next Page
indicator again,
-6-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
and fills in the indicator for form 6. The first user then hits the Back Page
button twice, and
passes the CrossPad, with forms l, 3, and 6 clipped thereon, to a second user.
The second
user can then fill in forms l, 3, and 6 without needing to fill in the
appropriate form identifier
indicators (although the Next and Back page buttons will still need to be
pressed when the
second user goes from form to form).
FIG. 4 is a flow diagram showing the operation of software used in a preferred
embodiment. When the form identifier indicator 230 is filled in by a user of
the Crosspad 35,
the image and location of the form identifier, along with image and location
data for any other
information written on the form by a user, is stored in the CrossPad memory.
The CrossPad
stores this information as ''ink data," which is an IBM proprietary data
format, though any
suitable format for storing this data may be used. This data is referred to
generally herein as
ink data.
At step 410 the user uploads the collected ink data to a personal computer by
plugging the CrossPad into a serial connection cable (although infrared,
ethernet, or other
connection devices could also be used) attached to the personal computer and
then pressing a
button on the Crosspad. In an alternate embodiment, the ink data is
transferred first to a
converter, which then sends the data to a remotely located computer,
preferably over the
Internet. The remotely located computer can be a personal computer, but is
preferably a
server-class computer (e.g., an Internet server), since it is expected that
the computer will be
receiving data from a plurality of sources.
At step 412 an ink transfer program stored on the personal computer receives
the
transferred ink data and at step 415 saves the data to the computer's hard
drive as an ink data
file. In a preferred embodiment, an IBM Ink Manager Transfer program, supplied
with the
Crosspad, performs this step and creates a file known as a "notebook."
After the ink data file is uploaded, the ink transfer program launches an ink
converter
program (referred to herein as TMDInkExec) at step 417. At step 420 TMDInkExec
reads
the uploaded ink data file and converts each ink page into a non-proprietary
image file (such
as group 4 TIFF, but other formats could also be used) by calling ink
conversion modules
included in an IBM-provided library, IBM Electric Ink SDK. TMDInkExec calls
the
appropriate ink conversion modules according to instructions contained in
documentation
included with IBM Electric Ink SDK. In an alternate embodiment, the conversion
at step 420


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
from notebook format into non-proprietary image format can also occur at a
remotely located
computer: the ink data is transferred to the remote computer using, e.g.,
email or file transfer
protocol, then the remote computer (or computers) poll multiple receiving
queues (e.g.,
email-boxes) and perform step 420, processing the received batches of ink
data.
Upon completion of step 420, the TMDInkExec program at step 422 notifies a
second
program, TMDInkMonitor, to further process the TIFF data. The TMDInkMonitor
program,
at step 424, displays a graphic interface for the user to optionally assign or
confirm a subject
ID to the batch of pages in the ink data file (for example, if the pages are
forms that have been
filled in for a particular medical patient, the patient's ID would be
entered). If a scanner is
used instead of a CrossPad, a subject ID label may be affixed to the form,
recognized. and the
recognized number assigned by the system. If a subject ID is not assigned by
the user (and
has not been recorded by the CrossPad), the TMDInkMonitor program assigns a
default
subject ID (preferably the previous subject ID incremented by 1 ).
A batch upload ID is assigned at step 426 to the corresponding batch of TIFF
image
files, and at step 428 the TIFF data is written into a database Lender both
the batch upload ID
and the subject ID. The TIFF image files are then copied out of the database
and processed at
step 430 by a recognition engine that turns the images into recognized data by
applying OMR
(optical mark read, or optical mark recognition) technology. A recognition
engine is software
that provides a documented set of services to another software module or
program related to
the transformation (recognition) of image data into computer readable text. A
Caere
Developer's Kit 2000 Recognition Engine, sold by Caere Corporation of Los
Gatos,
California, is used in a preferred embodiment, but numerous recognition
engines are
commercially available (e.g., from Mitek Systems, Inc., of San Diego,
California, and
National Computer Systems, Inc., of Minneapolis, Minnesota) and could also be
used in a
preferred embodiment by those skilled in the art.
At step 432 TMDInkMonitor calls a Form ID determination module that in turn
derives the form ID corresponding to the recognized data. The steps performed
by the Form
ID determination module are illustrated in FIGS. 5 and 6.
The form ID determination module uses an OMR Recognition Module and an OMR
filling method. A recognition module is software that conforms to the
specifications of a
recognition engine and that transforms different types of image data into
computer readable
_g_


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
text. A filling method specifies the type of image data that is expected to
appear inside a
zone. Some filling methods are associated with a unique recognition module;
others are
supported by more than one. Some examples of filling methods are: multiple
font machine
print, OMR, hand-print, bar-code. dot-matrix printing, OCRA, and Braille.
A zone is an area in an image, up to full page size, containing a feature of
interest to
the user. In a preferred embodiment, the zones are rectangular, but those
skilled in the art
will recognize that differently-shaped zones can also be used. The image data
covered by the
zones are handled and processed separately. General zone types are graphical
(no
recognition). textual, and others such as bar code or OMR. To facilitate
recognition of the
image data in the zone areas, each zone is assigned properties based on the
expected
information that it will contain. The basic zone properties are x-y
coordinates, size,
recognition module, filter, language dictionary, and user dictionary. A
language dictionary is
a file that contains a language's lexical elements plus rules for generating
words from them.
The file can be used to mark non-dictionary words in recognized text, or to
change non-
dictionary words to acceptable words. Some recognition modules consult the
language
dictionary to help them make recognition decisions. A user dictionary contains
lists of words,
strings, and regular expressions defining character patterns, supplementing
the content of the
language dictionary file. One or more user dictionaries can be associated with
a particular
zone. A regular expression in a user dictionary defines a permissible pattern
of characters by
specifying character classes, ranges or individual characters that are valid
for one or more
character positions in a zone to improve recognition accuracy. A regular
expression is also
called a mask or a pattern. Regular expressions are usually used in form-like
situations,
where zone content is in some way predictable.
A Form Identification Zone Template, or FIZT) is used in identifying a form
from an
image. A zone template is a set of zone and page properties (corresponding,
for example, to a
form to be recognized) that is specified and stored, for later retrieval and
processing, in a
medium such as a disk file or database record. The FIZT consists of a set of
zones which are
laid out in a grid pattern (see FIG. 7). Filters can be applied at either the
zone level or the
form level to improve recognition results. Filters are specifications for the
elimination of
certain categories of characters from being considered as possible valid
results. Examples of
filters include: digits only, uppercase letters, lowercase letters, and
punctuation.
-9-


CA 02397151 2002-07-11
WO 01/52519 PCT/USOI/00619
Because the subject invention utilizes OMR technology, in a further preferred
embodiment, non-standard symbols (such as an encircled N - analogous to the
symbol ~) are
used to indicate pertinent information on a form. Examples of such pertinent
information
comprise the following: cut-and-paste areas; information that needs to be
corrected; and a
reference to another data file.
The zones for form recognition are ordered left to right and top to bottom as
shown by
the shaded numbers in FIG. 7. Each zone corresponds to a form number in the
set to which
the form belongs. The zones are laid out with enough space between them to
minimize the
possibility of a false form identification caused by the ~~riter marking
outside of the circle or
caused by skewing when a form is scanned in. Each form in a set is assigned a
zone that
corresponds to its form number, and an unfilled dot is pre-printed within the
assigned zone on
the form page (see FIGS. 2 and 3). Each zone, and thus each dot, is assigned a
number. A
table ("dot map'') maps the dot number to a form number in the form set. An
example table
for dot mapping is as follows:
Dot Number Form ID Number


1 00300001-O1


2 00300001-02


OO3OOOO1-O3


The "dot map" is stored for later retrieval in a storage medium such as a
database table or a
disk file.
Referring to FIG. 5, at step 510 the form ID determination module retrieves
recognized data from mark sense zones. At step 520 a search is performed for
the first filled-
2~ in dot and corresponding zone, and at step 530 its dot number is determined
from the zone
definitions. At step 540, the corresponding form ID number is retrieved from a
dot mapping
table based on the dot number (for example, in the above table, dot number '1'
corresponds to
form ID number '00300001-O1'). If at step 545 the form ID number has not been
determined,
at step 550 an ink-data-based image of the page is displayed to a user and at
step 555 the user
is prompted to enter the proper form ID number. At step 560, the user enters
the form ID
number. Although a preferred embodiment uses numbers to identify forms, those
skilled in
the art will recognize that alphanumeric or other sy~rnbols could also be
used. The process
shown in FIG. 5 is shown in somewhat more detail in FIG. 14.
-10-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
Referring to FIG. 6, after the form ID number has either been determined by
the
system or entered by a user, step 64~ is performed. At step 64~ the data is
further recognized
(applying, for example, optical character recognition, or handprint
recognition-also known as
ICR, for Intelligent Character Recognition; a preferred method of recognition
is depicted in
FIG. 12), based on the form ID number, and at step 6~0 the recognized data and
the ID
number of the form on which the images were written are stored as an unaudited
file in the
database. The recognized data is organized according to the categories of the
form and
optionally displayed on a user's audit screen. The data is displayed in a
format, such as a
table, that is convenient for auditing. At step 660 the user optionally
audits, verifies, and
corrects, if appropriate, each recognized data element on the audit screen. In
a preferred
embodiment, a graphical user interface (GUI) 810 (see FIG. 8) enables a user
to view the data
in a "contextual orientation": if, for example, some data is not recognized, a
user can have an
image of a portion of the paper form displayed on the screen, with an image of
the ink data
for that page overlaid onto the image of the paper form so as to present to
the user an on-
1 ~ screen image of a portion of the written-on paper form. This enables the
user to audit and
verify the data elements without having the written-on form present, and to
view all of the
images written on the paper form. This in turn enables an operator to perform
auditing and
verification at a site remotely located from the site where the paper forms
have been written
on. Further, when the data are, for example, medical reports, contextually-
oriented on-screen
viewing of the handwriting on the form enables a user to reconstruct the mood
of the writer
wher: the report was taken, as well as the personality of the patient.
At step 665 the audited data is saved to the database.
FIG. 9 depicts an overview of a preferred form alignment process. At step 910,
the
form identification process described in FIGS. ~ and 6, with associated text,
is performed. At
step 915, if the form identification step 910 was not successful, at step 920
the form
alignment process terminates, with the form marked as "unknown."
If the form identification step 910 was successful, then at step 925, a Locate
Ink
Bubble Center Process (see FIGS. 10 & 11 and associated text) is performed on
an ink image
bubble. This step results in the calculation of the coordinates (XInkCenter,
YInkCenter) of
the ink image bubble.
-11-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
Following step 925, the offset (XOffset, YOffset) between a reference bubble
and the
ink image bubble is calculated: XOffset = XInkCenter - XMast; YOffset =
YInkCenter -
YMast. XMast and YMast are denoted in FIG. 10 by XO and Y0, respectively.
At step 940, ink alignment is performed: the XOffset and YOffset values are
applied
to the ink image, and the image is shifted accordingly.
At step 9~0, the aligned ink image is recognized (see FIG. 12 and associated
text).
Finally, at step 960, the recognized, aligned ink image is merged with the
reference image
(the stored image of the identified form.)
FIG. 10 illustrates the coordinate and zone definitions used in a preferred
Locate Ink
Bubble Center process, and FIG. 11 depicts the steps of that process.
Referring to FIG. 10, a coordinate system as depicted, with the y-coordinate
increasing in the downward direction. A reference search zone 1020 is
determined by the
identified form and a Form Identification Zone Template. Because the form was
identified, a
bubble ink circle 1030, with center at (XInk,YInk), was been found within a
particular
reference bubble zone 1010. Reference bubble zone 1010 is such that a
reference bubble has
center (X0, YO), which is also the center of reference bubble ink zone 1010.
Reference
search zone 1020 is defined by extending bubble ink zone 1010 a distance
Xexpand
horizontally in each direction (left and right) and a distance Yexpand
vertically in each
direction (up and down).
Thus, reference search zone 1020 is a rectangle whose corners have coordinates
(X1,
Y 1 ) (upper left corner), (X2, Y 1 ) (upper right corner), (X2, Y2) (lower
right corner), and (X 1,
Y2) (lower left corner).
A Locate Ink Bubble Center process is depicted in FIG. 11. At step 1105,
variables
are initialized. BubbleZoneLeft is the x-coordinate of the left side of bubble
zone 1010.
BubbleZoneRight is the x-coordinate of the right side of bubble zone 1010.
BubbleZoneTop
is the y-coordinate of the top side of bubble zone 1010. BubbleZoneBottom is
the y-
coordinate of the bottom side of bubble zone 1010. A variable YScan is set to
Yl .
The variables YTop, YBottom, XLeft, and XRight are initially set to Y2, Y1,
X2, and
Xl, respectively. At the end of the Locate Ink Bubble Center process, the
variable YTop will
contain the value of the y-coordinate of the top of bubble ink 1030, YBottom
will contain the
value of the y-coordinate of the bottom of bubble ink 1030, XLeft will contain
the value of
-12-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
the x-coordinate of the left side of bubble ink 1030,, and XRight will contain
the value of the
x-coordinate of the right side of bubble ink 1030. Note that the process is
robust, in that even
if bubble ink 1030 is not a perfect disk, the process will determine a
''center'' for bubble ink
1030.
At step 1110, a variable XScan is set to X1. At step 1115, a check is made
whether a
black pixel is located at (XScan, YScan) (which initially is at (X1, Y1), the
upper-left corner
of reference search zone 1020. If the answer is no, then at step 1160 the
variable XScan is
incremented. If when the value is checked at step 1165 the incrementation does
not result in
a value greater than X2 (i.e., if the scan has not reached the right side of
reference search zone
1020), step 1115 is repeated. If at step 1165 the incrementation at step 1160
of the variable
XScan does result in XScan having a value greater than X2, then at step 1170
the variable
YScan is incremented. If when the value of YScan is checked at step 1175 the
value of
YScan is greater than Y2 , then at step 1180 XInkCenter and YInKCenter are
calculated. If at
step 1175 the value of YScan is not greater than Y2, then step 1110 is
repeated. The
software scans from left to right, top to bottom over reference search zone
1020.
If the answer at step 1115 is yes, then at step 1120 a check is made as to
whether the
current value of the variable YScan is less than the current value of YTop. If
yes, then at step
1125 YTop is set equal to YScan. Thus, the top of bubble ink 1030 has been
found, and its y-
coordinate has been determined to be the current value of YScan. Step 1130
follows step
1120 if the answer at step 1120 is no, and follows step 1125 if the answer at
step 1120 is yes.
At step 1130, a check is made as to whether the current value of XScan is less
than the
current value of XLeft. If the answer is yes, then at step 1140 XLeft is set
equal to XScan.
Step 1145 follows step 1140 if the answer at step 1130 is yes, and follows
step 1130 if the
answer at step 1130 is no.
At step 1145, a check is made as to whether XScan is greater than XRight. If
the
answer is yes, then at step 1150 XRight is set equal to XScan. Step 1155
follows step 1145 if
the answer at step 1145 is no, and follows step 1150 if the answer at step
1145 is yes. At step
1155, YBottom is set equal to YScan. Step 1160, discussed above, follows step
1155. As the
process continues, the value of YTop decreases until it reflects the value of
the y-coordinate
of the uppermost pixel in bubble ink 1030; the value of YBottom increases
until it reflects the
value of the y-coordinate of the lowermost pixel in bubble ink 1030; the value
of XLeft
-13-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
decreases until it reflects the value of the x-coordinate of the leftmost
pixel in bubble ink
1030; and the value of XRight increases until it reflects the value of the x-
coordinate of the
rightmost pixel in bubble ink 1030. Thus, when step 1180 is finally reached,
the calculation
of XInkCenter (_ (XLeft + XRight) / 2) and YInkCenter ( _ (YTop + YBottom) /
2) will
precisely determine the location of the center of bubble ink 1030 if bubble
ink 1030 forms a
perfect disk, and will reasonable approximate the center of bubble ink 1030 in
any case. This
method of determining the center of bubble ink 1030 is a preferred embodiment
only, and is
not intended to limit the scope of the invention as described by the appended
claims. Those
skilled in the art will recognize that there are a variety of known methods of
determining the
I O center of an irregularly shaped two-dimensional object that could be
substituted for the above
method without departing from the scope of the disclosed invention.
Returning to FIG. 9, at step 930 an offset between ink bubble 1030 and a
reference
bubble (with center (XMast, YMast) is calculated using the formulas XOffset =
XInkCenter -
XMast and YOffset = YInkCenter - YMast. The values XOffset and YOffset are
then applied
I ~ to shift the ink image. For example, if XOffset is positive and YOffset is
negative, the ink
image is shifted to the left by a distance I XOffset j , and downward by a
distance I, YOffset ~ .
Preferably, distance is measured in number of pixels. Thus, in a preferred
embodiment, the ink image pixels are moved horizontally by j XOffset ~ pixels,
and vertically
by ~ YOffset ~ pixels. The following C++ code illustrates a preferred method
for performing
20 the ink image shift:
lIIIIIIIIIIIIIIIIIIlllllllllllllllllllllllllllllllllllllllllllllllllllllllll
// Function: CalculateBubbleOffset
//
25 //algorithm for calculating auto-alignment offsets
// Parameters:


// in


// hllmg BITMAPHANDLE * with ink pixels containing
bubble


30 // iBubbIeTop Y coordinate of top of bubble reference
zone


// iBubbIeLeft X coordinate of left of bubble reference
zone


II iBubbIeBottom Y coordinate of bottom of bubble
reference zone


// iBubbIeRight X coordinate of right of bubble
reference zone


II


35 // out


// *piOffsetX lculated XOffset or 0 as default
set to ca


// *piOffsetY lculated YOffset or 0 as default
set to ca


// returns: 0 if no bubble found
40 // 1 if bubble found
- 14-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
//
int CalculateBubbleOffset( BITMAPHANDLE *hllmg,
int iBubbIeTop,
int iBubbIeLeft,
int iBubbIeBottom,
int iBubbIeRight,
int *piOffsetX,
int *piOffsetY)
#define EXPAND Y 50 // reference search zone offset in Y
direction
#define EXPAND_X 40 // reference search zone offset in X
direction
#define ALIGN THRESHOLD 50 I/ maximum allowed auto-align in pixels
intiReturn = 0;
// initiallly no offset
*piOffsetX = *piOffsetY = 0;
COLORREF cr;
int x,y;
int iMinY = -1;
int iMinX = -1;
int iMaxY = -1;
int iMaxX = -1;
int iStartY = -max(iBubbIeTop - EXPAND_Y, 0);
int iStartX = _max(iBubbIeLeft - EXPAND X, 0);
int iEndY = iBubbIeBottom + EXPAND_Y;
int iEndX = iBubbIeRight + EXPAND X;
bool bFound = false;
for (y = iStartY; y <= iEndY; y++)
h=0;
for (x = iStartX; x <= iEndX; x++)
// check for black pixel
cr= GetPixeIColor(hLlmg, y, x);
if (cr == RGB(0,0,0))
f
if (!bFound)
bFound = true;
// initialize all coords to first black pixel
if (iMinY =_ -1 )
iMinY = y;
if (iMaxY =_ -1 )
{
iMaxY = y;
)
-15-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
if (iMinX =_ -1 )
{
iMinX = x;
}
if (iMaxX =_ -1 )
{
iMaxX = x;
}
// update mins and maxes
if (x > iMaxX)
iMaxX = x;
if (x > iMaxY)
1 S iMaxY = y;
if (x < iMinX)
iMinX = x;
}
}
}
if (bFound)
2S {
iReturn = 1;
// calculate center of bubble ink (Xlnk, Ylnk)
int iBubbIeCenterX = iMinX + (iMaxX - iMinX);
int iBubbIeCenterY = iMinY + (iMaxY - iMinY);
// calculate center of reference bubble (X0, YO)
int iReferenceCenterX = iBubbIeLeft + (iBubbIeRight - iBubbIeLeft);
int iReferenceCenterY = iBubbIeTop + (iBubbIeBottom - iBubbIeTop);
JS
// calculate automatic form alignment offsets (XOffset, YOffset)
*piOffsetX = iBubbIeCenterX - iReferenceCenterX;
*piOffsetY = iBubbIeCenterY - iReferenceCenterY;
// make sure offset is no bigger than max
*piOffsetX = min(*plOffsetX, ALIGN_THRESHOLD);
*piOffsetY = -min(*plOffsetY, ALIGN THRESHOLD);
else
4S {
TRACEO(_T("CalculateBubbleOffset no ink found in bubble zone.\n"));
}
return iReturn;
SO }
// end Function: CalculateBubbleOffset
llllllllllllllllllllllllllllllllllllllllllllllllllllllllllllllllllllllllllll
-16-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
The ink pixels are offset by merging the loaded ink with a programmatically-
created
white page "null'' image. A bitwise add algorithm that sets the pixel result
to a maximum of
255 (white) is used to merge the pixels. The following code performs this
operation:
CombineBitmapPixels(&bmpWhitePage, // created "white page" pixel data
structure
(also result destination)
0, // X origin for white page pixels
0, // Y origin for white page pixels
BITMAPWIDTH(&bmpWhitePage), // total X pixels in white page
BITMAPHEIGHT(&bmpWhitePage), // total Y pixels in white page
&bmplnk, // loaded ink pixel data structure
iXOffset, // X origin for ink pixels (XOffset)
iYOffset, II Y origin for ink pixels (YOffset)
CB OP ADD); // merge algorithm type
Once the ink image has been shifted (aligned), OMR and OCR processing occurs -
at
step 950 the aligned image is recognized.
FIG. 12 depicts a preferred ink image recognition process. At step 1210 a
field
definition file for a form is loaded. At step 1220, an ink image is loaded
into memory. At
step 1230, the field definition file and the image are passed to a recognition
engine for
recognition, and at step 1240 the recognized results are saved to a database.
Returning to FIG. 9, at step 960 the aligned, recognized ink image is aligned
with a
reference image of the identified (blank) form. This process is depicted in
FIG. 13. At step
1310, the reference image of the form is loaded into memory. At step 1320 the
ink image is
loaded into memory.
At step 1330 the ink image is merged with the reference image. The ink pixels
are
merged with the reference image pixels programmatically
using a bitwise OR algorithm to merge the pixels. The following code
performs this operation:
CombineBitmapPixels( &bmplnk, // loaded ink pixel data structure (also result
destination)
0, /I X origin for ink pixels
0, // Y origin for ink pixels
BITMAPWIDTH(&bmplnk), II total X pixels in ink
BITMAPHEIGHT(&bmplnk), // total Y pixels in ink
&bmpReference, // loaded reference pixel data structure
0, // X origin for reference pixels
0, // Y origin for ink pixels
CB OP OR); // merge algorithm type
Finally at step 1340, the merged image is saved to a disk or other storage
medium.
-17-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
FIG. 14 depicts a form identification process of a preferred embodiment. Some
of
this process is also shown in FIG. 5. At step 1410. a form set Form ID zone
definition file is
loaded into memory. At step 1420 an ink image is loaded into memory. At step
1425, a
recognition is called and the loaded zone definition and ink image are used.
At step 1430 a
variable FormID is set equal to 1. At step 1435, a check is made as to whether
the character
in the recognition results at the FormID is equal to 1. If so. the form has
been identified, and
the process ends at step 1450. If not, then step 1440. is performed, wherein
FormID is
incremented. At step 1445, a check is made as to whether FormID is greater
that the number
of forms in the form set. If so, then the process ends at step 1455 with the
form not
identified. If not, then the process returns to step 1435.
In a further embodiment of the subject invention, preprinted forms that do not
have a
form indicator printed thereon are used. To enable the use of such forms, a
template,
preferably metal or plastic, is affixed to a digitizer pad. The template acts
as a stencil (see
FIG. 15): a user fills in holes in the guide; each hole corresponds to a
different form. In this
manner, the form identification process described above is applied to forms
that do not have
preprinted form indicators.
In a preferred embodiment, a configuration such as that depicted in FIG. 15 is
used
The template is preferably a plastic overlay 1510 that fits over side paper
guides (see FIG. 16,
which shows a digitizer pad with three side paper guides 1610 on the left
side)) on a digitizer
pad (preferably a CrossPad or derivative thereof). The template 1510
preferably has three
slots 1530 that enable it to fit over three raised paper guides on the
digitizer pad. Those
skilled in the art will recognize that other methods of attachment would be
substituted for
digitizer pads of other configurations. For example, if there is only one
raised paper guide on
the digitizer pad, and it is on the right-hand side of the pad, template 1510
would have only
one slot, along its right side.
Returning to the preferred embodiment depicted in FIG. 15, the three slots
1530 are
approximately (~ 2 mm) 48 mm apart, 1 mm wide, Smm from the left-hand edge of
template
1510, and approximately (~ 1 m) 48 mm long. The top of topmost slot 1530 is
preferably 15
mm from the top edge of template 1510, and the bottom of bottom-most slot 1530
is
preferably 15 mm from the bottom edge of template 1510.
- 18-


CA 02397151 2002-07-11
WO 01/52519 PCT/USO1/00619
Template 1310 itself is approximately 2 mm thick, 273 mm long. and 27 mm wide.
Each form indicator bubble 1520 is preferably 2.~ mm in diameter, and the
minimum
preferred distance between bubbles 1520 is 4 mm.
While the embodiments shown and described are fully capable of achieving the
objects of the invention. it is to be understood that these embodiments are
shown only for the
purpose of illustration and not for the purpose of limitation. For example, a
CrossPad could
be easily be modified so that a user is able to skip from page to page (e.g.,
from page 14
directly to page 3) instead of having to repeatedly hit the page-up and page-
down buttons.
-19-

Representative Drawing
A single figure which represents the drawing illustrating the invention.
Administrative Status

For a clearer understanding of the status of the application/patent presented on this page, the site Disclaimer , as well as the definitions for Patent , Administrative Status , Maintenance Fee  and Payment History  should be consulted.

Administrative Status

Title Date
Forecasted Issue Date Unavailable
(86) PCT Filing Date 2001-01-09
(87) PCT Publication Date 2001-07-19
(85) National Entry 2002-07-11
Dead Application 2004-10-14

Abandonment History

Abandonment Date Reason Reinstatement Date
2003-10-14 FAILURE TO RESPOND TO OFFICE LETTER
2004-01-09 FAILURE TO PAY APPLICATION MAINTENANCE FEE

Payment History

Fee Type Anniversary Year Due Date Amount Paid Paid Date
Application Fee $300.00 2002-07-11
Maintenance Fee - Application - New Act 2 2003-01-09 $100.00 2002-07-11
Owners on Record

Note: Records showing the ownership history in alphabetical order.

Current Owners on Record
VANSICKLE, SCOTT
LEONE, ARMAND JR.
GROSSBERG, DAVID
READ, SUSAN B.
BERGELSON, ALAN P.
Past Owners on Record
None
Past Owners that do not appear in the "Owners on Record" listing will appear in other documentation within the application.
Documents

To view selected files, please enter reCAPTCHA code :



To view images, click a link in the Document Description column. To download the documents, select one or more checkboxes in the first column and then click the "Download Selected in PDF format (Zip Archive)" or the "Download Selected as Single PDF" button.

List of published and non-published patent-specific documents on the CPD .

If you have any difficulty accessing content, you can call the Client Service Centre at 1-866-997-1936 or send them an e-mail at CIPO Client Service Centre.


Document
Description 
Date
(yyyy-mm-dd) 
Number of pages   Size of Image (KB) 
Abstract 2002-07-11 2 84
Representative Drawing 2002-07-11 1 27
Cover Page 2002-12-10 1 54
Claims 2002-07-11 9 300
Drawings 2002-07-11 15 333
Description 2002-07-11 19 942
PCT 2002-07-11 3 117
Assignment 2002-07-11 4 125
Correspondence 2002-12-06 1 25
PCT 2002-07-12 3 182
PCT 2002-07-11 1 86