Language selection

Search

Patent 2820744 Summary

Third-party information liability

Some of the information on this Web page has been provided by external sources. The Government of Canada is not responsible for the accuracy, reliability or currency of the information supplied by external sources. Users wishing to rely upon this information should consult directly with the source of the information. Content provided by external sources is not subject to official languages, privacy and accessibility requirements.

Claims and Abstract availability

Any discrepancies in the text and image of the Claims and Abstract are due to differing posting times. Text of the Claims and Abstract are posted:

  • At the time the application is open to public inspection;
  • At the time of issue of the patent (grant).
(12) Patent Application: (11) CA 2820744
(54) English Title: PORTABLE ELECTRONIC DEVICE WITH SEMI-TRANSPARENT, LAYERED WINDOWS
(54) French Title: DISPOSITIF ELECTRONIQUE PORTABLE A FENETRES MULTICOUCHES SEMI-TRANSPARENTES
Status: Deemed Abandoned and Beyond the Period of Reinstatement - Pending Response to Notice of Disregarded Communication
Bibliographic Data
(51) International Patent Classification (IPC):
  • G06F 3/01 (2006.01)
  • G06F 3/023 (2006.01)
  • G06F 15/02 (2006.01)
  • H03M 11/00 (2006.01)
  • H04W 88/02 (2009.01)
(72) Inventors :
  • MARTIN-COCHER, GAELLE CHRISTINE (Canada)
  • SCOTT, SHERRYL LEE LORRAINE (Canada)
(73) Owners :
  • BLACKBERRY LIMITED
(71) Applicants :
  • BLACKBERRY LIMITED (Canada)
(74) Agent: BORDEN LADNER GERVAIS LLP
(74) Associate agent:
(45) Issued:
(86) PCT Filing Date: 2011-11-16
(87) Open to Public Inspection: 2012-06-28
Examination requested: 2013-06-07
Availability of licence: N/A
Dedicated to the Public: N/A
(25) Language of filing: English

Patent Cooperation Treaty (PCT): Yes
(86) PCT Filing Number: PCT/CA2011/001266
(87) International Publication Number: WO 2012083416
(85) National Entry: 2013-06-07

(30) Application Priority Data:
Application No. Country/Territory Date
10194572.3 (European Patent Office (EPO)) 2010-12-10
12/965,560 (United States of America) 2010-12-10

Abstracts

English Abstract

Methods and systems for selecting input in a portable electronic device comprising a display and a plurality of input members are disclosed. The methods and system use semi-transparent windows displayed in partially overlapping layers to present output options to the user of the device. For example, the method includes detecting a pre-determined type of actuation of one of the input members of the electronic device and determining the representations associated with the actuated input member. The method further includes outputting the representations using the display, each representation appearing in a semi transparent window, the semi-transparent windows being displayed in partially overlapping layers. The method further includes receiving a selection input and outputting the representation displayed in the top-most semi transparent window using the display.


French Abstract

La présente invention concerne des procédés et des systèmes permettant de sélectionner une saisie dans un dispositif électronique portable comprenant un écran d'affichage et plusieurs éléments de saisie. Les procédés et systèmes utilisent des fenêtres semi-transparentes qui sont affichées sous forme de couches se chevauchant partiellement afin de présenter des options de création à l'utilisateur du dispositif. Par exemple, le procédé consiste à détecter un type prédéfini d'actionnement d'un des éléments de saisie du dispositif électronique et à déterminer les représentations associées à l'élément de saisie actionné. Le procédé consiste également à produire les représentations à l'aide de l'écran d'affichage, chaque représentation apparaissant dans une fenêtre semi-transparente, les fenêtres semi-transparentes étant affichées sous forme de couches se chevauchant partiellement. Le procédé consiste également à recevoir une saisie sélectionnée et à produire la représentation affichée dans la fenêtre semi-transparente supérieure à l'aide de l'écran d'affichage.

Claims

Note: Claims are shown in the official language in which they were submitted.


WHAT IS CLAIMED IS:
1. A method of selecting input in a portable electronic device comprising a
display
and a plurality of input members, the method comprising:
detecting a pre-determined type of actuation of at least one of the plurality
of input members;
determining a plurality of representations associated with the actuated input
member;
outputting the plurality of representations on the display, each
representation of the plurality of representations appearing in a
semi-transparent window, the plurality of semi-transparent windows
being output in partially overlapping layers;
receiving a selection input; and
outputting a representation on the display, the representation being in a
semi-transparent window that is a top layer of the layers.
2. The method of claim 1, wherein the pre-determined type of actuation is a
press-
and-hold of the actuated input member and the selection input is a release
of the actuated input member.
3. The method of claim 1, wherein the plurality of representations include a
plurality of accented characters corresponding to at least one character
associated with the actuated input member.
4. The method of claim 1, wherein the plurality of representations include a
plurality of emoticons, each emoticon beginning with a character
associated with the actuated input member.
5. The method of claim 1, wherein the plurality of representations include a
plurality of short message service abbreviations, the first character of each
abbreviation associated with a character associated with the actuated input
member.
18

6. The method of claim 1 wherein the pre-determined type of actuation is an
actuation corresponding with the end of a sentence.
7. The method of claim 6, wherein the plurality of representations include a
plurality of punctuation marks.
8. The method of claim 6, wherein the plurality of representations include a
plurality of emoticons.
9. The method of claim 1, further comprising:
detecting a pre-determined input; and
in response to detecting the input, modifying the output of the semi-
transparent windows so that the semi-transparent window that is the
top layer moves to a back layer of the layers.
10. The method of claim 9, wherein the pre-determined input includes
continuing
to press-and-hold the actuated input member for a pre-determined amount
of time.
11. A computer-readable medium having computer-readable code executable by
at least one processor of the portable electronic device to perform the method
of claim 1.
12. A portable electronic device comprising:
a processor;
an output apparatus;
a plurality of input members; and
a memory comprising a plurality of representations associated with at least
some of the plurality of input members;
the processor being adapted to:
detect a pre-determined type of actuation of at least one of the
plurality of input members,
determine a plurality of representations associated with the actuated
input member,
19

cause the output of the plurality of representations using the output
apparatus, such that each representation of the plurality of
representations appears in a semi-transparent window, the
plurality of semi-transparent windows being output in partially
overlapping layers,
receive a selection input, and
cause a representation to be output using the output apparatus, the
representation being in a semi-transparent window that is a
top layer of the layers.
13. The device of claim 12, wherein the processor is further adapted to:
detect a pre-determined input; and
in response to detecting the input, modify the output of the semi-
transparent windows so that a semi-transparent window in a top
layer of the layers moves to a back layer of the layers.
14. The device of claim 13, wherein the pre-determined input includes
continuing
to press-and-hold the actuated input member for a pre-determined amount
of time and the selection input is a release of the actuated input member.
15. A method of enabling disambiguation of an input into a portable electronic
device, the portable electronic device including an input apparatus, a
display,
and a processor, and a memory having a plurality of language objects stored
therein, each language object being associated with a frequency, the input
apparatus including a plurality of input members, each of at least a portion
of
the input members having a plurality of characters assigned thereto, the
method comprising:
detecting a text input including a plurality of input member actuations, at
least one of the input member actuations being an ambiguous input;
generating a number of character permutations corresponding with the text
input, at least some of the permutations corresponding with a
language object;
20

generating an output set of at least a portion of the character permutations;
and
displaying the output set on the display, each character permutation in the
output set appearing in a semi-transparent window and the semi-
transparent windows being displayed in partially overlapping layers.
16. A method of selecting input in a portable electronic device comprising a
display and a plurality of input members, the method comprising:
detecting a pre-determined type of actuation of the plurality of input
members;
outputting, on the display, a plurality of text shortcuts in a first
semi-transparent designated window with one of the plurality of text
shortcuts being marked for selection;
outputting, on the display, a description of the text shortcut marked for
selection in a second semi-transparent window, wherein the first
semi-transparent window and the second semi-transparent window
are output in partially overlapping layers;
receiving a selection of the text shortcut marked for selection; and
outputting the text shortcut marked for selection on the display.
17. The method of claim 16, wherein the pre-determined type of actuation is a
punctuation mark ending a sentence and a space.
18. The method of claim 16, further comprising outputting a text equivalent of
the
text shortcut marked for selection in a third semi-transparent window,
wherein the first, the second, and the third semi-transparent windows are
output in partially overlapping layers.
19. The method of claim 16, further comprising determining that text shortcuts
are
enabled before outputting the first and the second windows.
20. The method of claim 16, further comprising receiving a scroll input and in
response to the scroll input, changing the text shortcut marked for
selection.
21

21. The method of claim 16, further comprising receiving a window scroll input
and
in response to the window scroll input, causing a semi-transparent window
that is a top layer of the layers to move to a bottom layer of the layers.
22

Description

Note: Descriptions are shown in the official language in which they were submitted.


CA 02820744 2013-06-07
WO 2012/083416
PCT/CA2011/001266
PORTABLE ELECTRONIC DEVICE WITH SEMI-TRANSPARENT,
LAYERED WINDOWS
FIELD OF TECHNOLOGY
[001] The present disclosure relates to providing input to portable
electronic devices, including but not limited to portable electronic devices
having
touch screen displays and, more specifically, to a user-interface using semi-
transparent, layered windows for selecting input in such devices.
BACKGROUND
[002] Electronic devices, including portable electronic devices, have
gained widespread use and may provide a variety of functions including, for
example, telephonic, electronic messaging, and other personal information
manager (PIM) application functions. Portable electronic devices include, for
example, several types of mobile stations such as simple cellular telephones,
smart telephones, wireless personal digital assistants (PDAs), and laptop
computers with wireless communication capabilities based on, for example, the
802.11 or Bluetoothe communications protocols.
[003] Portable electronic devices such as PDAs or smart telephones are
generally intended for handheld use and ease of portability. Smaller devices
are
generally desirable for portability. A touch-sensitive display, also known as
a
touchscreen display, is particularly useful on handheld devices, which are
small
and have limited space for user input and output. The information displayed on
the touch-sensitive displays may be modified depending on the functions and
operations being performed. With continued demand for decreased size of
portable electronic devices to facilitate portability, touch-sensitive
displays
continue to decrease in size.
[004] The decrease in the size of the portable electronic devices and their
display areas has resulted in screens overloaded with information. For
example,
when using electronic devices with a reduced keyboard, disambiguation results
often obscure text already composed by the user. Displays for accessing
special
characters also cover much of the screen. Furthermore, accessing special
characters and other symbols, such as emoticons, can be cumbersome because
1

CA 02820744 2013-06-07
WO 2012/083416 PCT/CA2011/001266
the user must interrupt the typing process to search for special keys, perform
special keystroke combinations, or use menus to input special characters and
other symbols.
BRIEF DESCRIPTION OF THE DRAWINGS
[001] The accompanying drawings, which are incorporated in and
constitute a part of this specification, illustrate several example
embodiments of
the present disclosure. In the drawings:
[002] Figure 1 is a block diagram of a portable electronic device,
consistent with disclosed example embodiments;
[003] Figure 2 is a top plan view of a portable electronic device, consistent
with disclosed example embodiments;
[004] Figure 3 is a flow diagram of an example process using a layered
user-interface to select representations, consistent with disclosed example
embodiments;
[005] Figures 4, 5, and 6A-B each show an example output of an
improved portable electronic device for selecting representations, consistent
with
disclosed example embodiments;
[006] Figures 7 to 9 each show an example output created during a
process to select representations, consistent with disclosed example
embodiments;
[007] Figure 10 is a flow diagram of an example process for selecting text
shortcuts, consistent with disclosed example embodiments;
[008] Figures 11 to 13 each show an example output of an improved
portable electronic device for selecting emoticons, consistent with disclosed
example embodiments; and
[009] Figures 14 to 16 each show an example output of an improved
portable electronic device used to display text disambiguation options,
consistent
with disclosed example embodiments.
DETAILED DESCRIPTION
[010] Reference will now be made in detail to the exemplary embodiments
of the invention, examples of which are illustrated in the accompanying
drawings.
2

CA 02820744 2013-06-07
WO 2012/083416 PCT/CA2011/001266
For simplicity and clarity of illustration, reference numerals may be repeated
among the figures to indicate corresponding or analogous elements. Numerous
details are set forth to provide an understanding of the embodiments described
herein. The embodiments may be practiced without these details. In other
instances, well-known methods, procedures, and components have not been
described in detail to avoid obscuring the embodiments described. The
description is not to be considered as limited to the scope of the embodiments
described herein.
[011] The disclosure generally relates to a portable electronic device.
Examples of portable electronic devices include mobile, or handheld, wireless
communication devices such as pagers, cellular phones, cellular smart-phones,
wireless organizers, personal digital assistants, wirelessly enabled notebook
computers, netbooks, tablets, and so forth. The portable electronic device may
also be a portable electronic device without wireless communication
capabilities,
such as a handheld electronic game device, digital photograph album, digital
camera, or other portable device.
[012] A block diagram of an example of a portable electronic device 100 is
shown in Figure 1. Portable electronic device 100 includes multiple
components,
such as processor 102 that controls the overall operation of the portable
electronic
device 100. Processor 102 may be, for instance, and without limitation, a
microprocessor (pP). Communication functions, including data and voice
communications, are performed through communication subsystem 104. Data
received by the portable electronic device 100 is optionally decompressed and
decrypted by a decoder 106. Communication subsystem 104 receives messages
from and sends messages to a wireless network 150. Wireless network 150 may
be any type of wireless network, including, but not limited to, data wireless
networks, voice wireless networks, and networks that support both voice and
data
communications. Power source 142, such as one or more rechargeable batteries
or a port to an external power supply, power portable electronic device 100.
[013] Processor 102 interacts with other components, such as Random
Access Memory (RAM) 108, memory 110, and display 112. In example
embodiments display 112 has a touch-sensitive overlay 114 operably connected
or coupled to an electronic controller 116 that together comprise touch-
sensitive
3

CA 02820744 2013-06-07
WO 2012/083416
PCT/CA2011/001266
display 112. Processor 102 interacts with touch-sensitive overlay 114 via
electronic controller 116. User-interaction with a graphical user interface is
performed through the touch-sensitive overlay 114. Information, such as text,
characters, symbols, images, icons, and other items that may be displayed or
rendered on portable electronic device 100, are displayed on the touch-
sensitive
display 112 via the processor 102. Although described as a touch-sensitive
display with regard to Figure 1, display 112 is not limited to a touch-
sensitive
display and can include any display screen for portable devices.
[014] Processor 102 also interacts with one or more actuators 120, one or
more force sensors 122, auxiliary input/output (I/O) subsystem 124, data port
126,
speaker 128, microphone 130, short-range communications 132, and other device
subsystems 134. Processor 102 interacts with accelerometer 136, which may be
utilized to detect direction of gravitational forces or gravity-induced
reaction forces.
[015] To identify a subscriber for network access, portable electronic
device 100 uses a Subscriber Identity Module or a Removable User Identity
Module (SIM/RUIM) card 138 for communication with a network, such as wireless
network 150. Alternatively, user identification information may be programmed
into memory 110.
[016] Portable electronic device 100 includes operating system 146 and
software programs or components 148 that are executed by the processor 102
and may be stored in a persistent, updatable store such as memory 110.
Additional applications or programs are loaded onto portable electronic device
100
through the wireless network 150, auxiliary I/O subsystem 124, data port 126,
short-range communications subsystem 132, or any other suitable subsystem
134.
[017] A received signal such as a text message, an e-mail message, or
web page download is processed by communication subsystem 104 and input to
processor 102. Processor 102 processes the received signal for output to
display
112 and/or to auxiliary I/O subsystem 124. A subscriber may generate data
items,
for example e-mail or text messages, which may be transmitted over wireless
network 150 through communication subsystem 104. For voice communications,
the overall operation of the portable electronic device 100 is similar.
Speaker 128
outputs audible information converted from electrical signals, and microphone
130
4

CA 02820744 2013-06-07
WO 2012/083416 PCT/CA2011/001266
converts audible information into electrical signals for processing. Speaker
128,
display 112, and data port 126 are considered output apparatus of device 100.
[018] A touch-sensitive display 112 may be any suitable touch-sensitive
display, such as a capacitive, resistive, infrared, surface acoustic wave
(SAW)
touch-sensitive display, strain gauge, optical imaging, dispersive signal
technology, acoustic pulse recognition, and so forth, as known in the art. A
capacitive touch-sensitive display includes capacitive touch-sensitive overlay
114.
Overlay 114 is an assembly of multiple layers in a stack including, for
example, a
substrate, a ground shield layer, a barrier layer, one or more capacitive
touch
sensor layers separated by a substrate or other barrier, and a cover. The
capacitive touch sensor layers are any suitable material, such as patterned
indium
tin oxide (ITO).
[019] One or more touches, also known as touch contacts, touch events,
or actuations, are detected by touch-sensitive display 112. The processor 102
or
controller 116 determines attributes of the touch, including a location of a
touch.
Touch location data includes an area of contact or a single point of contact,
such
as a point at or near a center of the area of contact. The location of a
detected
touch may include x and y components, e.g., horizontal and vertical
components,
respectively, with respect to one's view of touch-sensitive display 112. For
example, the x location component may be determined by a signal generated from
one touch sensor, and the y location component may be determined by a signal
generated from another touch sensor. A signal may be provided to controller
116
in response to detection of a touch. A touch may be detected from any suitable
object, such as a finger, thumb, appendage, or other items, for example, a
stylus,
pen, or other pointer, depending on the nature of touch-sensitive display 112.
Multiple simultaneous touches or gestures are also detected. These multiple
simultaneous touches may be considered chording events.
[020] In some example embodiments, one or more actuators 120 may be
depressed by applying sufficient force to the touch-sensitive display 112 to
overcome the actuation force of the actuator 120. Actuator 120 is actuated by
pressing anywhere on touch-sensitive display 112. Actuator 120 provides input
to
the processor 102 when actuated. Actuation of the actuator 120 results in
provision of tactile feedback.

CA 02820744 2013-06-07
WO 2012/083416
PCT/CA2011/001266
[021] In certain embodiments, a mechanical dome switch may be utilized
as one or more of actuators 120. In this example, tactile feedback is provided
when the dome collapses due to imparted force and when the dome returns to the
rest position after release of the switch.
[022] Alternatively, actuator 120 may comprise one or more piezoelectric
(piezo) devices that provide tactile feedback for the touch-sensitive display
112.
Contraction of the piezo actuators applies a spring-like force, for example,
opposing a force externally applied to the touch-sensitive display 112. Each
piezo
actuator includes a piezoelectric device, such as a piezoelectric (PZT)
ceramic
disk adhered to a metal substrate. The metal substrate bends when the PZT disk
contracts due to build up of charge at the PZT disk or in response to a force,
such
as an external force applied to touch-sensitive display 112. The charge may be
adjusted by varying the applied voltage or current, thereby controlling the
force
applied by the piezo disks. The charge on the piezo actuator may be removed by
a controlled discharge current that causes the PZT disk to expand, releasing
the
force thereby decreasing the force applied by the piezo disks. The charge may
advantageously be removed over a relatively short period of time to provide
tactile
feedback to the user. Absent an external force and absent a charge on the
piezo
disk, the piezo disk may be slightly bent due to a mechanical preload.
Actuator
120, touch-sensitive display 112, force sensor 122, microphone 130, and data
port
126 are input apparatuses for device 100.
[023] A top plan view of portable electronic device 100 is shown
generally in Figure 2. Example portable electronic device 100 includes housing
200 in which various components as shown in Figure 1 are disposed. For
example, various input apparatuses and output apparatuses, processor 102, and
memory 110 for storing at least programs 148 are disposed in housing 200.
Processor 102 is responsive to input signals from input apparatuses, such as
the
display 112 or actuator 120, and optionally provides output signals to output
apparatuses, such as the display 112 or speaker 128. Processor 102 also
interfaces with memory 110 and is enabled to execute programs 148.
[024] As can be understood from Figure 2, the output apparatus includes
display 112 and speaker 128, each of which are responsive to one or more
output
signals from processor 116 or processor 102. The input apparatuses includes
6

CA 02820744 2013-06-07
WO 2012/083416 PCT/CA2011/001266
keyboard 220. As described above, input members 225 on keyboard 220 may be
rendered on touch-sensitive display 112. For example, each input member can
be defined by specific coordinates of display 112. Alternatively, input
members
225 are mechanical keys using, for example, a mechanical dome switch actuator
or a piezoelectric actuator. In certain embodiments, input members 225 may
form
a QWERTY keyboard or other known keyboard layouts, either in reduced or full
format. In a reduced keyboard layout, input members are assigned a number of
characters. In other example embodiments, input members 225 may form an
alphabetical keyboard layout. Whether input members 225 are rendered using
touch sensitive display 112 or are mechanical, input members 225 may be
capable of a press-and-hold operation. In a press-and-hold operation a user
actuates (presses) the input member and continues pressing the input member
for
a period of time. Processor 102 or controller 116 are configured to detect a
first
input when input member is pressed and a second input when the user continues
to press an individual input member.
[025] In the presently described example embodiment shown in Figure 2,
each input member 225 corresponds to a number of characters or other
linguistic
elements. Input members 225 as shown in Figure 2 correspond generally to three
characters, with some input members corresponding to two or one characters.
However, in other example embodiments, keyboard 220 corresponds to a full
keyboard, and each input member 225 corresponds generally to only one
character. As used herein, characters refers to letters, numbers, or symbols
found on a keyboard. Special characters refers broadly to letters, numbers, or
symbols that are part of a font, but not necessarily displayed on the
keyboard,
such as accented letters, diacritics or foreign currency symbols.
[026] Although not shown in Figure 2, handheld electronic device 100
may include other input apparatuses, such as a scroll wheel, an optical
trackpad,
or a ball located either on the face or side of device 100. These input
apparatuses
provide additional input processor 102. For example, a scroll wheel may
provide
one input to processor 102 when rotated and a second input to processor 102
when actuated. An optical trackpad may provide one input to processor 102 when
swiped and a second input to processor 102 when pressed or tapped. Input
7

CA 02820744 2013-06-07
WO 2012/083416
PCT/CA2011/001266
members 225, actuators 120, and other input apparatuses, such as a scroll
wheel
or trackpad, are generally considered input members.
[027] Figure 3 is a flow diagram of an example process 300 using a
layered user-interface to select a special character, consistent with
disclosed
example embodiments. The method is carried out by software or firmware, for
example as part of programs 148, that is stored in Random Access Memory
(RAM) 108 or memory 110, and is executed by, for example, the processor 102 as
described herein, or by controller 116. Process 300 is used to select
representations associated with the characters corresponding to an input
member
for output to an output apparatus. Representations include accented
characters,
symbols, special characters, or punctuation marks. Representations may also
include text shortcuts, such as emoticons or chat acronyms used, for example,
in
short message service (SMS), instant message (IM), or BlackBerry Messenger
(BBM) sessions, that begin with a character corresponding to the input member.
The process is used to select a representation using as little display 112
space as
possible while still presenting multiple options to the user.
[028] At Step 305, processor 102 detects a pre-determined type of
actuation of an input member. As employed herein, the expression "actuation"
and variations thereof shall refer broadly to any way of activating an input
member, including pressing down on, tapping, or touching the input member. In
certain embodiments, the pre-determined type of actuation is a press-and-hold
of
the input member. In other example embodiments, the pre-determined type of
actuation includes chording of the input member and another input member. For
example, such a chording could consist of pressing the input member associated
with the character "E" and the "Enter" input member at the same time. In some
example embodiments the actuation includes a specific sequence, such as the
entry of a punctuation mark used to complete a sentence followed by the entry
of
a "space" input member.
[029] At Step 310, processor 102 determines the representations that
are associated with the actuated input member. In certain embodiments, the
representations include accented characters. For example the representations
of
"e," "e," "6," and "e" may be associated with the input member for the
character
"e." A representation may also include symbols, such that the dollar sign "$"
is
8

CA 02820744 2013-06-07
WO 2012/083416 PCT/CA2011/001266
associated with the input member for the character "d." In other example
embodiments representations also include emoticons. For example, (happy
face) and 0 (sad face) may be associated with the input member for the colon
character, which is the character that begins the text equivalent of these
emoticons, namely ":)" and ":(". In yet other example embodiments, all
emoticons
are associated with the input member for the colon character, not just those
that
start with a colon. A representation may also include chat acronyms, such that
LOL, L8R, and LMK are associated with the input member for the character "L."
In other example embodiments, punctuation marks used at the end of a sentence
are representations associated with the period input member. Memory 110 stores
a table of the representations associated with each input member.
[030] At Step 315, processor 102 optionally orders the representations
so that the most frequently used representation appears first in a list. As
previously described, memory 110 stores an association between a
representation and an input member. In addition, memory 110 also stores a
frequency object for each representation. In the example embodiment presently
described, processor 102 orders the representations based on the frequency
objects. In addition, processor 102 updates the frequency objects when a user
selects a representation. Thus, the frequency objects may reflect the
frequency
with which a user uses a particular representation. In other example
embodiments, processor 102 orders the representations so that the most
probable
representation appears first in a list. In such embodiments, processor 102
uses a
dictionary, wordlist, vocabulary, or other corpus of words, stored in memory
110,
to determine what representation is most likely to come next, given what has
already been input by the user.
[031] In Step 320, processor 102 creates a semi-transparent window for
each representation. A window is a designated display area rendered on display
112. Next, in Step 325, processor 102 displays the semi-transparent windows in
partially overlapping (offset) layers. Such a layered user-interface allows
the user
to clearly see not only the representation on the top layer, but also the
representation contained in the window(s) just under the top window, thereby
economizing display area. In some example embodiments, the user is also
capable of viewing the representations even further down in the layers. The
offset
9

CA 02820744 2013-06-07
WO 2012/083416
PCT/CA2011/001266
allows a user to see the representation through the top window. Rather than
appearing directly behind the representation in the top window, the offset
allows
the representation in the window behind the top window to appear, for example,
slightly higher and to the right of the representation in the top window.
[032] Figures 4-6 each show an example layered user-interface of semi-
transparent offset windows, in accordance with example embodiments described
herein. As seen in Figure 4, the top semi-transparent window 405 contains a
representation of an "é" character. Behind it, window 410 contains a
representation of an "e" character. As seen in Figure 4, the character "e" in
a
subsequent semi-transparent window 410 is visible through semi-transparent
window 405. Other representations may be visible through windows 405 and 410,
although the representations would be fainter than the "e" of window 410 as
the
layers progress. Figure 5 shows chat acronyms in a layered user-interface of
semi-transparent windows in offset layers. Figures 6A and 6B show emoticons in
such a layered user-interface. The amount of screen space used by the layers
depends on the amount of offset. The smaller the offset, the less screen space
needed by the layers of windows. For example, the representations in Figure 6A
are larger than the representations in Figure 6B. Because of this, offsets 620
(in a
y direction) and 625 (in an x direction) of Figure 6A are larger than offsets
620'
and 625' of Figure 6B. This occurs because more space is needed to move the
representation in window 610 above and to the right of the representation in
window 605 than is needed to move the representation in window 610' above and
to the right of the representation in window 605'. Although offsets 620 and
625
are shown equal in size in Figure 6A, the offsets need not be of equal size.
The
offset is generally determined by the amount of space needed to make the
representation in window 410 appear above or adjacent to the representation in
window 405.
[033] In Step 330, processor 102 determines if a scroll input has been
detected. In some example embodiments, a scroll input is detected when the
actuated input member is still being held after a pre-determined length of
time.
For example, the display of Figure 4 appears after a press-and-hold actuation
of
the input member for the character "e." When the user continues to hold the
input
member for the character "e," after one second, for example, processor 102

CA 02820744 2013-06-07
WO 2012/083416
PCT/CA2011/001266
determines this constitutes a scroll input. In some example embodiments where
the pre-determined type of actuation includes chording, processor 102
determines
a scrolling input has been received through actuation of a down or up arrow,
use
of a scroll wheel, track pad, track ball, optical mouse, gesturing on a touch
screen,
or any other input commonly used for scrolling. If the scrolling input has
been
detected (Step 330, Yes), then processor 102 causes the top window to move to
the bottom of the layers. In an example using Figure 4, processor 102 moves
window 405 with the "é" representation to the position behind window 415. This
results in window 410 with the "6" representation displayed on the top layer.
In
certain embodiments, a representation in the top window is considered marked
for
selection. Thus, after the movement of window 405 to the bottom, the
representation in window 410 is marked for selection.
[034] If no scroll input is detected (Step 330, No) then, in Step 340,
processor 102 determines whether a selection input has been received. If no
selection input has been received (Step 340, No), then processing continues at
step 330, with processor 102 awaiting further input. If a selection input has
been
received (Step 340, Yes), then in Step 345, processor 102 selects the
representation displayed in the window in the top layer for output to the
output
apparatus as the desired input and removes the display of semi-transparent
layered windows. In certain embodiments, a selection input includes releasing
the
actuated input member. In other example embodiments, where the pre-
determined type of actuation includes chording, a selection input includes
using
the "Enter" input member, tapping or touching on touch-sensitive display 112,
or
an optical trackpad selection.
[035] An example of a process using a layered user-interface to select a
character with an accent mark will now be explained using Figures 7 through 9.
A
user enters text for a message, such as the message shown in Figure 7. In the
example of Figures 7 to 9, the user desires to type the word "Noel" as part of
the
message. After the user types the first two characters of "Noel," processor
102
awaits the next input, with cursor 705 marking the position of the next text
entry,
as shown in Figure 7. To enter the next character, the user may press-and-hold
the input member used to input the character "e." Processor 102, recognizing
this
as a pre-determined type of actuation, then retrieves the representations
11

CA 02820744 2013-06-07
WO 2012/083416
PCT/CA2011/001266
associated with the "e" input member, and creates the layered display seen in
Figure 7, items 710, 715, 720, and 725.
[036] The "é" representation shown in window 710 is not the character
desired by the user, so the user continues to press-and-hold the "e" input
member. After a pre-determined amount of time, such as one second, processor
102 detects that the user is still holding (pressing on) the input member
associated with character "e", and recognizes this as a scroll input.
Detection of
the scroll input causes processor 102 to move the top window, currently window
710, to the bottom of the layers. Figure 8 shows an example of the display
that
results from the movement of window 710 to the bottom layer.
[037] The transparency of the windows aids the user in determining
when to release the pressed input member, because the user is able to see what
character or other symbol will be on top after the next rotation. When the
user
sees that the correct character is showing in the top layer, the user releases
the
"e" input member. When processor 102 detects the release of the input member,
it interprets this as the receipt of a selection input. After receiving the
selection
input, processor 102 inserts the "e" at the position of cursor 705, resulting
in the
display shown in Figure 9. While this example used a press-and-hold as the pre-
determined type of actuation, other types of actuations may be used in the
process.
[038] Figure 10 is a flow diagram of another example input selection
process used to select text shortcuts, such as emoticons or chat acronyms,
consistent with disclosed example embodiments. The process is carried out by
software or firmware, stored as part of programs 148 that is stored in Random
Access Memory (108) or memory 110, and is executed by, for example, processor
102 or controller 116. This process is used to select for output text
shortcuts,
such as emoticons or chat acronyms, that often occur at the beginning or end
of a
sentence. The process is used to select these types of representations while
using as little display 112 space as possible while still presenting multiple
options
to the user.
[039] In Step 1005, processor 102 detects an end of a sentence. For
example, the end of a sentence may be detected by detecting an actuation of
one
or more input members used to mark the end a sentence followed by a space.
12

CA 02820744 2013-06-07
WO 2012/083416
PCT/CA2011/001266
Such punctuation marks may include a period, a question mark, or an
exclamation
mark. In Step 1010, processor 102 determines whether the user wants text
shortcuts displayed. In certain embodiments, the user indicates through
keyboard
options that emoticons and chat acronyms should not be automatically
displayed.
In other example embodiments, the user indicates that representations such as
emoticons and chat acronyms should never be displayed.
[040] If the user does not desire text shortcuts to be displayed (Step 1015,
No), the process ends. But if the user indicates text shortcuts may be
displayed
(Step 1015, Yes), then in Step 1020, processor 102 creates a first semi-
transparent window that contains a plurality of text shortcuts, such as
emoticons.
In certain embodiments, one of the emoticons is marked for selection. Such a
marking includes, but is not limited to, a box around the emoticon, a
different
background color for the emoticon, or the emoticon appearing larger than the
other emoticons. In other example embodiments, a plurality of chat acronyms is
displayed instead of emoticons.
[041] In Step 1025, processor 102 creates a second semi-transparent
window that contains the text equivalents of the plurality of emoticons in the
first
semi-transparent window. In Step 1030, processor creates a third semi-
transparent window that contains the names of the plurality of emoticons
contained in the first window.
[042] In Step 1035, processor 102 displays multiple (three in one
embodiment) semi-transparent windows in partially overlapping (offset) layers.
For example, the first window with the plurality of emoticons appears on top
and
the text equivalent and name of the emoticon marked for selection appears
through the first semi-transparent window. In certain embodiments, the user
selects which of the three windows is displayed on top by default. For
example, if
a user desires to see the text equivalents on top, processor 102 displays a
window showing the text equivalents of the emoticons on top, and the emoticons
and the emoticon names appear in semi-transparent windows behind the text
equivalent window. In such embodiments, one of the text equivalents is marked
for selection and the emoticon and name corresponding to the text equivalent
marked for selection are seen through the top window. In other example
embodiments, the user may desire to see the names on the top window, with the
13

CA 02820744 2013-06-07
WO 2012/083416
PCT/CA2011/001266
emoticons and the text equivalents displayed in the lower layers. In such
embodiments, the emoticon and the text equivalent corresponding to the name
marked for selection would show through the top window.
[043] In Step 1040, processor 102 determines if a selection input has
been received. If so (Step 1040, Yes), then in Step 1045, the emoticon marked
for selection is output and the semi-transparent layers of windows are removed
from the display. If not (Step 1040, No), in Step 1050, processor 102
determines
if a scroll input has been received. If so (Step 1050, Yes), in Step 1055,
processor 102 changes the emoticon marked for selection and then returns to
Step 1035 to re-create the display of the layered windows. For example, if the
emoticon marked for selection changes from (a smile) to 0 (a sad face), then
processor 102 causes the name and text equivalent of the (sad face) emoticon
to show behind the top window. If no scroll input is detected (Step 1050, No),
processor 102 waits for further input.
[044] In certain embodiments, processor 102 receives a window scroll
input (not shown in Figure 10). As described above, the window scroll input
may
include a chording, such as holding down a "Shift" or "Control" input member
while
pressing an up or down arrow or using a mouse, trackball, optical pad, or
scroll
wheel. The window scroll input may also include an up and down scrolling on a
mouse, optical pad, or trackball while a side-to-side motion may be used as
the
scroll input to change the emoticon marked for selection. When processor 102
receives a window scroll input, it causes the top semi-transparent window to
move
to the bottom of the layers. Using a window scroll input a user changes the
window displayed on top between the emoticon window, the text equivalent
window, and the name window.
[045] An example of a process using a layered user-interface to select
an emoticon will now be explained using Figures 11 - 13. A user may type a
text
message, as shown in Figure 11. The position of cursor 705 demonstrates where
the next output will appear in the message. As shown in Figure 11, the user
types
the phrase "Just kidding. "and in response processor 102 detects the end of a
sentence by detecting that the user has entered a period followed by a space.
Processor 102 also detects the end of a sentence when other sentence
delimiters,
such as a question mark and a space or an exclamation mark and a space are
14

CA 02820744 2013-06-07
WO 2012/083416 PCT/CA2011/001266
entered. Once processor 102 detects the end of a sentence, processor 102
creates a display of emoticons, text equivalents of emoticons, and emoticon
names in semi-transparent, layered windows, as shown in Figure 12.
[046] One such emoticon is marked for selection, as shown, for example,
by box 1205. The display of semi-transparent, layered windows includes window
1210, which displays emoticons, window 1215, which displays the text
equivalent
of the emoticon marked for selection, and window 1220, which displays the name
of the emoticon marked for selection. The user uses a scroll input to change
the
emoticon marked for selection, or uses a window scroll input to make window
1210 move behind window 1220, causing window 1215 to be displayed on top.
When window 1215 is displayed on top, a plurality of text equivalents for
emoticons are displayed, with the name and the emoticon corresponding to the
text equivalent marked for selection visible through window 1215, as shown in
Figure 13.
[047] Returning to Figure 12, processor 102 next receives a selection
input from the user. After receiving the selection input, processor 102
outputs the
selected emoticon at the position of cursor 705. If the text equivalents in
window
1215 are displayed as the top window, as shown in Figure 13, selection of a
text
equivalent still causes processor 102 to output the corresponding emoticon at
the
position of cursor 705 instead of the text equivalent. In other example
embodiments, selection of a text equivalent will cause processor 102 to output
the
text equivalent at the position of cursor 705.
[048] Figures 14 to 16 show example output of an improved portable
electronic device used to display text disambiguation options, consistent with
disclosed example embodiments. In portable electronic devices using a reduced
keyboard, each input member is assigned to a number of characters, usually two
or three characters. When such an input member is actuated, the result is an
ambiguous input because it is not immediately clear what character the user
intends to output. Inputting text on such a keyboard commonly uses one of two
methods. In the first method a user uses a multi-tap method to enter the
character desired by pressing an input member in rapid succession one time for
the first character, two times for the second, three times for the third and
so on.

CA 02820744 2013-06-07
WO 2012/083416
PCT/CA2011/001266
[049] The second method is for the handheld device to predict the word
desired using stored data on common words and the frequency with which the
words occur in the language. Using a disambiguation system, for example, as a
user types processor 102 determines the possible letter combinations, or
permutations, for the characters represented by the input members actuated.
Processor 102 then compares these combinations to language objects, such as
words, common character combinations (n-grams), and the frequencies with
which these language objects occur in the language stored in static and custom
dictionaries. Such language objects and frequency objects are stored, for
example, in memory 110.
[050] Processor 102 displays the most probable character combinations,
or permutations, in a list on the screen as the user types. In some
circumstances,
a user may need to select a specific combination of characters, also called a
prefix, locking the prefix. When a prefix is locked, further combination of
characters represented by actuations of additional input members are added to
the prefix. Further description of how a disambiguation function works is
found in
U.S. Patent Application No. 11/098,783, the specification of which is
incorporated
herein by reference.
[051] The various permutations of the characters assigned to input
members are displayed in semi-transparent, layered windows, as shown in
Figures 14-16. For example, if a user actuates input members labeled "AS" and
"OP", the input is ambiguous and represented by four permutations of
characters:
"SO," "SP," "AP," and "AO." Processor 102 generates these permutations based
on the characters assigned to the input member and the order in which the
input
members are actuated.
[052] Processor 102 presents these permutations to the user in
semi-transparent, layered windows 1405-1420, as shown in Figure 14. "SO"
occupies the top window because it has the highest frequency object of the
four
permutations, because it represents a full word, or because it is the most
probable
permutation given one or more words preceding the current input. The user can
see "SP" through window 1405, and "AP" through windows 1405 and 1410. The
user can scroll the windows using any input members traditionally used to
navigate a list of permutations for disambiguation. Furthermore, the user is
able
16

CA 02820744 2013-06-07
WO 2012/083416
PCT/CA2011/001266
to select the permutation shown in the top window, thus locking the prefix for
further disambiguation.
[053] Next, if the user actuates the "OP" input member again and then the
input member for "L," the character permutations include "APPL," "APOL,"
"SPOL," "AOOL," "AOPL," "SOPL", and "SOOL." Processor 102 displays these
permutations on display 112 as seven semi-transparent, layered windows, as
shown in Figure 15. "APPL" appears in the top layer because it is associated
with
a language object with the highest frequency or because it is the most
probable
permutation given one or more preceding words. If the user enters a scrolling
input, for example actuating a down arrow or using a track wheel, then "APOL"
is
displayed on in the top window, as shown in Figure 16. In response, processor
102 changes default portion 1425 to reflect the new permutation.
[054] Those of skill in the art of disambiguation will realize that this
method
of displaying permutations is used for as many permutations as needed, and
continues until the user selects a word or enters a delimiter. The overlapping
nature of the windows allows the portable handheld device to display as many
permutations as needed without using much screen space. The transparency of
the windows allows the user to anticipate the next selection, reducing the
chances
that the user will scroll past the desired permutation.
[055] The present disclosure may be embodied in other specific forms
without departing from its spirit or essential characteristics. Other
embodiments of
the invention will be apparent to those skilled in the art from consideration
of the
specification and practice of the invention disclosed herein. The described
embodiments are to be considered in all respects only as illustrative and not
restrictive, with the true scope and spirit of the invention being indicated
by the
following claims rather than by the foregoing description. All changes that
come
within the meaning and range of equivalency of the claims are to be embraced
within their scope.
17

Representative Drawing
A single figure which represents the drawing illustrating the invention.
Administrative Status

2024-08-01:As part of the Next Generation Patents (NGP) transition, the Canadian Patents Database (CPD) now contains a more detailed Event History, which replicates the Event Log of our new back-office solution.

Please note that "Inactive:" events refers to events no longer in use in our new back-office solution.

For a clearer understanding of the status of the application/patent presented on this page, the site Disclaimer , as well as the definitions for Patent , Event History , Maintenance Fee  and Payment History  should be consulted.

Event History

Description Date
Inactive: IPC expired 2022-01-01
Inactive: IPC assigned 2020-08-19
Inactive: IPC assigned 2020-08-19
Inactive: COVID 19 - Deadline extended 2020-03-29
Inactive: IPC expired 2020-01-01
Inactive: IPC removed 2019-12-31
Inactive: Dead - No reply to s.30(2) Rules requisition 2019-04-17
Application Not Reinstated by Deadline 2019-04-17
Deemed Abandoned - Failure to Respond to Maintenance Fee Notice 2018-11-16
Inactive: Abandoned - No reply to s.30(2) Rules requisition 2018-04-17
Inactive: S.30(2) Rules - Examiner requisition 2017-10-17
Inactive: Report - No QC 2017-10-13
Letter Sent 2017-06-07
Inactive: Single transfer 2017-05-26
Amendment Received - Voluntary Amendment 2017-05-11
Inactive: S.30(2) Rules - Examiner requisition 2016-12-09
Inactive: Report - No QC 2016-12-09
Amendment Received - Voluntary Amendment 2016-05-02
Inactive: S.30(2) Rules - Examiner requisition 2015-10-30
Inactive: Report - QC passed 2015-10-26
Inactive: Agents merged 2015-05-14
Amendment Received - Voluntary Amendment 2015-01-30
Inactive: S.30(2) Rules - Examiner requisition 2014-07-31
Inactive: Report - No QC 2014-07-31
Maintenance Request Received 2013-11-18
Inactive: Cover page published 2013-09-16
Inactive: IPC assigned 2013-07-18
Inactive: IPC assigned 2013-07-18
Application Received - PCT 2013-07-18
Inactive: First IPC assigned 2013-07-18
Letter Sent 2013-07-18
Inactive: Acknowledgment of national entry - RFE 2013-07-18
Inactive: IPC assigned 2013-07-18
Inactive: IPC assigned 2013-07-18
Inactive: IPC assigned 2013-07-18
National Entry Requirements Determined Compliant 2013-06-07
Request for Examination Requirements Determined Compliant 2013-06-07
All Requirements for Examination Determined Compliant 2013-06-07
Application Published (Open to Public Inspection) 2012-06-28

Abandonment History

Abandonment Date Reason Reinstatement Date
2018-11-16

Maintenance Fee

The last payment was received on 2017-11-01

Note : If the full payment has not been received on or before the date indicated, a further fee may be required which may be one of the following

  • the reinstatement fee;
  • the late payment fee; or
  • additional fee to reverse deemed expiry.

Please refer to the CIPO Patent Fees web page to see all current fee amounts.

Fee History

Fee Type Anniversary Year Due Date Paid Date
Request for exam. (CIPO ISR) – standard 2013-06-07
Basic national fee - standard 2013-06-07
MF (application, 2nd anniv.) - standard 02 2013-11-18 2013-11-18
MF (application, 3rd anniv.) - standard 03 2014-11-17 2014-11-04
MF (application, 4th anniv.) - standard 04 2015-11-16 2015-11-03
MF (application, 5th anniv.) - standard 05 2016-11-16 2016-11-01
Registration of a document 2017-05-26
MF (application, 6th anniv.) - standard 06 2017-11-16 2017-11-01
Owners on Record

Note: Records showing the ownership history in alphabetical order.

Current Owners on Record
BLACKBERRY LIMITED
Past Owners on Record
GAELLE CHRISTINE MARTIN-COCHER
SHERRYL LEE LORRAINE SCOTT
Past Owners that do not appear in the "Owners on Record" listing will appear in other documentation within the application.
Documents

To view selected files, please enter reCAPTCHA code :



To view images, click a link in the Document Description column. To download the documents, select one or more checkboxes in the first column and then click the "Download Selected in PDF format (Zip Archive)" or the "Download Selected as Single PDF" button.

List of published and non-published patent-specific documents on the CPD .

If you have any difficulty accessing content, you can call the Client Service Centre at 1-866-997-1936 or send them an e-mail at CIPO Client Service Centre.


Document
Description 
Date
(yyyy-mm-dd) 
Number of pages   Size of Image (KB) 
Cover Page 2013-09-16 1 50
Description 2013-06-07 17 973
Drawings 2013-06-07 12 243
Claims 2013-06-07 5 165
Representative drawing 2013-06-07 1 17
Abstract 2013-06-07 1 74
Description 2015-01-30 17 895
Claims 2015-01-30 4 98
Claims 2016-05-02 4 121
Acknowledgement of Request for Examination 2013-07-18 1 176
Reminder of maintenance fee due 2013-07-18 1 112
Notice of National Entry 2013-07-18 1 202
Courtesy - Abandonment Letter (Maintenance Fee) 2018-12-28 1 174
Courtesy - Certificate of registration (related document(s)) 2017-06-07 1 102
Courtesy - Abandonment Letter (R30(2)) 2018-05-29 1 164
PCT 2013-06-07 12 539
Fees 2013-11-18 1 31
Examiner Requisition 2015-10-30 4 275
Amendment / response to report 2016-05-02 8 328
Examiner Requisition 2016-12-09 3 206
Amendment / response to report 2017-05-11 3 124
Examiner Requisition 2017-10-17 4 247