Language selection

Search

Patent 2757527 Summary

Third-party information liability

Some of the information on this Web page has been provided by external sources. The Government of Canada is not responsible for the accuracy, reliability or currency of the information supplied by external sources. Users wishing to rely upon this information should consult directly with the source of the information. Content provided by external sources is not subject to official languages, privacy and accessibility requirements.

Claims and Abstract availability

Any discrepancies in the text and image of the Claims and Abstract are due to differing posting times. Text of the Claims and Abstract are posted:

  • At the time the application is open to public inspection;
  • At the time of issue of the patent (grant).
(12) Patent Application: (11) CA 2757527
(54) English Title: MOTION GESTURES INTERFACE FOR PORTABLE ELECTRONIC DEVICE
(54) French Title: INTERFACE DE GESTES DE MOUVEMENT DESTINEE A UN DISPOSITIF ELECTRONIQUE PORTATIF
Status: Dead
Bibliographic Data
(51) International Patent Classification (IPC):
  • G06F 3/01 (2006.01)
  • G06F 3/0487 (2013.01)
  • H04W 88/02 (2009.01)
  • G06F 15/02 (2006.01)
(72) Inventors :
  • LUCIC, DRAZEN (Canada)
  • KEIP, PAUL (Canada)
(73) Owners :
  • BLACKBERRY LIMITED (Canada)
(71) Applicants :
  • RESEARCH IN MOTION LIMITED (Canada)
(74) Agent: RIDOUT & MAYBEE LLP
(74) Associate agent:
(45) Issued:
(86) PCT Filing Date: 2010-11-12
(87) Open to Public Inspection: 2012-05-12
Examination requested: 2011-11-14
Availability of licence: N/A
(25) Language of filing: English

Patent Cooperation Treaty (PCT): Yes
(86) PCT Filing Number: PCT/CA2010/001760
(87) International Publication Number: WO2012/061917
(85) National Entry: 2011-11-14

(30) Application Priority Data: None

Abstracts

English Abstract




The present disclosure provides a method of interacting with a portable
electronic device and a portable electronic device configured to perform the
same.
In accordance with one embodiment, the method comprises: detecting motion of
the portable electronic device; determining whether detected motion matches a
first motion gesture or a second motion gesture; when the first motion gesture
is
detected, showing a designated user interface element in a user interface
screen
displayed on a touch-sensitive display of the portable electronic device; and
when
the second motion gesture is detected, hiding the designated user interface
element from the user interface screen displayed on the touch-sensitive
display of
the portable electronic device.


Claims

Note: Claims are shown in the official language in which they were submitted.




CLAIMS:

1. A method of interacting with a portable electronic device, the method
comprising:

detecting motion of the portable electronic device;

determining whether detected motion matches a first motion gesture or
second motion gesture;

when the first motion gesture is detected, showing a designated user
interface element in a user interface screen displayed on a touch-sensitive
display
of the portable electronic device; and

when the second motion gesture is detected, hiding the designated user
interface element from the user interface screen displayed on the touch-
sensitive
display of the portable electronic device.

2. The method of claim 1, wherein the first motion gesture and second motion
gesture are directional motion gestures having a primary direction of motion,
wherein the primary direction of motion of the first motion gesture and second

motion are oriented in generally opposite directions to each other.

3. The method of claim 1 or claim 2, wherein the second motion gesture is a
reversed motion sequence of the first motion gesture.

4. The method of any one of claims 1 to 3, wherein the first motion gesture
comprises an up-down motion and the second motion gesture comprises a down-up
motion.

5. The method of any one of claims 1 to 3, wherein the first motion gesture
comprises a down-up motion and the second motion gesture comprises an up-down
motion.


34



6. The method of any one of claims 1 to 3, wherein the first motion comprises
a
left-right motion and the second motion gesture comprises a right-left motion.

7. The method of any one of claims 1 to 6, wherein the designated user
interface element is a virtual keyboard.

8. The method of any one of claims 1 to 6, wherein the designated user
interface element is a context-sensitive menu.

9. The method of claim 1, wherein the first motion gesture is a left-right
gesture and the second motion gesture is a right-left gesture, and the
designated
user interface element is a context-sensitive menu.

10. The method of claim 1, wherein the first motion gesture is an up-down
flick
gesture and the second motion gesture is a down-up flick gesture, and the
designated user interface element is a virtual keyboard.

11. The method of claim 10, further comprising:

determining whether detected motion matches a left-right gesture or right-
left gesture;

when the left-right gesture is detected, showing a context-sensitive menu in
a user interface screen displayed on a touch-sensitive display of the portable

electronic device; and

when the right-left gesture is detected, hiding the context-sensitive menu in
the user interface screen displayed on the touch-sensitive display of the
portable
electronic device.

12. The method of any one of claims 1 to 11, wherein the designated user
interface element is shown when not already displayed on the touch-sensitive
display of the portable electronic device, and the designated user interface
element
is hidden when displayed on the touch-sensitive display of the portable
electronic
device.





13. A method of interacting with a portable electronic device, the method
comprising:

detecting motion of the portable electronic device;

determining whether detected motion matches known motion gestures;
when a toss gesture is detected, sending an electronic message under
composition to at least one address specified by the electronic message under
composition;

when a left-right gesture is detected, displaying a next electronic message in

an inbox or message list of an electronic messaging application; and

when a right-left gesture is detected, displaying a previous electronic
message in an inbox or message list of the electronic messaging application.
14. A method of interacting with a portable electronic device, the method
comprising:

detecting motion of the portable electronic device;

determining whether detected motion matches known motion gestures;
when a toss gesture is detected, sending a data object to a second electronic
device using a short-range communication protocol;

when a left-right gesture is detected, reproducing content of a next data
object in a datastore of a media player application; and

when a right-left gesture is detected, reproducing content of a previous next
data object in a datastore of a media player application.

15. A portable electronic device, comprising:
a housing;

a processor received within the housing;

36



a touch-sensitive display coupled to the processor and having a touch-
sensitive overlay exposed by the housing; and

an accelerometer display coupled to the processor;

wherein the processor is configured for performing the method of any one of
claims 1 to 14.


37

Description

Note: Descriptions are shown in the official language in which they were submitted.



CA 02757527 2011-11-14
T

METHOD OF INTERACTING WITH A PORTABLE ELECTRONIC DEVICE
TECHNICAL FIELD

[0001] The present disclosure relates to portable electronic devices,
including
but not limited to portable electronic devices having touch screen displays
and their
control.

BACKGROUND
[0002] Electronic devices, including portable electronic devices, have gained
widespread use and may provide a variety of functions including, for example,
telephonic, electronic messaging and other personal information manager (PIM)
application functions. Portable electronic devices include, for example,
several
types of mobile stations such as simple cellular telephones, smart telephones,
wireless personal digital assistants (PDAs), and laptop computers with
wireless
802.11 or BluetoothTM capabilities.

[0003] Portable electronic devices such as PDAs or smart telephones are
generally intended for handheld use and ease of portability. Smaller devices
are
generally desirable for portability. A touch-sensitive display, also known as
a
touchscreen display, is particularly useful on handheld devices, which are
small and
have limited space for user input and output. As new functions and
capabilities are
added to portable electronic devices, the number of onscreen elements provided
by
such devices increases. Accordingly, improvements in controlling portable
electronic devices which accommodate the demand for screen space on touch-
sensitive displays are desirable.

BRIEF DESCRIPTION OF THE DRAWINGS

[0004] Figure 1 is a simplified block diagram of components including internal
components of a portable electronic device to which embodiments of the current
disclosure may be applied;

1


CA 02757527 2011-11-14

[0005] Figure 2 is a perspective view of an example of a portable electronic
device to which embodiments of the current disclosure may be applied;

[0006] Figures 3A and 3B are front views of a portable electronic device
illustrating example user interface screens with which embodiments of present
disclosure may be applied;

[0007] Figure 4 is a front view of the portable electronic device of Figures
3A
and 3B with a direction of movement shown by a block arrow with corresponding
acceleration-time graphs for the movement;

[0008] Figure 5A is a front view of the portable electronic device of Figures
3A
and 3B with a direction of an upward flick gesture shown by a block arrow;

[0009] Figure 5B is a front view of the portable electronic device of Figures
3A
and 3B with a direction of a downward flick gesture shown by a block arrow;
[0010] Figure 5C is a front view of the portable electronic device of Figures
3A
and 3B with a direction of a toss movement shown by a block arrow with
corresponding acceleration-time graphs for the movement;

[0011] Figure 5D is a front view of the portable electronic device of Figures
3A
and 3B with a direction of a left-right cycle gesture shown by a block arrow;
[0012] Figure 5E is a front view of the portable electronic device of Figures
3A
and 3B with a direction of a right-left cycle gesture shown by a block arrow;

[0013] Figure 5F is an acceleration-time graph for a pair of shake gestures of
the portable electronic device of Figures 3A;

[0014] Figure 5G is an acceleration-time graph for a repeated shaking gesture
of the portable electronic device of Figures 3A along the x-axis;

2


CA 02757527 2011-11-14

[0015] Figure 6 is a flowchart illustrating a method of interacting with a
portable electronic device using a touch-sensitive display in accordance with
one
example embodiment of the present disclosure;

[0016] Figure 7 is a flowchart illustrating a method of interacting with a
portable electronic device using a touch-sensitive display in accordance with
another example embodiment of the present disclosure; and

[0017] Figure 8 is a flowchart illustrating a method of interacting with a
portable electronic device using a touch-sensitive display in accordance with
a
further example embodiment of the present disclosure.

DETAILED DESCRIPTION OF EXAMPLE EMBODIMENTS

[0018] The present disclosure provides methods of interacting with a portable
mobile device using designated motion gestures to control the content
displayed on
a touch-sensitive display, to control actions performed by the portable mobile
device, or both. In one example, a pair of opposite motion gestures is used to
show and hide a designated user interface element such as a virtual keyboard
in a
user interface screen displayed on the touch-sensitive display, thereby
obviating
the need to press a mechanical key or touch the touch-sensitive display to
show or
hide the designated user interface element.

[0019] In accordance with one embodiment of the present disclosure, there is
provided a method of interacting with a portable electronic device, the method
comprising: detecting motion of the portable electronic device; determining
whether detected motion matches a first motion gesture or a second motion
gesture; when the first motion gesture is detected, showing a designated user
interface element in a user interface screen displayed on a touch-sensitive
display
of the portable electronic device; and when the second motion gesture is
detected,
hiding the designated user interface element from the user interface screen
displayed on the touch-sensitive display of the portable electronic device.

3


CA 02757527 2011-11-14

[0020] In accordance with another embodiment of the present disclosure,
there is provided a method of interacting with a portable electronic device,
the
method comprising: detecting motion of the portable electronic device;
determining
whether detected motion matches a first motion gesture or a second motion
gesture; when the first motion gesture is detected, showing a designated user
interface element in a user interface screen displayed on a touch-sensitive
display
of the portable electronic device; and when the second motion gesture is
detected,
showing a second designated user interface element in a user interface screen
displayed on the touch-sensitive display of the portable electronic device.

[0021] In accordance with a further embodiment of the present disclosure,
there is provided a method of interacting with a portable electronic device,
the
method comprising: detecting motion of the portable electronic device;
determining
whether detected motion matches known motion gestures; when a toss gesture is
detected, sending an electronic message under composition to at least one
address
specified by the electronic message under composition; when a left-right
gesture is
detected, displaying a next electronic message in an inbox or message list of
an
electronic messaging application; and when a right-left gesture is detected,
displaying a previous electronic message in an inbox or message list of the
electronic messaging application.

[0022] In accordance with a further embodiment of the present disclosure,
there is provided a method of interacting with a portable electronic device,
the
method comprising: detecting motion of the portable electronic device;
determining
whether detected motion matches known motion gestures; when a toss gesture is
detected, sending a data object to a second electronic device using a short-
range
communication protocol; when a left-right gesture is detected, reproducing
content
of a next data object in a datastore of a media player application; and when a
right-
left gesture is detected, reproducing content of a previous next data object
in a
datastore of a media player application.

[0023] In accordance with a further embodiment of the present disclosure,
there is provided a portable electronic device, comprising: a housing; a
processor
4


CA 02757527 2011-11-14

received within the housing; a touch-sensitive display coupled to the
processor and
having a touch-sensitive overlay exposed by the housing; and an accelerometer
coupled to the processor, wherein the processor is configured to perform the
methods described herein.

[0024] In accordance with a further embodiment of the present disclosure,
there is provided a portable electronic device, comprising: a housing; a
processor
received within the housing; a touch-sensitive display coupled to the
processor and
having a touch-sensitive overlay exposed by the housing; and an accelerometer
coupled to the processor; wherein the processor is configured for: detecting
motion
of the portable electronic device; determining whether detected motion matches
a
first motion gesture or second motion gesture; when the first motion gesture
is
detected, causing a designated user interface element to be shown in a user
interface screen displayed on a touch-sensitive display of the portable
electronic
device; and when the second motion gesture is detected, causing the designated
user interface element to be hidden from the user interface screen displayed
on the
touch-sensitive display of the portable electronic device.

[0025] The present disclosure generally relates to portable electronic devices
which may be carried in a user's hands (i.e., handheld electronic devices).
Examples of portable electronic devices include, but are not limited to,
pagers,
mobile phones, smartphones, wireless organizers, PDAs, portable media players,
portable gaming devices, Global Positioning System (GPS) navigation devices,
electronic book readers, cameras, and notebook and tablet computers.
Embodiments of the present disclosure may be applied to other portable
electronic
devices not specifically described in the above examples.

[0026] Reference will now be made to the accompanying drawings which
show, by way of example, embodiments of the present disclosure. For simplicity
and clarity of illustration, reference numerals may be repeated among the
Figures
to indicate corresponding or analogous elements. Numerous details are set
forth to
provide an understanding of the embodiments described herein. The embodiments
may be practiced without these details. In other instances, well-known
methods,
5


CA 02757527 2011-11-14

procedures, and components have not been described in detail to avoid
obscuring
the embodiments described. The description is not to be considered as limited
to
the scope of the embodiments described herein.

[0027] Reference is made to FIG. 1, which illustrates in block diagram form, a
portable electronic device 100 to which example embodiments described in the
present disclosure can be applied. The portable electronic device 100 includes
multiple components, such as a processor 102 that controls the overall
operation of
the portable electronic device 100. Communication functions, including data
and
voice communications, are performed through a communication subsystem 104.
Data received by the portable electronic device 100 is decompressed and
decrypted
by a decoder 106. The communication subsystem 104 receives messages from and
sends messages to a wireless network 150. The wireless network 150 may be any
type of wireless network, including, but not limited to, data wireless
networks, voice
wireless networks, and networks that support both voice and data
communications.
A power source 142, such as one or more rechargeable batteries or a port to an
external power supply, powers the portable electronic device 100.

[0028] The processor 102 interacts with other components, such as Random
Access Memory (RAM) 108, memory 110, a display 112 (such as a liquid crystal
display (LCD)) with a touch-sensitive overlay 114 coupled to an electronic
controller
116 that together comprise a touch-sensitive display 118, one or keys or
buttons
120, a navigation device 122, one or more auxiliary input/output (I/O)
subsystems
124, a data port 126, a speaker 128, a microphone 130, short-range
communications subsystem 132, and other device subsystems 134. It will be
appreciated that the electronic controller 116 of the touch-sensitive display
118
need not be physically integrated with the touch-sensitive overlay 114 and
display
112. User-interaction with a graphical user interface (GUI) is performed
through
the touch-sensitive overlay 114. The GUI displays user interface screens on
the
touch-sensitive display 118 for displaying information or providing a touch-
sensitive
onscreen user interface element for receiving input. This content of the user
interface screen varies depending on the device state and active application,
among
other factors. Some user interface screens may include a text field sometimes
6


CA 02757527 2011-11-14

called a text input field. The processor 102 interacts with the touch-
sensitive
overlay 114 via the electronic controller 116. Information, such as text,
characters,
symbols, images, icons, and other items that may be displayed or rendered on a
portable electronic device, is displayed on the touch-sensitive display 118
via the
processor 102.

[0029] The portable electronic device 100 also comprises a motion detection
subsystem 140 comprising at least one sensor which is coupled to the processor
102 and which is controlled by one or a combination of a monitoring circuit
and
operating software. The sensor has a sensing element which detects
acceleration
from motion and/or gravity. The sensor generates and outputs an electrical
signal
representative of the detected acceleration. Changes in movement of the
portable
electronic device 100 results in changes in acceleration which produce
corresponding changes in the electrical signal output of the sensor. In at
least
some embodiments, the sensor is an accelerometer 136 such as a three-axis
accelerometer having three mutual orthogonally sensing axes. The accelerometer
136 detects changes in the acceleration of the portable electronic device 100.
Other
types of motion sensors may be used by the motion detection subsystem 140 in
addition to, or instead of, the accelerometer 136 in other embodiments. The
other
motion sensors may comprise a proximity sensor, gyroscope, or both, which
detect
changes in the proximity and orientation of portable electronic device 100.
[0030] Changes in acceleration, proximity and orientation detected by the
accelerometer 136, proximity sensor and/or gyroscope may be interpreted by the
portable electronic device 100 as motion of the portable electronic device
100.
When the changes in acceleration, proximity and orientation are within
threshold
tolerance(s) of regularity or predictability, when the changes in
acceleration,
proximity and orientation match predetermined motion criteria (e.g., stored in
the
memory 110), the changes may be interpreted by the portable electronic device
100 as a pattern of motion. Multiple patterns of motion may be recognized by
the
portable electronic device 100.

7


CA 02757527 2011-11-14

[0031] Referring to Figure 4, an example accelerometer response to a
movement of the portable electronic device 100 in the y-direction from rest
followed by a stopping'of the movement will be shown. The direction of
movement
is shown by a block arrow. Corresponding acceleration-time graphs for the
movement illustrate example acceleration signals which may be generated by the
accelerometer 136 (or motion detection subsystem 140) in response to the
movement (or motion sequence). For this motion sequence, the acceleration in
the
x-direction 410 sensed by the accelerometer 136 stays fairly constant at
approximately zero (0) G, while the acceleration in the y-direction 420
increases as
the portable electronic device 100 starts moving, and then turns negative as
the
device is brought to a stop. The motion pattern in the signal will be affected
by the
speed and force with which a user performs a particular motion sequence.

[0032] By configuring the processor 102 to recognize certain motion patterns
in the acceleration signal from the accelerometer 136, the processor 102 can
determine whether the portable electronic device 100 has been moved in a
certain
motion sequence. Predetermined motion sequences recognized by the processor
102 in accordance with a designated pattern of motion will herein be referred
to as
motion gestures. Motion gestures performed by the user may cause acceleration
in
one or more sensing axes and in one or more directions.

[0033] Figures 5A to 5E illustrate, by way of example, a number of motion
gestures which may be detected by the portable electronic device 100. Figure
5A
shows a first flick gesture in which the portable electronic device 100 is
moved in
the positive y-direction and then back in the negative y-direction. Figure 5B
shows
a second flick gesture in which the portable electronic device 100 is moved in
the
negative y-direction and then in the positive y-direction. The second flick
gesture is
reverse flick gesture which is a reversed motion sequence of the first flick
gesture.
[0034] Figure 5C shows a toss gesture in which the portable electronic device
100 is rotated clockwise about an axis of rotation 530. The toss gesture is
similar to
the motion used to throw a flying disc such as a Frisbee . The angle of
rotation 8
and distance between the accelerometer 136 and the axis of rotation 530 may
8


CA 02757527 2011-11-14

affect the acceleration signal generated. In some embodiments, a toss gesture
can
be based on the acceleration signals generated when the angle of rotation 0 is
around 90 degrees, and the distance between the accelerometer 136 and the axis
of rotation can be estimated by assuming the axis of rotation is located about
the
wrist joint of an average user. In other embodiments, the processor 102 can be
configured to recognize a toss gesture to have occurred based on the
acceleration
signals generated for any predetermined angle of rotation 0, direction of
rotation,
or axis of rotation 530.

[0035] The toss gesture shown in Figure 5C is sometimes referred to as a toss
"away" gesture since the gesture starts with the portable electronic device
100 held
towards the user and moves away from the user. A toss "towards" gesture is
related, but opposite to, a toss "away" gesture. The toss "towards" gesture
starts
with the portable electronic device 100 held away from the user and moves
towards
from the user. The acceleration-time graph for a toss "towards" gesture would
be
similar to the acceleration-time graph for a toss "away" gesture with the
curve for
the x-axis inverted and the curve for the y-axis the same.

[0036] Figure 5D shows a left-right cycle gesture wherein the portable
electronic device 100 is moved from left to right in the positive x-direction.
Figure
5E shows a right-left cycle gesture wherein the portable electronic device 100
is
moved from right to left in the negative x-direction.

[0037] Figure 5F shows an acceleration-time graph for a pair of shake
gestures which may be detected by the portable electronic device 100. Figure
5G
shows an acceleration-time graph for a repeated shaking gesture of the
portable
electronic device of Figures 3A along the x-axis. The acceleration in Figures
5F and
5G is shown in Gal over a time duration measured in seconds using each of the
three sensing axes (i.e., x, y and z axes) of a three-axis accelerometer. The
z-axis
in Figure 5F is calibrated for a steady-state reading of -1 g (-1000 Gal)
whereas the
z-axis in Figure 5G is calibrated for a steady-state reading of +1 g (1000
Gal),
otherwise the acceleration-time graphs are comparable in terms of device
characteristics.

9


CA 02757527 2011-11-14

[0038] The shaking shown in Figure 5G is characterized by alternating
increase and decreases in acceleration. At the start of the acceleration the
portable electronic device 100 was substantially still representing a period
of
relative stability. Because the acceleration of Figure 5G represents a lateral
shaking motion of the portable electronic device 100 along the x-axis, the
acceleration from the y-axis and z-axis is relatively stable. The acceleration
also
illustrates that the z-axis was substantially parallel to gravity during the
shaking
movement as it experiences a force of acceleration of approximately 980 Gal
(9.8
m/s2).

[0039] The shaking movement illustrated in Figure 5G is characterized by
acceleration on the x-axis which alternates between positive acceleration
spikes and
negative acceleration spikes. In the positive acceleration spikes, the
accelerometer
acceleration along the x-axis increases from a general baseline measurement in
the
stable period prior to the shaking movement. Similarly, in the negative
acceleration (e.g. deceleration) spikes, the acceleration along the x-axis
decreases
from the baseline in the stable period prior to the shaking movement. In the
example shown, prior to and during the shaking movement, the x-axis is
generally
perpendicular to the earth's gravitational force. In this orientation, the
acceleration
on the x-axis is approximately zero Gal when the portable electronic device
100 is
not since force of gravity acting on the y and z axes in this position is
approximately zero. Accordingly, in the shown example shown, the positive
acceleration periods may be defined as the periods in which the accelerometer
acceleration on the x-axis is greater than the baseline when the device 100
was
not moving, and the negative acceleration periods may be defined as the
periods in
which the acceleration on the x-axis is less than baseline when the device 100
was
not moving.

[0040] The motion gestures described above have been described by way of
example and not intended to be limiting unless explicitly stated otherwise
herein.
The processor 102 may be configured to determine when any motion gesture. In
some embodiments, the portable electronic device 100 may provide a gesture
defining mode which allows users to configure the processor 102 to recognize a


CA 02757527 2011-11-14

user defined gestures. In the gesture defining mode, a user may perform a
gesture
a predetermined number of times. The processor 102 then stores the associated
motion patterns and/or predetermined motion criteria in memory 110 for
detecting
the user defined gestures. The motion patterns and/or predetermined motion
criteria may then be mapped to user interface changes and/or commands or
actions
performed by the portable electronic device 100, for example, using a
configuration
menu provided in the gesture defining mode. When the user interface changes
and/or commands or actions are supported by the active application 148 or
operating system 146 in a device state, performing the user defined gestures
will
cause the portable electronic device 100 to perform the user interface changes
and/or commands or actions associated (e.g., mapped) to those user defined
gestures.

[0041] As will also be appreciated by persons skilled in the art,
accelerometers may produce digital or analog output signals. Generally, two
types
of outputs are available depending on whether an analog or digital
accelerometer is
used: (1) an analog output requiring buffering and analog-to-digital (A/D)
conversion; and (2) a digital output which is typically available in an
industry
standard interface such as an SPI (Serial Peripheral Interface) or 12C (Inter-
Integrated Circuit) interface. When the accelerometer is analog, the memory
110
includes machine-readable instructions for calculating acceleration based on
electrical output input from the accelerometer 136. The processor 102 executes
the
machine-readable instructions to calculate acceleration which may be used by
the
operating system 146 and/or applications 148 as input. Depending on the
acceleration input, the operating system 146 and/or applications 148 may
perform
operations causing changes to the state of the portable electronic device 100,
including but not limited to a change in the operational state or a change in
the
content displayed on the display screen 112.

[0042] The output of the accelerometer 136 is typically measured in terms of
the gravitational acceleration constant at the Earth's surface, denoted g,
which is
approximately 9.81 m/s2 (32.2 ft/s2) as the standard average, or in terms of
units
Gal (cm/s2). The accelerometer 136 may be of almost any type including, but
not
11


CA 02757527 2011-11-14

limited to, a capacitive, piezoelectric, piezoresistive, or gas-based
accelerometer.
The range of accelerometers vary up to the thousands of g's, however for
portable
electronic devices "low-g" accelerometers may be used. Example low-g
accelerometers which may be used are MEMS digital accelerometers from Analog
Devices, Inc. (ADI), Freescale Semiconductor, Inc. (Freescale) and
STMicroelectronics N.V. of Geneva, Switzerland. Example low-g MEMS
accelerometers are model LIS331DL, LIS3021DL and LIS3344AL accelerometers
from STMicroelectronics N.V. The LIS3344AL model is an analog accelerometer
with an output data rate of up to 2 kHz which has been shown to have good
response characteristics in analog sensor based motion detection subsystems.
[0043] The auxiliary I/O subsystems 124 could include other input devices
such as one or more control keys, a keyboard or keypad, navigational tool
(input
device), or both. The navigational tool may be a depressible (or clickable)
joystick
such as a depressible optical joystick, a depressible trackball, a depressible
scroll
wheel, or a depressible touch-sensitive trackpad or touchpad. The other input
devices could be included in addition to, or instead of, the touch-sensitive
display
118, depending on the embodiment.

[0044] To identify a subscriber for network access, the portable electronic
device 100 uses a Subscriber Identity Module or a Removable User Identity
Module
(SIM/RUIM) card 138 for communication with a network, such as the wireless
network 150. Alternatively, user identification information may be programmed
into memory 110.

[0045] The portable electronic device 100 includes an operating system 146
and software programs or components 148 that are executed by the processor 102
and are typically stored in a persistent, updatable store such as the memory
110.
Additional applications or programs may be loaded onto the portable electronic
device 100 through the wireless network 150, the auxiliary I/O subsystem 124,
the
data port 126, the short-range communications subsystem 132, or any other
suitable subsystem 134.

12


CA 02757527 2011-11-14

[0046] A received signal such as a text message, an e-mail message, or web
page download is processed by the communication subsystem 104 and input to the
processor 102. The processor 102 processes the received signal for output to
the
display 112 and/or to the auxiliary I/O subsystem 124. A subscriber may
generate
data items, for example e-mail messages, which may be transmitted over the
wireless network 150 through the communication subsystem 104. For voice
communications, the overall operation of the portable electronic device 100 is
similar. The speaker 128 outputs audible information converted from electrical
signals, and the microphone 130 converts audible information into electrical
signals
for processing.

[0047] FIG 2 shows a front perspective view of an example of a portable
electronic device 100. The portable electronic device 100 includes a housing
200
that houses internal components including internal components shown in FIG 1.
In
the embodiment shown in FIG. 1, the housing 200 is elongate having a length
greater than its width. The housing 200 has opposed top and bottom ends
designated by references 202, 204 respectively, and two left and right sides
extending transverse to the top and bottom ends 202, 204, designated by
references 206, 208 respectively. Although the housing 200 is shown as a
single
unit, it could, among other possible configurations, include two or more case
members hinged together (such as, for example, a flip-phone configuration or a
clam shell-style laptop computer). Other device configurations are also
possible.
[0048] The housing 200 also frames the touch-sensitive display 118 such that
the touch-sensitive display 118 is exposed for user-interaction therewith when
the
portable electronic device 100 is in use. It will be appreciated that the
touch-
sensitive display 118 may include any suitable number of user-selectable
features
rendered thereon, for example, in the form of virtual buttons for user-
selection of,
for example, applications, options, or keys of a keyboard for user entry of
data
during operation of the portable electronic device 100.

[0049] The touch-sensitive display 118 may be any suitable touch-sensitive
display, such as a capacitive, resistive, infrared, surface acoustic wave
(SAW)
13


CA 02757527 2011-11-14

touch-sensitive display, strain gauge, optical imaging, dispersive signal
technology,
acoustic pulse recognition, and so forth, as known in the art. A capacitive
touch-
sensitive display includes a capacitive touch-sensitive overlay 114. The
overlay 114
may be an assembly of multiple layers in a stack including, for example, a
substrate, a ground shield layer, a barrier layer, one or more capacitive
touch
sensor layers separated by a substrate or other barrier, and a cover. The
capacitive touch sensor layers may be any suitable material, such as patterned
indium tin oxide (ITO).

[0050] The buttons 120 may be separately operable buttons or may be
located below the touch-sensitive display 118 on a front face 210 of the
portable
electronic device 100. The buttons 120 generate corresponding input signals
when
activated. The buttons 120 may be constructed using any suitable button (or
key)
construction such as, for example, a dome-switch construction. The actions
performed by the portable electric device 100 in response to activation of
respective buttons 120 are context-sensitive. The action performed depends on
a
context that the button was activated. The context may be, but is not limited
to, a
device state, application, screen context, selected item or function, or any
combination thereof.

[0051] Referring now to FIG. 2, an accelerometer 136 is shown located within
portable electronic device 100. The accelerometer 136 includes three mutual
orthogonally sensing axes denoted x, y and z which are aligned with the form
factor
of the portable electronic device 100. In some embodiments, the accelerometer
136 is aligned such that a first sensing axis (e.g., the x-axis) extends
longitudinally
between left and right sides 206, 208 of the portable electronic device 100, a
second sensing axis (e.g., the y-axis) extends laterally between top and
bottom
ends 202, 204, and a third sensing axis (e.g., the z-axis) extends
perpendicularly
through the x-y plane defined by the x and y axes at the intersection (origin)
of
these axes. In such a configuration, when the portable electronic device 100
is
oriented horizontally, the x and y axes are parallel to the horizontal axis
and the z
axis has the force of gravity operating directly upon it. The sensing axes x,
y, z
14


CA 02757527 2011-11-14

could be aligned with different features of the portable electronic device 100
in
other embodiments.

[0052] A flowchart illustrating one example embodiment of a method 600 of
interacting with a portable electronic device using a touch-sensitive display
in
accordance with one example embodiment of the present disclosure is shown in
Figure 6. The method 600 may be carried out by software executed, for example,
by the processor 102. Coding of software for carrying out such a method 600 is
within the scope of a person of ordinary skill in the art provided the present
disclosure. The method 600 may contain additional or fewer processes than
shown
and/or described, and may be performed in a different order. Computer-readable
code executable by the processor 102 to perform the method 600 may be stored
in
a computer-readable medium such as the memory 110.

[0053] First, a user interface screen having a text input field for input text
is
displayed on the touch-sensitive display 118 of the portable electronic device
100
(602). Figures 3A and 3B show user interface screens for a Web browser
application displayed on the touch-sensitive display 118. In Figure 3A, a part
of a
webpage is displayed by the Web browser application. In the shown example, the
entire webpage does not fit within the display area of the touch-sensitive
display
118 and so a user must scroll down to see the remainder of the webpage.

[0054] Next, the portable electronic device 100 monitors for and detects
motion of the portable electronic device 100 (604). Motion is typically
detected
using the motion sensor of the motion detection subsystem 140, such as the
accelerometer 136 which uses acceleration measurements to detect motion. The
portable electronic device 100 monitors acceleration measurements reported by
the
accelerometer 136 and detects motion when acceleration matches predetermined
criteria. The motion detection subsystem 140 and/or accelerometer 136 may
generate an analog or digital acceleration signal in response to motion and
acceleration. Similar motions generate similar acceleration signal patterns.



CA 02757527 2011-11-14

[0055] Next, the portable electronic device 100 determines whether detected
motion matches a first motion gesture or a second motion gesture (decision
block
606) based on patterns of motion recognized by the portable electronic device
100.
The second motion gesture is different from the first motion gesture. The
portable
electronic device 100 has a motion analyzing unit which analyses the
acceleration
measurements in terms of factors such as amplitude/magnitude over time,
frequency, or other factors to determine whether the detected motion matches a
known motion gesture such as the first or second motion gesture.

[0056] When the first motion gesture is detected, a virtual (or soft) keyboard
320 is shown (e.g., invoked) on the user interface screen displayed on the
touch-
sensitive display 118 of the portable electronic device 100 (608). The virtual
keyboard 320 comprises a number of virtual (or soft) keys 325 as shown in
Figure
3B. Typically, this only occurs when the virtual keyboard 320 is not already
displayed on the touch-sensitive display 118. In such embodiments, when the
first
motion gesture is detected while the virtual keyboard 320 is already
displayed, the
first motion gesture is ignored. Alternatively, the virtual keyboard 320 may
be
hidden and re-shown in response to detecting the first motion gesture for GUI
effect, or a secondary function may be performed by the portable electronic
device
100 such as, for example, character input (e.g., of a special character) or
performance of a command or action.

[0057] Showing the virtual keyboard 320 on the touch-sensitive display 118
comprises rendering at least the virtual keyboard 320 and displaying the
rendered
virtual keyboard 320 on the display 112. Showing may comprise rendering the
entire user interface screen including the virtual keyboard 320 and displaying
the
rendered user interface screen on the touch-sensitive display 118. In other
embodiments, only the virtual keyboard 320 is rendered and displayed while the
remainder of the user interface screen is unchanged and is not rendered for
efficient graphics processing on the portable electronic device 100. Showing
the
virtual keyboard 320 may also comprise configuring processor 102 to recognize
touch inputs associated with the virtual keyboard 320, such as touch inputs
associated with the keys of the virtual keyboard 320.
16


CA 02757527 2011-11-14

[0058] When the second motion gesture is detected, the virtual keyboard 320
is hidden in the user interface screen displayed on the touch-sensitive
display 118
of the portable electronic device 100 (610). Typically, this occurs when the
virtual
keyboard 320 is displayed on the touch-sensitive display 118. In such
embodiments, when the second motion gesture is detected while the virtual
keyboard 320 is already hidden, the second motion gesture is ignored.
Alternatively, the virtual keyboard 320 may be shown and re-hidden in response
to
detecting the second motion gesture for GUI effect, or a secondary function
may be
performed by the portable electronic device 100 such as, for example,
character
input (e.g., of a special character) or performance of a command or action.
[0059] Hiding the virtual keyboard 320 on the touch-sensitive display 118
comprises rendering a portion of the user interface screen in the location of
the
virtual keyboard 320 and displaying the portion of the user interface screen
to be
shown when the virtual keyboard 320 is hidden. Hiding the designated user
interface element may comprise rendering the entire user interface screen
without
the virtual keyboard 320 and displaying the rendered user interface screen on
the
touch-sensitive display 118. In other embodiments, only the portion of the
user
interface screen used by the virtual keyboard 320 is rendered and displayed
while
the remainder of the user interface screen is unchanged and is not rendered
for
efficient graphics processing on the portable electronic device 100.

[0060] When the detected motion does not match the first motion gesture or
second motion gesture, the motion is ignored. Alternatively, if the detected
motion
matches another motion gesture recognized by the portable electronic device
100,
the command or action associated with that other motion gesture may be
performed, depending on the embodiment.

[0061] The availability of the virtual keyboard 320 for invocation may depend
on the presence of a text input field for input text such as an address bar
305 or
search bar 310. Typically, the availability of the virtual keyboard 320 for
invocation
depends on a text input field being active. The text input field may be made
an
active field by appropriate input including, for example, selection of the
text input
17


CA 02757527 2011-11-14

field using an onscreen position indicator. Selection with the key 325 with
the
onscreen position indicator may involve highlighting or focusing the text
input field.
Selecting the text input field may cause the appearance of the text input
field to be
changed from a first visual state to a second visual state different from the
first
visual state. Changing the appearance of the text input field may cause the
colour
to change from an initial colour (e.g. white or grey) to a different colour
(e.g.,
blue).

[0062] The virtual keyboard 320 may be a full QWERTY keyboard or a reduced
QWERTY keyboard. Each key 325 in the virtual keyboard 320 may be associated
with one or more indicia representing an alphabetic character, a numeral
character
or a command (such as a space command, return command, or the like). The
plurality of the keys having alphabetic characters may be arranged in a
standard
keyboard layout such as a QWERTY layout, a QZERTY layout, a QWERTZ layout, an
AZERTY layout, a Dvorak layout, a Russian keyboard layout, a Chinese keyboard
layout, or other suitable layout. These standard layouts are provided by way
of
example and other similar standard layouts may be used. The keyboard layout
may be based on the geographical region in which the portable electronic
device
100 is intended for use. Touching a key 325 in the virtual keyboard 320 causes
a
character associated with the key 325 to be input and displayed in a text
input field
on the touch-sensitive display 118, or causes a command or other input
associated
with the key 325 to be performed by the portable electronic device 100.

[0063] Touching a key 325 comprises touching a location of the touch-
sensitive display 118 which is coincident with the key 325 on the display 112.
A
location is coincident with the key 325 in that the centroid of the touch
event is
within an input area of the user interface screen assigned for receiving input
for
activating the key 325. The input area of the key 325 in some embodiments may
be
different than the displayed area of the key 325 on the display 112, typically
the
input area being larger than the displayed area in such embodiments to
accommodate touch offset of the user.

18


CA 02757527 2011-11-14

[0064] In at least some embodiments, the first motion gesture and second
motion gesture are directional motion gestures having a primary direction of
motion, wherein the primary direction of motion of the first motion gesture
and
second motion are oriented in generally opposite directions to each other. The
second motion gesture may be a reversed motion sequence of the first motion
gesture. The first motion gesture and second motion gesture may be, in at
least
some embodiments, flick gestures oriented in generally opposite directions to
each
other. Typically, the first motion gesture (e.g., first flick motion gesture)
comprises
a generally up-down motion and the second motion gesture (e.g., second flick
motion gesture) comprises a generally down-up motion. This mapping of motion
gestures to showing and hiding the virtual keyboard 320 provides a more
intuitive
solution in that the actions of the user for showing and hiding the virtual
keyboard
320 mimic the physical movement required to open a flip phone to expose a
physical keypad or keyboard and close the flip phone to conceal the physical
keypad or keyboard. The motion gestures are also similar to the physical
movement
required to open a slider phone to expose a physical keypad or keyboard and
close
the slider phone to conceal the physical keypad or keyboard.

[0065] In other embodiments, the first motion gesture (e.g., first flick
motion
gesture) may comprise a generally down-up motion and the second motion gesture
(e.g., second flick motion gesture) may comprise a generally up-down motion.
In
yet other embodiments, the first motion gesture and the second motion gesture
may be the same.

[0066] In other embodiments, the first motion gesture may comprise a left-
right cycle gesture and the second motion gesture may comprise a right-left
cycle
gesture. This combination of motion gestures is an alternative combination of
directional motion gestures having reverse or opposite primary direction of
motion.
This alternative combination of motion gestures could be used instead of a
flick
gesture and a reverse flick gesture to provide a pair of opposite motion
gestures is
used to show and hide a different user interface element such as a context-
sensitive menu.

19


CA 02757527 2011-11-14

[0067] The method 600 uses the first and second motion gestures to show
and hide the virtual keyboard 320 without the need to press a mechanical key
or
touch the touch-sensitive display 118 as is conventionally done. When a
mechanical key is not needed to show or hide the virtual keyboard 320, the key
can
be omitted from the portable electronic device 100 reducing costs and simplify
device design and construction. When interaction with the touch-sensitive
display
118 is not required to show or hide the virtual keyboard 320 (such as swiping
or
otherwise activating an icon or other onscreen element on a touch-sensitive
display
118), accidental activation of touch gesture commands can be avoided. The
method 600 also overcomes problems with solution which automatically display a
virtual keyboard when a text input field is in active focus. However, this
condition
is undesired in many circumstances, most notably because it presents the
possibility for a user to accidentally select a text input field bringing it
into active
focus and triggering the portable electronic device 100 to display the virtual
keyboard.

[0068] While described in the context of the virtual keyboard 320, the method
600 can be applied to a different designated user interface element such as a
context-sensitive menu associated with the operating system 146, active
application or active onscreen element. The context-sensitive menu provides a
limited set of commands or actions associated with the operating system 146,
active application or active onscreen element. For example, when viewing an
email,
the context-sensitive menu may contain commands relating to email messaging
such as reply, forward, delete, etc. Similar to when used to invoke the
virtual
keyboard 320, the method 600 may be advantageous when used to show and hide
a context-sensitive menu in that it avoids interacting with a mechanical key
or
onscreen element displayed on a touch-sensitive display 118 to trigger the
display
of the context-sensitive menu.

[0069] In some embodiments, the processor 102 may be configured to detect
different types of motion gestures to display different user interface
elements. For
example, flick gestures may be used to show and hide the virtual keyboard 320
whereas cycle gestures may be used to show and hide the context-sensitive
menu.


CA 02757527 2011-11-14

For example, a left-right cycle gesture may be used to show the context-
sensitive
menu and a right-left cycle gesture may be used to hide the context-sensitive
menu, or vice versa.

[0070] The processor 102 may be configured to detect motion only when a
predetermined condition exists. This may reduce power consumption and may
reduce inadvertent gestures, for example caused by movement while in a user's
pocket or bag, from triggering a response by the portable electronic device
100.
This may also increase the accuracy of identifying motion gestures since the
predetermined condition provides an indication that the gesture is intended if
the
predetermined condition exists. In such cases, the processor 102 needs only to
match the detected motion to available gestures in the specified context or
state of
the portable electronic device 100 rather than determining whether the motion
detected by the portable electronic device 100 is a known motion gesture. The
predetermined condition may be depression of a designated button 120 (e.g., a
press and hold of the designated button 120), depression of the depressible
optical
joystick, display of a designated user interface screen, selection of a
designated
user interface element such as a text input field, or other suitable
predetermined
condition.

[0071] A flowchart illustrating one example embodiment of a method 700 of
interacting with a portable electronic device using a touch-sensitive display
in
accordance with one example embodiment of the present disclosure is shown in
Figure 7. The method 700 may be carried out by software executed, for example,
by the processor 102. Coding of software for carrying out such a method 700 is
within the scope of a person of ordinary skill in the art provided the present
disclosure. The method 700 may contain additional or fewer processes than
shown
and/or described, and may be performed in a different order. Computer-readable
code executable by the processor 102 to perform the method 700 may be stored
in
a computer-readable medium such as the memory 110.

[0072] First, a messaging application is started and a user interface screen
for
the messaging application is displayed on the touch-sensitive display 118 of
the

21


CA 02757527 2011-11-14

portable electronic device 100, typically in response to user input (702).
From a
default user interface screen of the messaging application, such as an inbox,
the
user can navigate to other user interface screens such as a message
composition
user interface screen for composing an electronic message, or a messaging
viewing
user interface screen in which a received message is displayed on the touch-
sensitive display 118.

[0073] The messaging application may be, but is not limited to, an email
messaging application for composing and sending email messages, an SMS (Short
Message Service) messaging application for composing and sending SMS text
messages, a Multimedia Messaging Service (MMS) messaging application for
composing and sending MMS text messages, an instant messaging (IM) application
for composing and sending IM messages, a peer-to-peer or device-to-device
messaging application for composing and sending peer-to-peer messages, or a
personal information manager (PIM) for composing and sending a number of
different types of electronic messages.

[0074] Next, the portable electronic device 100 monitors for and detects
motion of the portable electronic device 100 (704). Next, the portable
electronic
device 100 determines whether detected motion matches a toss gesture, a left-
right cycle gesture, or a right-left cycle gesture (decision block 706) based
on
patterns of motion recognized by the portable electronic device 100. The
portable
electronic device 100 has motion analyzing unit which analyses the
acceleration
measurements in terms of factors such as amplitude/magnitude over time,
frequency, or other factors to determine whether detect motion matches a known
motion gesture such as the toss gesture, left-right cycle gesture and right-
left cycle
gesture. As noted above, a toss gesture comprises a rotation around an axis
normal to a plane of the portable electronic device 100 (e.g., normal to a
plane of a
surface of the touch-sensitive display 118).

[0075] When a toss gesture is detected, any electronic message under
composition is sent using the communication subsystem 104 over the wireless
network 150 when at least one address for the electronic message is defined
(708).
22


CA 02757527 2011-11-14

When at least one address for the electronic message is not defined, a prompt
to
enter at least one address for the electronic message may be provided after
which
the electronic message will be sent. A notification that electronic message
has
been sent may be displayed on the display 112 to inform the user. When an
electronic message is not under composition, the portable electronic device
100
does not monitor for toss gestures and any toss gesture which is performed is
not
detected. Alternatively, the portable electronic device 100 may monitor for
and
detect toss gestures but ignores any toss gesture when an electronic message
is
not under composition.

[0076] When the left-right cycle gesture is detected, the electronic messaging
application causes a next message in an inbox, message folder or message list
of
the electronic messaging application to be displayed (710). The next message
is
determined relative to a currently selected message, typically in
chronological order
from older to newer messages. The currently selected message may be indicated
in the inbox, message folder or message list of the electronic messaging
application
displayed on the display 112, for example, by highlighting or focusing the
message
in the inbox, message folder or message list or other suitable method of
visual
indication. Highlighting or focusing the currently selected message causes the
appearance of the corresponding message in the inbox, message folder or
message
list to be changed from a first visual state to a second visual state
different from
the first visual state. Changing the appearance of the message in the inbox,
message folder or message list, in at least some embodiments, may comprise
changing a colour of a background or field of the message entry in the inbox,
message folder or message list, the text of the message entry in the inbox,
message folder or message list, or both. The currently selected message may be
displayed on the display 112. Alternatively, the currently selected message
may
not be shown or otherwise indicated on the display 112.

[0077] When the right-left cycle gesture is detected, the electronic messaging
application causes a previous message in the inbox, message folder or message
list
of the electronic messaging application to be displayed (712). The previous

23


CA 02757527 2011-11-14

message is determined relative to a currently selected message, typically in
chronological order from older to newer messages.

[0078] When an electronic message is not selected, the portable electronic
device 100 does not monitor for left-right cycle gestures or right-left cycle
gestures
and any left-right cycle gesture or right-left cycle gesture which is
performed is not
detected is ignored. Alternatively, the portable electronic device 100 may
monitor
for and detect left-right cycle gestures and right-left cycle gestures but
ignores any
detected when an electronic message is not selected. Alternatively, the next
message or previous message may be determined based on a default message such
as the most recently received message. In some embodiments, when an electronic
message is being composed and a message composition user interface screen is
displayed on the touch-sensitive display 118 when a left-right cycle gesture
or
right-left cycle gesture is detected , the electronic message under composed
may
be automatically saved as a draft message before displaying the next message
or
previous message.

[0079] When the detected motion does not match the toss gesture, left-right
cycle gesture or right-left cycle gesture, the motion is ignored.
Alternatively, if the
detected motion matches another motion gesture recognized by the portable
electronic device 100, the command or action associated with that other motion
gesture may be performed, depending on the embodiment.

[0080] A flowchart illustrating one example embodiment of a method 800 of
interacting with a portable electronic device using a touch-sensitive display
in
accordance with one example embodiment of the present disclosure is shown in
Figure 8. The method 800 may be carried out by software executed, for example,
by the processor 102. Coding of software for carrying out such a method 800 is
within the scope of a person of ordinary skill in the art provided the present
disclosure. The method 800 may contain additional or fewer processes than
shown
and/or described, and may be performed in a different order. Computer-readable
code executable by the processor 102 to perform the method 800 may be stored
in
a computer-readable medium such as the memory 110.
24


CA 02757527 2011-11-14

[0081] First, a media player application is started and a user interface
screen
for the media player application is displayed on the touch-sensitive display
118 of
the portable electronic device 100, typically in response to user input (802).
Next,
the portable electronic device 100 monitors for and detects motion of the
portable
electronic device 100 (804). Next, the portable electronic device 100
determines
whether detected motion matches a toss gesture, a left-right cycle gesture, or
a
right-left cycle gesture (decision block 806) based on patterns of motion
recognized
by the portable electronic device 100. The portable electronic device 100 has
motion analyzing unit which analyses the acceleration measurements in terms of
factors such as amplitude/magnitude over time, frequency, or other factors to
determine whether detect motion matches a known motion gesture such as the
toss
gesture, left-right cycle gesture and right-left cycle gesture.

[0082] When a toss gesture is detected, a selected data object such as a
digital picture or graphic object, video object, or audio object (e.g., song)
is sent to
a second electronic device operably coupled to the portable electronic device
100
(808). When a destination is not defined, a prompt to enter a destination for
the
selected data object may be provided after which the selected data object will
be
sent. A data object is not currently selected, a prompt to select a data
object may
be provided after which the selected data object will be sent. The second
electronic
device may be a computer, smartphone, digital picture frame, portable media
player, portable gaming device, portable navigation device, or any other
electronic
device. For security reasons, the second electronic device is typically an
electronic
device with which the portable electronic device 100 has previously paired.
Pairing
allows the devices to connect and communicate with each other, typically
without
user intervention.

[0083] When the data object is an audio object or video object, sending the
data object may comprise streaming the audio (e.g., song/track) defined by the
audio object or streaming the video defined by the video object to the second
electronic device.



CA 02757527 2011-11-14

[0084] The portable electronic device 100 may be operably coupled to the
second electronic device using a short-range communications protocol supported
by
the short-range communications subsystem 132 including, but not limited to,
Universal Serial Bus (USB), Wi-Fi , Bluetooth , UltraWideband (UWB), an
Infrared
Data Association (IrDA), Z-Wave or ZigBee wireless network or other suitable
wireless local area network (WLAN) protocol. When the portable electronic
device
100 is not coupled to an electronic device, a prompt to connect to an
electronic
device may be provided after which the selected data object will be sent.

[0085] Alternatively, the selected data object may be sent to a recipient
using
the communication subsystem 104 over the wireless network 150 when a
destination (e.g., blog, webpage, RSS feed, etc.) is defined. When a
destination is
not defined, a prompt to enter a destination for the selected data object may
be
provided after which the selected data object will be sent. A notification
that
selected data object has been sent may be displayed on the display 112 to
inform
the user.

[0086] When the left-right cycle gesture is detected, the media player
application causes content of a next data object of the same data type in a
datastore of the media player application, such as a database of data objects
of the
same type stored in the memory 110, to be reproduced. When the data object is
a
digital picture or graphic object, reproducing comprises displaying the
digital picture
or graphic defined by the digital picture or graphic object on the display
112. When
the data object is a video object, reproducing comprises playing the video
defined
by the video object on the display 112 and speaker 128 or routing an
electrical
acoustic audio signal to the data port 126 for output to headphones or other
external speaker. When the data object is an audio object, reproducing
comprises
playing the audio (e.g., song or track) defined by the audio object using the
speaker 128 or routing an electrical acoustic audio signal to the data port
126 for
output to headphones or other external speaker.

[0087] The next data object is determined relative to a currently selected
data
object, for example, in alphabetical order or chronological order from older
to
26


CA 02757527 2011-11-14

newer. The currently selected data object may appear as an entry in a playlist
of
the media player application. The currently selected data object may be
indicated
in a displayed playlist using highlighting or focusing the corresponding entry
in the
displayed playlist or other suitable method of visual indication. Highlighting
or
focusing an entry in the displayed playlist causes the appearance of the
corresponding entry in the displayed playlist to be changed from a first
visual state
to a second visual state different from the first visual state. Changing the
appearance of an entry in the displayed playlist, in at least some
embodiments,
may comprise changing a colour of a background or field of the entry in the
displayed playlist, the text of the entry in the displayed playlist, or both.
Alternatively, the currently selected data object may not be shown or
otherwise
indicated on the display 112.

[0088] The currently selected data object may be in reproduction, for
example, when the currently selected data object is a digital picture or
graphic
object, the currently selected digital picture or graphic may be being
displayed on
the display 112. Similarly, when the currently selected data object is an
audio
object (e.g., song or track), the currently selected song or track may be
being
played, for example, with the speaker 128. When the currently selected data
object is a video object, the currently selected video object may be being
played on
the display 112 and speaker 128.

[0089] When the right-left cycle gesture is detected, the media player
application causes content of a previous data object of the same data type in
a
datastore of the media player application, such as a database of data objects
of the
same type stored in the memory 110, to be reproduced. When the data object is
a
digital picture or graphic object, reproducing comprises displaying the
digital picture
or graphic defined by the digital picture or graphic object on the display
112. When
the data object is a video object, reproducing comprises playing the video
defined
by the video object on the display 112 and speaker 128 or routing an
electrical
acoustic audio signal to the data port 126 for output to headphones or other
external speaker. When the data object is an audio object, reproducing
comprises
playing the audio (e.g., song/track) defined by the audio object using the
speaker
27


CA 02757527 2011-11-14

128 or routing an electrical acoustic audio signal to the data port 126 for
output to
headphones or other external speaker.

[0090] The previous data object is determined relative to a currently selected
data object, for example, in alphabetical order or chronological order from
older to
newer.

[0091] When a data object is not selected, the portable electronic device 100
does not monitor for left-right cycle gestures or right-left cycle gestures
and any
left-right cycle gesture or right-left cycle gesture which is performed is not
detected
is ignored. Alternatively, the portable electronic device 100 may monitor for
and
detect left-right cycle gestures and right-left cycle gestures but ignores any
detected when a data object is not selected. Alternatively, the next or
previous
data object may be determined based on a default data object such as the last
accessed data object of the given type in a media folder, database, or
playlist, or
the newest data object of the given type.

[0092] When the detected motion does not match the toss gesture, left-right
cycle gesture or right-left cycle gesture, the motion is ignored.
Alternatively, if the
detected motion matches another motion gesture recognized by the portable
electronic device 100, the command or action associated with that other motion
gesture may be performed, depending on the embodiment.

[0093] In other embodiments, the toss gesture, left-right cycle gesture and
right-left cycle gestures described above could be applied to calendars in a
calendar
application, which could be part of a PIM on the portable electronic device
100.
Detection of a left-right cycle gesture by the portable electronic device 100
may
cause a previous view of a current view type to be displayed. Detection of a
right-
left cycle gesture by the portable electronic device 100 may cause a next view
of a
current view type to be displayed. A calendar application typically has
several view
types including, but not limited to, an event view, an agenda view, a month
view, a
week view, a day view, etc. The event view shows event details about a
particular
event. The agenda view shows event details about events for the current day.
The
28


CA 02757527 2011-11-14

month view shows the current month including any events in the current month.
The week view shows the current week including any events in the current week.
The day view shows the current day including any events in the current day.
Detection of a toss gesture by the portable electronic device 100 may invite a
second electronic device, such a paired device, to an appointment which is
described in an event view displayed on the display 112, or selected (e.g.,
highlighted) event in an agenda view, month view, week view, day view or other
view displayed on the display 112. The operation of the calendar application
in
connection with the toss gesture, left-right cycle gesture and right-left
cycle
gestures and the above-described commands would operate generally similar to
the
method 800 except for the different functionality described above.

[0094] In other embodiments, the toss gesture, left-right cycle gesture and
right-left cycle gestures described above could be applied to the Web browser
application. Detection of a left-right cycle gesture by the portable
electronic device
100 may cause a back command to be performed by the Web browser. Detection
of a right-left cycle gesture by the portable electronic device 100 may cause
a
forward command to be performed by the Web browser. Detection of a toss
gesture (or shake gesture) by the portable electronic device 100 may cause
creation of a favourite for the current Uniform Resource Locator (URL),
bookmarking of streamed media, or downloading content or queuing content for
download depending on the context. The context-sensitive factors for selecting
the
context-sensitive action may depend on several factors, such as whether
streamed
content is available or selected (e.g., highlighted) in the content (e.g., Web
page)
displayed by the Web browser on the display 112, whether downloadable content
is
available or selected (e.g., highlighted) in the content (e.g., Web page)
displayed
by the Web browser on the display 112. For example, if nothing is selected
when a
toss away gesture is detected, the portable electronic device 100 may send the
page URL to a paired electronic device device and may bookmark the page if
nothing is selected when a toss towards gesture is detected. However,
selecting an
object (e.g., touching an object on the page with a touch-sensitive display
118)
may send, bookmark or download that object.

29


CA 02757527 2011-11-14

[0095] In other embodiments, the toss gesture, left-right cycle gesture and
right-left cycle gestures described above could be applied to cycling between
sources of notification. A notification queue is provided in which all new
notifications, regardless of type, are queued by as on a notification time
stamp
describing when the notification was generated or received. The notification
queue
may be agnostic with respect to the source of notification or notification
type.
The notification queue may be ordered newest to oldest or oldest to newest,
depending on device settings and user preferences. The order of the
notification
queue may be a configurable.

[0096] Notification cycling, in some embodiments, may only be supported
when a messaging application or PIM is the active application 148 on the
portable
electronic device 100, i.e. the foreground application. To be supported when
other
application 138 are active, the gestures used in notification cycling should
be used
by the active application 148 to avoid conflict. Alternatively, the gestures
used in
notification cycling may be rendered temporality unavailable/unsupported for a
threshold duration from the receipt of the notification (e.g., within 5
seconds of the
receipt of a notification). This allows the gestural control of the
notification cycling
to override the gestural control of the active application 148 to prevent
conflicts.
[0097] Detection of a right-left cycle gesture by the portable electronic
device
100 causes the newest notification or the source of the newest notification,
such as
the newest event or electronic message, to be displayed on the display 112. It
will
be appreciated that a notification can act as a source in some instances, for
example, when the notification is a reminder or alarm. When the source of the
notification is displayed, it is removed from the notification queue. When no
notifications .are in the notification queue, a right-left cycle gesture which
is
detected by the portable electronic device 100 when no unread electronic
message
exists is ignored. When the notification queue is limited to new message
notifications.

[0098] Detection of a further right-left cycle gesture by the portable
electronic device 100 when the newest notification or the source of the newest


CA 02757527 2011-11-14

notification is displayed on the display causes the next newest electronic
message
to be displayed on the display 112. Detection of yet a further right-left
cycle
gesture by the portable electronic device 100 when the notification or the
source of
the notification is displayed on the display 112 causes the next newest
electronic
message to be displayed on the display 112, and so on.

[0099] Detection of a left-right cycle gesture by the portable electronic
device
100 when the notification or the source of the notification is displayed on
the
display 112 causes the previously displayed user interface screen, i.e. the
previously displayed message or inbox (if no message was previously displayed)
to
be displayed on the display 112. The re-display/return to the previously
displayed
user interface screen acts as a reset for a notification cycling gesture.

[0100] For example, if a .user is composing an email message and a
notification of a new instant message occurs (e.g., vibration informing the
user of
the new IM), performing a right-left cycle gesture cause the new instant
message
(e.g., within an IM thread) to be displayed on the display 112. Performing a
left-
right cycle causes the email message which the user was composing to be
displayed
on the display 112. Alternatively, if a user is instant messaging and a
notification
of a new RSS (Really Simple Syndication) article in Web feeds occurs (e.g.,
vibration) following by a notification of a new instant message occurs (e.g.,
vibration), performing a right-left cycle gesture causes the new IM message to
be
displayed on the display 112. Performing a further right-left cycle gesture
causes
the new RSS article be displayed on the display 112. Performing a left-right
cycle
causes the conversation in which the user was working to be displayed on the
display 112. If the notification queue is works oldest to newest rather than
newest
to oldest, performing the first right-left cycle gesture would cause the new
RSS
article to be displayed on the display 112 and performing a further right-left
cycle
gesture would cause the new IM message to be displayed on the display 112.
[0101] In other embodiments, the notification queue may be limited to
notifications of a particular type, for example notifications of new messages
of a
particular type. The toss gesture, left-right cycle gesture and right-left
cycle
31


CA 02757527 2011-11-14

gestures described above may be used to cycle through messages of the same
type. A shake gesture, other gesture or input (e.g. depressing of a designated
button or key, or touching of an onscreen element) may be used to change the
particular type of notification, e.g. particular type of message being cycled.

[0102] In some embodiments, detection of a shaking gesture when may
cause the portable electronic device 100 to switch applications 148 among
currently
active applications. In some embodiments, the portable electronic device 100
may
monitor for and detect the shaking gesture when an application 148 is
displayed on
the display 112, i.e. in the foreground. In other embodiments, the portable
electronic device 100 may only monitor for and detect the shaking gesture when
an
application 148 is not displayed, i.e., when the home screen is displayed on
the
display or an application 148 is otherwise not in the foreground. This allows
the
shaking gesture to be used by the application 148 for other purposes.

[0103] There are numerous possible permutations of acceleration gesture
(motion gesture) and command combinations; however, not all acceleration
gesture
and command combinations are procedural efficient to implement or intuitive
for a
user. The present disclosure describes a number of acceleration gesture and
command combinations which can be implemented in a relatively straightforward
manner within a GUI without becoming awkward in terms of processing or user
experience, and without conflicting with other gestural command inputs, touch
command inputs and or other command inputs. These acceleration gesture and
command combinations described herein are believed to provide a more intuitive
user interface for providing the described functionality with less processing
complexity than menu-driven or button/key-driven alternatives.

[0104] While the present disclosure is described primarily in terms of
methods, the present disclosure is also directed to a portable electronic
device
configured to perform at least part of the methods. The portable electronic
device
may be configured using hardware modules, software modules, a combination of
hardware and software modules, or any other suitable manner. The present
disclosure is also directed to a pre-recorded storage device or computer-
readable
32


CA 02757527 2011-11-14

medium having computer-readable code stored thereon, the computer-readable
code being executable by at least one processor of the portable electronic
device for
performing at least parts of the described methods.

[0105] The present disclosure may be embodied in other specific forms
without departing from its spirit or essential characteristics. The described
embodiments are to be considered in all respects as being only illustrative
and not
restrictive. The present disclosure intends to cover and embrace all suitable
changes in technology. The scope of the present disclosure is, therefore,
described
by the appended claims rather than by the foregoing description. All changes
that
come within the meaning and range of equivalency of the claims are intended to
be
embraced within their scope.

33

Representative Drawing
A single figure which represents the drawing illustrating the invention.
Administrative Status

For a clearer understanding of the status of the application/patent presented on this page, the site Disclaimer , as well as the definitions for Patent , Administrative Status , Maintenance Fee  and Payment History  should be consulted.

Administrative Status

Title Date
Forecasted Issue Date Unavailable
(86) PCT Filing Date 2010-11-12
(85) National Entry 2011-11-14
Examination Requested 2011-11-14
(87) PCT Publication Date 2012-05-12
Dead Application 2021-08-31

Abandonment History

Abandonment Date Reason Reinstatement Date
2020-08-31 FAILURE TO PAY APPLICATION MAINTENANCE FEE

Payment History

Fee Type Anniversary Year Due Date Amount Paid Paid Date
Request for Examination $200.00 2011-11-14
Registration of a document - section 124 $100.00 2011-11-14
Application Fee $400.00 2011-11-14
Maintenance Fee - Application - New Act 2 2012-11-13 $100.00 2012-11-01
Maintenance Fee - Application - New Act 3 2013-11-12 $100.00 2013-10-22
Maintenance Fee - Application - New Act 4 2014-11-12 $100.00 2014-10-22
Maintenance Fee - Application - New Act 5 2015-11-12 $200.00 2015-10-28
Maintenance Fee - Application - New Act 6 2016-11-14 $200.00 2016-10-18
Registration of a document - section 124 $100.00 2017-05-03
Maintenance Fee - Application - New Act 7 2017-11-14 $200.00 2017-10-26
Maintenance Fee - Application - New Act 8 2018-11-13 $200.00 2018-10-19
Owners on Record

Note: Records showing the ownership history in alphabetical order.

Current Owners on Record
BLACKBERRY LIMITED
Past Owners on Record
RESEARCH IN MOTION LIMITED
Past Owners that do not appear in the "Owners on Record" listing will appear in other documentation within the application.
Documents

To view selected files, please enter reCAPTCHA code :



To view images, click a link in the Document Description column. To download the documents, select one or more checkboxes in the first column and then click the "Download Selected in PDF format (Zip Archive)" or the "Download Selected as Single PDF" button.

List of published and non-published patent-specific documents on the CPD .

If you have any difficulty accessing content, you can call the Client Service Centre at 1-866-997-1936 or send them an e-mail at CIPO Client Service Centre.


Document
Description 
Date
(yyyy-mm-dd) 
Number of pages   Size of Image (KB) 
Abstract 2011-11-14 1 19
Description 2011-11-14 33 1,720
Claims 2011-11-14 4 112
Drawings 2011-11-14 11 184
Cover Page 2012-10-19 1 33
Representative Drawing 2013-06-03 1 8
Claims 2013-11-19 3 83
Description 2013-11-19 33 1,724
Description 2015-06-01 33 1,723
Claims 2015-06-01 2 75
Claims 2016-09-20 3 92
Amendment 2017-08-30 20 872
Claims 2017-08-30 5 183
Examiner Requisition 2018-02-09 4 194
Claims 2018-07-20 5 196
Amendment 2018-07-20 24 1,252
Examiner Requisition 2018-12-27 4 236
Assignment 2011-11-14 9 295
Amendment 2019-06-27 6 212
Prosecution-Amendment 2013-03-22 2 72
Prosecution-Amendment 2013-07-24 3 87
Prosecution-Amendment 2013-11-19 13 410
PCT 2010-12-11 7 292
Prosecution-Amendment 2014-12-02 5 262
Prosecution-Amendment 2015-06-01 16 609
Prosecution-Amendment 2016-09-20 12 444
Examiner Requisition 2016-03-23 4 295
Examiner Requisition 2017-03-15 4 235