Title of Invention

AN ILLUMINATED TOUCHPAD AND METHOD OF OPERATION THEREOF

Abstract Improved feedback mechanisms for touch pads are disclosed. One aspect relates to devices capable of illuminating the touch sensitive surface of the touch pad. Another aspect relates to methods for providing visual feedback at the touch pad.
Full Text ILLUMINATED TOUCHPAD
BACKGROUND OF THE INVENTION
Field of the Invention
[0001] The present invention relates generally to touch pads that provide visual
feedback. More particularly, the present invention relates to illuminated touch pads
that use light to provide feedback.
Description of the Related Art
[0002] There exist today many styles of input devices for performing operations in a
computer system. The operations generally correspond to moving a cursor and/or
making selections on a display screen. By way of example, the input devices may
include buttons or keys, mice, trackballs, touch pads, joy sticks, touch screens and the
like.
[0003] Touch pads, in particular, are becoming increasingly popular because of their
ease and versatility of operation as well as to their declining price. Touch pads allow
a user to make selections and move a cursor by simply touching an input surface via a
finger or stylus. In general, the touch pad recognizes the touch and position of the
touch on the input surface and the computer system interprets the touch and thereafter
performs an action based on the touch event.
[0004] Touch pads typically include an opaque touch panel, a controller and a
software driver. The touch panel registers touch events and sends these signals to the
controller. The controller processes these signals and sends the data to the computer
system. The software driver translates the touch events into computer events.
[0005] Although touch pads work well, improvements to their form feel and
functionality are desired. By way of example, it may be desirable to provide visual
stimuli at the touch pad so that a user can better operate the touch pad. For example,
the visual stimuli may be used (among others) to alert a user when the touch pad is
registering; a touch, alert a user where the touch is occurring on the touch pad, provide
feedback related to the touch event, indicate the state of the touch pad, and/or the like.
SUMMARY OF THE INVENTION
[0006] The invention relates, in one embodiment, to an illuminated input device. The
illuminated input device includes an object sensing mechanism capable of sensing a

user input over an input surface. The illuminated input device also includes a visual
feedback system configured to illuminate the input surface in association with a user
input.
[0007] The invention relates, in another embodiment, to a method of operating an
input device. The method includes sensing an object over an input surface. The
method also includes and illuminating at least a portion of the input surface when an
object is sensed.
[0008] The invention relates, in another embodiment, to a method of operating an
input device. The method includes illuminating at least a portion of an input surface
when an object is detected over the input surface. The method also includes adjusting
the illumination when the object is moved over the input surface.
[0009] The invention relates, in another embodiment, to a method of operating an
input device. The method includes detecting a user input over the input surface. The
method also includes determining an input state of the input device based on the user
input. The method additionally includes illuminating the input surface based on the
input state of the input device. Each input state having a different illumination
profile.
BRIEF DESCRIPTION OF THE DRAWINGS
[0010] The invention may best be understood by reference to the following
description taken in conjunction with the accompanying drawings in which:
[0011] Fig. 1 is a simplified block diagram of an input device, in accordance with one
embodiment of the present invention.
[0012] Fig. 2 is a method of operating an input device, in accordance with one
embodiment of the present invention.
[0013] Fig. 3 is a method of operating an input device, in accordance with one
embodiment of the present invention.
[0014] Fig. 4 is a simplified diagram of an illuminated touch pad, in accordance with
one embodiment of the present invention.
[0015] Fig. 5 is a simplified diagram of an illuminated touch pad, in accordance with
one embodiment of the present invention.
[0016] Fig. 6A is a side view of an illuminated touch pad, in accordance with one
embodiment of the present invention.

[0017] Fig. 6B is an exploded perspective view of the illuminated touch pad of Fig.
6A, in accordance with one embodiment of the present invention.
[0018] Fig. 7A is a side view of an illuminated touch pad, in accordance with one
embodiment of the present invention.
[0019] Fig. 7B is a top view, in part, of the touch pad of Fig, 7A, in accordance with
another embodiment of the present invention.
[0020] Fig. 8A is a side view of an illuminated touch pad, in accordance with one
embodiment of the present invention.
[0021] Fig. 8B is a top view, in part, of the touch pad of Fig. 8A, in accordance with
another embodiment of the present invention.
[0022] Fig. 9A is a side view of an illuminated touch pad, in accordance with one
embodiment of the present invention.
[0023] Fig. 9B is a top view, in part, of the touch pad of Fig. 7A, in accordance with
another embodiment of the present invention.
[0024] Fig 10 is a diagram of an illuminated touch pad, in accordance with one
embodiment of the present invention.
[0025] Fig. 11 is a diagram of a light panel that can be used in an illuminated touch
pad, in accordance with another embodiment of the present invention.
[0026] Fig. 12 is a method of operating an illuminated touch pad, in accordance with
one embodiment of the present invention.
[0027] Fig. 13A illustrates one implementation where an angular segment is
illuminated when the user places their finger over the angular segment, in accordance
with one embodiment of the present invention.
[0028] Fig. 13B illustrates one implementation where two angular segments are
illuminated at the same time when two finger are distinctly placed over the two
angular segments, in accordance with one embodiment of the present invention.
[0029] Fig. 14A illustrates one implementation where illumination points adjacent
and surrounding the location of the finger are illuminated when the user places their
finger over the input surface, in accordance with one embodiment of the present
invention.
[0030] Fig. 14B illustrates one implementation where the area around two fingers are
illuminated at the same time when the two finger are placed over the input surface at
the same time, in accordance with one embodiment of the present invention.

[0031] Fig. 15 is a method of operating an illuminated touch pad, in accordance with
one embodiment of the present invention.
[0032] Figs. 16A-16D illustrate one implementation where the illuminated portion
follows the motion of the finger as it is moved across the surface of the touch pad, in
accordance with one embodiment of the present invention.
[0033] Figs. 17A-17D illustrate one implementation where the illuminated portion
follows the motion of the finger as it is moved across the surface of the touch pad, in
accordance with one embodiment of the present invention.
[0034] Fig. 18 is a method of operating an illuminated touch pad, in accordance with
one embodiment of the present invention.
[0035] Fig. 19A is a method of determining the state of the touch pad, in accordance
with one embodiment of the present invention.
[0036] Fig. 19B is a method of illuminating a touch pad, in accordance with one
embodiment of the present invention.
[0037] Figs. 20A and 20B illustrate illuminating the touch pad with a first color or
intensity when the touch pad is in a first state, and illuminating the touch pad with a
second color or intensity when the touch pad is in a second state, in accordance with
one embodiment of the present invention.
[0038] Fig. 21 is a method of operating an illuminated touch pad, in accordance with
one embodiment of the present invention.
[0039] Figs. 22A and 22B illustrate increasing the intensity of the illumination when
an object is closer or exerts increased pressure relative to the touch surface, and
decreasing the intensity of the illumination when an object is further away or exerts
decreased pressure relative to the touch surface, in accordance with one embodiment
of the present invention.
[0040] Fig. 23 is a method of operating an illuminated touch pad, in accordance with
one embodiment of the present invention.
[0041] Fig. 24 is a method of operating an illuminated touch pad, in accordance with
one embodiment of the present invention.
[0042] Figs. 25A-25D illustrate providing low intensity illumination when a touch is
first detected, providing medium intensity illumination when the object is slowly
moved around the input surface (e.g., low acceleration), and providing high intensity
illumination when the object is quickly moved around the input surface (e.g., high
acceleration), in accordance with one embodiment of the present invention.

DETAILED DESCRIPTION OF THE INVENTION
[0043] The present invention pertains to improved feedback mechanisms for touch
pads. One aspect relates to devices capable of illuminating the touch sensitive surface
of the touch pad. Not just in backlighting so that the user knows where the touchpad
is located in low light conditions, but also to give other feedback related to how the
touch pad is being used. Another aspect relates to methods for providing feedback at
the touch pad. For example, changing intensity or color based on motion
characteristics and/or pressure, providing an illumination point that follows a finger
as it is moved about the touch sensitive surface, showing different states with varying
levels of brightness or color, etc.
[0044] Embodiments of the invention are discussed below with reference to Figs. 1-
25. However, those skilled in the art will readily appreciate that the detailed
description given herein with respect to these figures is for explanatory purposes as
the invention extends beyond these limited embodiments.
[0045] Fig. 1 is a simplified block diagram of an input device 10, in accordance with
one embodiment of the present invention. The input device 10 may be a standalone
peripheral device that connects to a host device through wired or wireless connections
or it may be integrated into a host device (e.g., hard wired). In either case, the input
device 10 is configured to provide inputs to the host device. Examples of host devices
include any consumer related electronic device such as computers, PDAs, media
players, telephones, etc.
[0046] In order to generate inputs as for example initiating commands, making
selections or tracking, the input device 10 includes an object sensing mechanism 12
configured to detect one or more objects in close proximity to and/or in contact with
an input surface 13. The object sensing mechanism 12 may be based on proximity
sensing and/or touch sensing.
[0047] In the case of proximity sensing, the input surface 13 may be the surface
directly underneath a proximity sensing field. The object sensing mechanism 12
generates input signals when, an object such as a finger (or stylus) is moved above the
input surface and within the sensing field (e.g., x and y plane), from an object
holding a particular position above the surface and within the sensing field and/or by
an object moving through or in and out of the sensing field (e.g., z direction).
Proximity detection may be based on technologies including but not limited to

capacitiive, electric field, inductive, hall effect, reed, eddy current, magneto resistive,
optical shadow, optical visual light, optical IR, optical color recognition, ultrasonic,
acoustic emission, radar, heat, sonar, conductive or resistive and the like.
[0048] In the case of touch sensing, the input surface 13 may be a touch surface that
is sensitive to direct physical touch. The object sensing mechanism 12 generates
input signals when an object in contact with the input surface 13 is moved across the
input surface (e.g., x and y plane), from an object holding a particular position on the
input surface and/or by an object tapping on the input surface. Touch sensing may be
based on technologies including but not limited to resistive, capacitive, infrared and
surface acoustic wave. Examples of touch sensing devices that utilize these
technologies include touch pads, touch screens, and the like.
[0049] To elaborate, the sensing region, i.e., input surface or the sensing field above
the input surface, is typically divided into several independent and spatially distinct
sensing points, nodes or regions. The sensing points, which are typically hidden from
view, are dispersed about the sensing region with each sensing point representing a
different position in the sensing region. The sensing points may be positioned in a
grid or a pixel array where each pixilated sensing point is capable of generating a
signal. In the simplest case, a signal is produced each time an object is positioned
over a sensing point. When an object is placed over multiple sensing points or when
the object is moved between or over multiple sensing points, multiple signals can be
generated. The sensing points generally map the sensing region into a coordinate
system such as a Cartesian coordinate system, a Polar coordinate system or some
other coordinate system. Furthermore, the touch sensing means may be based on
single point sensing or multipoint sensing. Single point sensing is capable of only
distinguishing a single object at any given time, while multipoint sensing is capable
of distinguishing multiple objects at the same time.
[0050] The input device 10 also includes a visual feedback system 14 configured to
output visual effects at the input surface 13 in association with the object sensing
system 12. The visual feedback system 14 is dedicated to enhancing the operation of
the input device 10 by providing visual feedback to the user when making touch or
proximity inputs via the object sensing system 12. For example, the visual effects
may be used to indicate the location of the input surface 13 thereby making inputting
easier for the user (e.g., backlighting).

[0051] Alternatively or additionally, the visual effects may be used during and after
the input event to dynamically indicate characteristics associated with the input
events. The characteristics may for example include when and where and the
number of inputs being made relative to the input surface 13. This type of feedback
also improves inputting by providing-visual cues to the user about the input device 10
as the input device 10 is used.
[0052] Alternatively or additionally, the visual effects may be used before an input
event to invoke the user to perform a particular input event at the input surface 13.
This type of feedback also improves inputting by helping the user make appropriate
inputs or helping them learn input gestures (e.g., timing, location and movements).
[0053] In most cases, the visual effects (outputs) are linked or tied to and associated
with the input events being performed. Although capable of, they typically do not
provide external outputs associated with events occurring outside the input device 10.
That is, the visual effects typically do not operate separately as an output for any
device other than the input device 10 (e.g., the visual feedback system should not be
considered a separate display).
[0054] The visual feedback system 14 includes one or more visual changing elements
15. The visual changing elements 15 may be separate of integral with the sensing
elements of the object sensing system 12. In some cases, the one or more of the
visual changing elements 15 may be mapped, associated with or tied to one or more
of the sensing nodes of the object sensing system. The number of visual changing
elements tied to a particular sensing node may be less than, equal to, or more than the
number of sensing nodes.
[0055] The resolution of the visual changing elements 15 can be widely varied. In
one embodiment, the resolution of the visual changing elements 15 is greater than the
sensing nodes (e.g., the number of visual changing elements is greater than the
number of sensing nodes). la another embodiment, the resolution of the visual
changing elements 15 is substantially equal to the resolution of the sensing nodes
(e.g., the number of visual changing elements is substantially equal to the number of
sensing nodes), hi yet "another embodiment, the resolution of the visual changing
elements 15 is less than the resolution of the sensing nodes (e.g., the number of visual
changing elements is less than the number of sensing nodes). The resolution
generally depends on the needs of the particular input device. In some cases, high
resolution is need to create dynamic visual effects such a graphical effects. In other

cases, only low resolution is required, as for example to visually change a region as
large as a finger.
[0056] The visual feedback system 14 may be widely varied. In one embodiment, the
visual feedback system 14 is a light based system that illuminates the input surface
13. In this embodiment, the visual changing elements 15 are embodied as light
emitting devices. The light emitting devices may include one or more light sources,
and a light distribution system for distributing the light at the input surface 13. In
some cases, the light from the light sources may be diffused so that the input surface
13 emits a characteristic glow (not a precise point of light, but rather a blurred
glowing effect or phosphorous glow). That is, the input surface can generate glowing
special effects that may for example provide backlighting to the input surface 13
and/or provide an outline, trace or shadow of the sensed object on the input surface
13. The glowing special effects may even indicate a state of the input device as for
example when the input device is in a tracking state or gesture state.
[0057] Alternatively, the visual feedback system may be a graphically based system
that generates graphics at the input surface. LCDs, OLEDs and electronic inks are
examples of graphically based systems. These devices however can be cost
prohibitive and more complex to implement when compared to light based systems.
[0058] The input device 10 also includes a controller 16 that is operatively coupled to •
the object sensing device 12 and visual feedback system 14. The controller 16
monitors signals generated at the input surface 13 and sends corresponding control
signals associated therewith to the host device, which interprets the signals in
accordance with its programming, (e.g., input events). The controller 16 also
generates visual effect commands for controlling the visual effects outputted by the
visual feedback system 14. Single or multiple commands can be generated to
change one, some or all of the visual changing elements at the same time. Further,
the commands may be based on the signals generated via the object sensing device
12.
[0059] In one embodiment, the controller 16 may instruct the visual changing
elements to change in a non-trivial manner in the region of the detected object in
order to indicate a location of the object relative to the input surface 13. ha another
embodiment, the commands may be based on instructions from the host device. For
example, the host device may instruct the input device 10 to backlight the input
surface 13 or alternatively to alter the input surface in such a way as to prompt the

user to perform a particular event relative to the input surface (e.g., the host device
may use the input surface to teach the user how to perform a particular gesture).
[0060] Fig. 2 is a method 50 of operating an input device, in accordance with one
embodiment of the present invention. The input device may for example be the input
device shown and described in Fig. 1. The method begins at block 52 where one or
more objects are detected at a sensing region. This may for example be accomplished
with the object sensing device described above. Following block 52, the method
proceeds to blocks 54 and 56. In block 54, visual effects are displayed at the sensing
region based on the detected objects. This may for example be accomplished with the
visual feedback system described above.
[0061] In one embodiment, the visual effects are performed in the region of the
detected object. For example, in the case of a light based system, the area under
and/or around the detected object may be illuminated. As should be appreciated, the
visual effects can be made to follow the object as it is moved around the sensing
region. In fact in some cases, the visual effects may include a leading edge, a body,
and/or a trailing edge. The leading edge indicates where the object is directed, the
body indicates the current location of the object, and the trailing edge indicates where
the object has been.
[0062] In another embodiment, the visual effects are performed to indicate the state
of the object sensing event. For example, if one object is detected, a first visual effect
may be performed, and if multiple objects are detected, a second visual effect may be
performed. This may be beneficial in cases where single objects are used for tracking
and multiple objects are used for gesturing.
[0063] In block 56, control signals are generated based on the detected objects. This
may for example be accomplished with the controller described above. The signals
are reported to a host device as an input event, and the host device interprets the
signals in accordance with its programming.
[0064] Fig. 3 is a method 70 of operating an input device, in accordance with one
embodiment of the present invention. The input device may for example be the input
device shown and described in Fig. 1. The method begins at block 72 where visual
effects are displayed at the sensing region. This may for example be accomplished
with the visual feedback system described above.
[0065] In one embodiment, the visual effects are based on a control signal from the
host device. For example, the host device may instruct the visual feedback system to

output visual effects to encourage a user to place an object at a particular location at
the input surface or to perform a gesture during a particular operation in the host
device (e.g., training sequence).
[0066] Following block 72, the method proceeds to blocks 74 and 76. In block 74,
the sensing region is monitored. This may for example be accomplished with the
object sensing device described above. In block 76, control signals are generated
when objects are detected in the sensing region. This may for example be
accomplished with the controller described above. The signals may be used by the
host device as an input event.
[0067] Fig. 4 is a simplified diagram of touch pad 100, in accordance with one
embodiment of the present invention. In this embodiment, the touch pad 100 includes
an optically transmissive touch sensing device 102 disposed over a light panel 104.
Both the touch sensing device 102 and the light panel 104 communicate with a
controller 106 that monitors touch inputs on the touch sensitive surface 108 of the
touch sensing device 102 and that directs the light panel 104 to emit light in the
direction of the touch sensing device in a controlled manner.
[0068] The touch sensing device may be widely varied. The touch sensing device
may for example be selected from any of those used for touch screens.
[0069] The light panel may also be widely varied. In one embodiment, the light
panel is a pixilated light device that includes a plurality of light sources mat are
distributed over an extended area such as the touch sensitive surface of the touch
sensing device. The light panel may include a plurality of light emitting diodes
(LEDs) that are laid out in a matrix such as rows and columns. Any number of LEDs
may be used. The number generally depends on the desired resolution of the light
panel. In the simplest case, LEDs are placed next to or adjacent one another in rows
and columns on a PCB that is sized similarly to the touch sensing device (e.g., covers
the same area).
[0070] Fig. 5 is a simplified diagram of a touch pad 120, in accordance with one
embodiment of the present invention. In this embodiment, the touch pad 120 includes
an opaque or alternatively an optically transmissive touch sensing device 122
disposed below a light panel 124. Both the touch sensing device 122 and the light
panel 124 communicate with a controller 126 that monitors touch inputs on the touch
sensitive surface 128 of the touch sensing device 122 and that directs the light panel
124 to emit light in a controlled manner.

[0071] The touch sensing device may be widely varied. The touch sensing device
may for example be selected from any of those used for touch pads or touch screens.
[0072] The light panel may also be widely varied. Unlike the light panel discussed in
Fig. 4, this light panel needs to allow touch sensing to occur therethrough. The light
panel may further need to.be somewhat diffused to hide the electronics displayed -
underneath the light panel. In one embodiment, the light panel includes one or more
light sources and a light distribution mechanism for distributing the light from the
light source over an extended area such as the touch sensitive surface of the touch
sensing device. The light distribution mechanism may include for example light
pipes or light guides that allow the light sources to be placed away from the sensing
field. In one embodiment, the light distribution mechamism is formed from a
dielectric material so that touch sensing can occur therethrough with impediments
(e.g., capacitance). By way of example, transparent or semi-transparent plastic
materials may be used.
[0073] Figs. 6A and 6B are diagrams of an illuminable touch pad 150, in accordance
with one embodiment of the present invention. The touch pad 150 includes a
translucent or semi-translucent touch screen 152 and a pixilated light panel 154
disposed below the touch screen 152. The touch screen 152 is divided into several
independent and spatially distinct sensing points, nodes or regions. The sensing
points, which are hidden from view (transparent), are dispersed about the touch
screen 152 with each sensing point representing a different position on the surface of
the touch, screen (or touch screen plane). The sensing points may be positioned in a
grid or a pixel array where each pixilated sensing point is capable of generating a
signal. In the simplest case, a signal is produced each time an object is positioned
over a sensing point. When an object is placed over multiple sensing points or when
the object is moved between or over multiple sensing point, multiple signals can be
generated.
[0074] In one embodiment, the touch screen 152 includes a plurality of capacitance
sensing nodes. The capacitive sensing nodes may be widely varied. For example, the
capacitive sensing nodes may be based on self-capacitance or mutual capacitance. In
self-capacitance, the "self capacitance of a single electrode is measured as for
example relative to ground. In mutual capacitance, the mutual capacitance between at
least first and second electrodes is measured. In either cases, each of the nodes

works independent of the other nodes so as to produce simultaneously occurring
signals representative of different points on the touch screen 152.
[0075] In order to produce a transparent touch screen 152, the capacitance sensing
nodes may be formed with a transparent conductive medium such as indium tin oxide
(ITO).
[0076] In self-capacitance sensing arrangements, the transparent conductive medium
is patterned into spatially separated electrodes and traces. Each of the electrodes
represents a different coordinate and the traces connect the electrodes to a capacitive
sensing circuit. The coordinates may be associated with Cartesian coordinate system
(x and y), Polar coordinate system (r, θ) or some other coordinate system. During
operation, the capacitive sensing circuit monitors changes in capacitance that occur at
each of the electrodes. The positions where changes occur and the magnitude of
those changes are used to help recognize the touch events. A change in capacitance
typically occurs at an electrode when a user places an object such as a finger in close
proximity to the electrode, i.e., the object steals charge thereby affecting the
capacitance.
[0077] In mutual capacitance, the transparent conductive medium is patterned into a
group of spatially separated lines formed on two different layers. Driving lines are
formed on a first layer and sensing lines are formed on a second layer. Although
separated by being on different layers, the sensing lines traverse, intersect or cut
across the driving lines thereby forming a capacitive coupling node. The manner in
which the sensing lines cut across the driving lines generally depends on the
coordinate system used. For example, in a Cartesian coordinate system, the sensing
lines are perpendicular to the driving lines thereby forming nodes with distinct x and
y coordinates. Alternatively, in a polar coordinate system, the sensing lines may be
concentric circles and the driving lines may be radially extending lines (or vice
versa). The driving lines are connected to a voltage source and the sensing lines are
connected to capacitive sensing circuit. During operation, a current is driven through
one driving line at a time, and because of capacitive coupling, the current is carried
through to the sensing lines at each of the nodes (e.g., intersection points).
Furthermore, the sensing circuit monitors changes in capacitance that occurs at each
of the nodes. The positions where changes occur and the magnitude of those changes
are used to help recognize the multiple touch events. A change in capacitance
typically occurs at a capacitive coupling node when a user places an object such as a

finger in close proximity to the capacitive coupling node, i.e., the object steals charge
thereby affecting the capacitance.
[0078] Referring now to the light panel 154, the light panel 154 includes a light
emitting surface that is typically divided into several independent and spatially
distinct illumination points, nodes or regions 156. The illumination points 156 are
dispersed about the light emitting surface with each illumination point 156
representing a different position in the light emitting surface. The illumination points
156 may be positioned in a grid or a pixel array where each pixilated illumination
point is capable of emitting light. The illumination points 156 generally map the
illumination region into a coordinate system such as a Cartesian coordinate system, a
Polar coordinate system or some other coordinate system. In some cases, the
illuminaitions points may be laid out in a pattern similar to the sensing points of the
touch panel 152 (e.g., same coordinate system, same number of points). In other
cases, the illumination points 156 may be laid out in a pattern that is different than the
sensing points of the touch panel 152 (e.g., different coordinate system, different
number of points).
[0079] The light panel 154 may be widely varied. In the illustrated embodiment, the
illumination points 156 are embodied as individual light emitting diodes that are
placed in a grid like manner thereby forming a pixilated illumination area, i.e., each
of the light emitting diodes forms an illumination node. The grid may be oriented
rows and columns (x and y) or angular/radial segments (as shown). Furthermore, the
LEDs are attached to the printed circuit board 160 and operatively coupled to the
controller 158 located on the backside of the printed circuit board 160.
[0080] The touch screen 152 is also operatively coupled to the controller 158 as for
example using a flex circuit that attached to the printed circuit board 160. During
operation, the controller 158 monitors the changes in capacitance and generates
control signals based on these changes. The controller 158 also separately adjusts the
intensity of each of the LEDs to illuminate portions or all of the touch screen 152 in a
controlled manner. That is, the light panel 154 can produce any number of various
light effects by selectively controlling the intensities of the LED's via the controller
158. Because the touch screen 154 is translucent, the light can be seen through the
touch screen 154.
[0081] In some cases, the touch pad 150 may further include a light diffuser 162.
The light diffuser 162 is configured to diffuse the light being emitted by the light

panel 154. This may be done to normalize the light intensity of the LEDs, to produce
a characteristic glow at the input surface, and/or to hide the physical parts of the touch
pad located underneath the light diffuser.
[0082] Although the light diffuser 162 can include color components, in most cases,
the light diffuser appears as a white or semi transparent white material. When
embodied with white elements, the light diffuser 162 takes on the color of light
emitted by the LEDs. Generally speaking, the light diffuser 162 is positioned
somewhere between the LEDs and the input surface. More particularly, the light
diffuser 162 can be placed above, within or underneath the touch screen. For
example, a light diffuser 162 can be placed on the upper surface, lower surface, or in
the layers of the touch screen. Alternatively or additionally, the light diffuser 162
may be integrated with or attached to the light panel or even be a separate component
disposed between the light panel 154 and touch screen 152 (as shown).
[0083] The light diffuser 162 may be embodied in many different forms including for
example surface treatments on one or more layers of the touch screen, additives in
one or more layers of the touch screen, an additional layer in the touch screen, rigid
plastic inserts disposed above or below the touch screen, flexible labels disposed
above or below the touch screen, and the like. The light diffuser 162 may even be the
ITO coating used to form the sensing components of the touch screen (e.g., the
greater the density of the ITO coating, the greater the amount of light that is diffused).
[0084] In the illustrated embodiment, the light diffuser 162 is a plastic insert that
includes light scattering additives. Furthermore, the light diffuser 162 is disposed
between the light panel 154 and the touch screen 152.
[0085] It should be pointed out that LED's offer many advantages over other light
sources. For example, LED's are relatively small devices that are energy efficient
and long lasting. LED's also run relatively cool and are low in cost. Furthermore,
LED's come in various colors such as white, blue, green, red and the like. The
pixilated LEDs may be configured to emit that same color of light or a different color
of light.
[0086] Furthermore, although shown as single LEDs, it should be noted that the
LEDs may be embodied as an integrated array of LEDs that are grouped together as
for example an array of red, blue, green and/or white LEDs that cooperate to produce
a resultant color (via color mixing). The resultant color may be a wide range of
colors, as for example, a majority of the colors from the color spectrum. During

operation, the controller can produced almost any color by adjusting the intensity of
each of the colored LED's. By way of example, in order to produce the highest shade
of red, the intensities of the green and blue are reduced to zero intensity and the
intensity of the red is increased to its peak intensity. The highest shades of green and
blue can be implemented in a similar manner. In addition, in order to produce a shade
of red and green, the intensities of the green and red are increased to levels above
zero intensity while the intensity of blue is reduced to zero intensity. Shades of green
and blue and blue and red can be implemented in a similar manner. Furthermore, in
order to produce shades of white, the intensities of the red, green and blue are
increased to the same levels above zero intensity, or alternatively the red, green and
blue LED's are turned off and a white LED is turned on.
[0087] Although the integrated LED array is described as using the three primary
colors, it should be noted that this is not a limitation and that other combinations may
be used. For example, the integrated LED array may be configured to include only
two of the primary colors or it may only include LED's with a single color.
[0088] When the LEDs are capable of generating any color, unique input surfaces can
be produced. By way of example, the touch pad can produce an input surface with
rainbow stripes, different colored spots, different colored quadrants or sections and
the like. The touch pad can also produce an input surface that has a dynamically
changing pattern. This is typically accomplished by activating distinct LED's at
different times or by adjusting the intensities of distinct LED's at different times.
[0089] Figs 7A and 7B are diagrams of a touch pad 200, in accordance with another
embodiment of the present invention. The touch pad 200 includes various layers
including a light panel 202, an electrode layer 204 and a printed circuit board 206
(PCB). The electrode layer 204 is positioned on the PCB 206 and the light panel 202
is placed above the electrode layer 204.
[0090] The electrode layer 204 includes a plurality of spatially separated electrodes
205 configured to detect changes in capacitance at an upper surface 208 of the light
panel 202. Each of the electrodes 205 is operatively coupled to a controller 210
located on the backside of the printed circuit board 206. During operation, the
controller 210 monitors the changes in capacitance and generates control signals
based on these changes.
[0091] The light panel 202 includes a light distribution panel 212 disposed over the
electrode layer 204 and one or more side mounted light emitting diodes 214 disposed

around the periphery of the light distribution panel 212. The side mounted light
emitting diodes 214 are configured to direct light into a different portion of the light
distribution panel 212. Alternatively, a light pipe may be used to direct light from an
LED located away from the light distribution panel 212. The light distribution panel
212 is configured to redirect the light made incident thereon via the light emitting
diodes 214 to an upper surface of the light distribution panel 212 thereby illuminating
the touch pad surface 201. The light distribution panel 212 is also configured to serve
as a dielectric layer that covers the electrode layer 204 in order to help form the
capacitance sensing circuit of the touch pad 200.
[0092] As shown, the LEDs 214 are attached to the printed circuit board 206 and
operatively coupled to the controller 210 located on the backside of the printed circuit
board 206. During operation, the controller 210 selectively adjusts the intensity of
each of the LEDs 214 to illuminate portions of or all of the light distribution panel
212 in a controlled manner.
[0093] Although shown as single LEDs, the LEDs may be embodied as an array of
LEDs as for example an array of red, blue and green LEDs. Arrayed LEDs such as
this may be capable of generating most colors in the color spectrum.
[0094] The light distribution panel 212 can be widely varied. In one embodiment, the
light distribution panel 212 is a separate component disposed within the housing 211
of the touch pad 200. For example, the light distribution panel 212 is inserted within
an opening in the housing 211 (as shown). In this arrangement, it may be preferable
to place the upper surface of the light distribution panel 212 flush with or recessed
below the outer surface of the housing 211. Furthermore, in order to provide a tight
fit that limits dust and particles from entering the touch pad 200, the light distribution
panel 212 may include edges that extend over the outer surface of the housing 211.
[0095] In another embodiment, the light distribution panel 212 is an integral part of
the housing 211. For example, the housing 211 is formed from a transparent or semi-
transparent material. This particular embodiment provides a continuous surface
without gaps or breaks, which can be aesthetically pleasing to the user.
[0096] In either embodiment, the light distribution panel 212 typically includes a
portion 213 that extends below the inner surface of the housing 211. This portion 213
provides a light receiving area at the sides of the light distribution panel 212 for
receiving light emitted by the side mounted LED's 214.

[0097] The light distribution panel 212, which can be formed from a single or
multiple layers, is typically formed from translucent or semi-translucent dielectric
materials including for example plastic materials such as polycarbonate, acrylic or
ABS plastic. It should be appreciated, however, that these materials are not a
limitation and thatany optically transmittable dielectric material may be used.
[0098] In most cases, the light distribution panel 212 or some other component of the
touch pad 200 includes light diffusing elements to diffuse the light made incident
thereon in order to normalize the light intensity of the LEDs, to produce a
characteristic glow at the input surface, and/or to hide the physical parts of the touch
pad located underneath the input surface. The light diffusing elements may be
provided on an inner surface, outer surface or they may be embedded inside the light
distribution panel 212. Additionally or alternatively, the light diffusing elements can
also be applied to a separate optical component disposed above the light distribution
panel 212.
[0099] In one embodiment, the light diffusing element: is an additive disposed inside
the light distribution panel 212. For example, the light distribution panel 212 may
include a plurality of light scattering particles dispersed between the top and bottom
surfaces of the light distribution panel. When the light is made incident on the inner
surface, it is transmitted through the light distribution panel 212 until is intersects a
light scattering particle disposed inside the panel. After intersecting the light
scattering particle, the light is scattered outwards in a plurality of directions, i.e., the
light is reflected off the surface and/or refracted through the light scattering particle
thereby creating the characteristic glow. By way of example, the light scattering
particles may be formed from small glass particles or white pigments. Furthermore,
by changing the amount of light scattering particles disposed in the panel, the
characteristics of the glow can be altered, i.e., the greater the particles the greater the
light scattering.
[00100] In another embodiment, the light diffusing element is a layer, coating
and/or texture that is applied to the inner, side or outer surfaces of the panel 212. For
example, the panel 212 may include a light scattering coating or a light scattering
texture disposed on the side or outer surface of the panel. By way of example, the
light scattering coating may be a paint, film or spray coating. In addition, the light
scattering texture may be a molded surface of the wall or a sandblasted surface of the
panel. When light is made incident on the inner or outer surface, it intersects the light

scattering coating or texture applied on the surface. After intersecting the light
scattering coating or the light scattering texture, the light is scattered outwards in a
plurality of directions, i.e., the light is reflected off the surface and/or refracted
through the light scattering particle thereby creating a characteristic glow.
[00101] In the illustrated embodiment, the light diffusing element is embodied
as a light diffusing label 216. The light diffusing label 216 is at least adhered to the
top surface of the light distribution panel 212. In some cases, the label 216 may even
extend over and be adhered to a top edge of the housing wall 211. In cases such as
this, the light diffusing label 216 may even be placed in a pocket formed by recesses
217 at the top edge of the housing wall 211 in order to make the top surface of the
light diffusing label 216 flush with the external surface of the housing wall 211. The
label 216 can have a graphic printed thereon, can have multiple colors and can have
varying thickness to assist in controlling the intensity and color of the illumination.
The label 216 may be formed from transparent or semitransparent dielectric materials
such as Mylar or Polycarbonate or any other dielectric material that is thin, optically
transmittable and includes some sort of light diffusing means.
[00102] Further, the light distribution panel 212 may be configured as a single
node, or it may be broken up into plurality of distinct nodes 218, each of which
includes its own dedicated light emitting diode for individual illumination thereof.
During operation, when light is released by a light emitting diode 214, the light is
made incident on the side of the light distribution panel 212 at the node 218. The
node 218 redirects and transmits the light from its side to an upper surface of the node
218. In order to prevent light bleeding between adjacent nodes 218, each node 218
may be optically separated by a reflecting or masking region disposed therebetween.
[00103] Each of the nodes 218 may be formed from a solid piece of material or
it may be formed from a combination of elements. In one embodiment, each of the
nodes 218 is formed from a translucent or semi-translucent plastic insert that when
combined with the other inserts forms the light distribution panel 212. In another
embodiment, each of the nodes 218 is formed from a bundle of fiber optic strands.
[00104] The configuration of the nodes 218 including layout, shape and size
may be widely varied. With regards to layout, the nodes 218 may be based on a Polar
or Cartesian coordinate system (or some other coordinate system). With regards to
shape, any shape including for example standard shapes such as circles, squares,
rectangles, triangles, may be used. With regards to size, the nodes 218 may be larger

than a finger or stylus, about the same size as a finger or stylus, or smaller than a
finger or stylus. In one embodiment, the nodes 218 are set up similarly to the
electrodes 205 of the electrode layer 204, i.e., the nodes 218 have generally the same
layout, number, size and shape as the electrodes 205. In another embodiment, the
nodes are set up differently. For example, the nodes 218 may have a different layout,
different number, different shape and/or different size when compared to the
electrodes 205.
[00105] In the illustrated embodiment, the touch pad 200 is circular and the
nodes 218 are embodied as distinct angular segments (e.g., pie shaped). Any number
of angular segments may be used. The number generally depends on the desired
resolution of the illuminating surface. In this particular embodiment, the resolution of
the light panel 202 is low and therefore each of the angular segments cover a plurality
of sensing electrodes 205.
[00106] In one embodiment, all the LEDs 214 are powered at the same time to
produce a fully illuminated touch pad 200. This may be analogous to backlighting.
In another embodiment, the LEDs 214 are powered in accordance with the
capacitance changes measured by each of the electrodes 205. For example, the
segments above the detected area may be illuminated while the segments above the
undetected areas may be turned off. This provides indication to the user as to their •
exact location of the touch surface. In yet another embodiment, selected segments
may be illuminated to encourage a user to place their finger in a particular area of the
touch pad.
[00107] Although only a single light panel 202 is shown, it should be
appreciated that this is not a limitation and that additional light panels may be used.
For example, one or more light panels may be further positioned underneath the first
light panel described above. In one embodiment, each light panel in a group of light
panels is configured to distribute a different color. For example, three light panels
including a red, green and blue light panel may be used. Using this arrangement,
different colored segments may be produced. By controlling their intensity, almost
any color can be produced (mixed) at the touch surface. In another embodiment, each
light panel in the group of light panels may have a different orientation. For example,
the angularly segmented nodes of the light distribution panel may be rotated relative
to the other light panels so that they are placed at different positions about an axis

(e.g., partially overlapping and angularly offset). Using this arrangement, leading and
trailing illumination can be produced.
[00108] Figs 8A and 8B are diagrams of an illuminated touch pad 250, in
accordance with one embodiment of the present invention. The touch pad 250 is
similar to the touch pad 200 shown in Figs 7A and 7B in that it includes a light panel
252, electrode layer 254 and PCB 256. It differs from the touch pad of Figs. 7A and
7B in that the light panel 252 additionally includes inner side mounted LEDs 258 to
go along with the outer side mounted LEDs 260. It also differs from the touch pad of
Figs. 7A and 7B in that the light distribution panel 262 of the light panel 252 breaks
up each of the angularly segmented nodes of Fig. 7 into a pair of radially positioned
nodes including inner nodes 264 and outer nodes 266 that cooperate with the
respective LEDs 258 and 260. As a result, each of the nodes 264 and 266 represent
both an angular and radial position in the plane of the touch pad 250. This works
particularly well in touch pads with a circular shape. In addition, unlike the touch pad
of Figs. 7 A and 7B, the touch pad 250 also includes a mechanical button 268 at the
center of the touch pad 250. The mechanical button 268 may be illuminated with one
or more center LEDs 270.
[00109] In this embodiment, both the light distribution panel 262 and the
electrode layer 254 have an annular shape that creates a void at the center of the touch
pad 250. The void provides a space for placement of the extra light emitting diodes
258 and 270 as well as the mechanical button 268. As shown, the inner LEDs 258 are
disposed along the inner periphery of the light distribution panel 262 next to distinct
inner nodes 264 of the light distribution pane 262. Furthermore, the outer LEDs 260
are disposed along the outer periphery of the light distribution panel 262 next to
distinct outer nodes 266 of the light distribution panel 262.
[00110] Furthermore, the center LED 270 is disposed near the center of the
touch pad 250 underneath a translucent button cap 272 of the mechanical button 270.
The button cap 272 is movable trapped between a diffusing label layer 274 and a
spring loaded switch 276 that is also located near the center of the touch pad 250.
When the button cap is pressed, it moves against the actuator of the spring loaded
switch thereby generating a button event.
[00111] In the illustrated embodiment, the electrode layer 254, LEDs 258, 260
and 270 and mechanical switch 276 are all attached to the printed circuit board 256,
and operatively coupled to a controller 280 located on the backside of the PCB 256.

During operation, the controller 280 monitors the signals generated at the electrode
layer 254 and switch 276, and provides commands for controlling the LEDs 258, 260
and 270.
[00112] Figs 9 A and 9B are diagrams of an illuminated touch pad 300, in
accordance with one embodiment of the present invention. The touch pad-300 is
similar to the touch pad shown in Figs 8A and 8B in that it includes a light panel 302,
electrode layer 304 and PCB 306. It differs from the touch pad of Figs 8A and 8B in
that the light panel 302 includes a second set of inner LEDs 310 to go along with a
first set of inner LEDs 308 and a second set of outer LEDs 314 to go along with a
first set of outer LEDs 312. The first sets are located above the second sets, and may
be masked from one another to prevent bleed through.
[00113] It also differs from the touch pad of Figs. 8A and 8B in that the light
distribution panel 316 of the light panel 302 further breaks up each of the angularly
segmented nodes into four radially positioned nodes including inner nodes 318,
inner/middle nodes 320, outer/middle nodes 322, and outer nodes 324 that optically
cooperate with their respective LEDs. Particularly, the first set of inner LEDs 308 are
positioned to illuminate the inner nodes 318, the second set of inner LEDs 310 are
positioned to illuminate the inner/middle nodes 320, the first set of outer LEDs 312
are positioned to illuminate the outer nodes 324, and the second set of outer LEDs
314 are positioned to illuminate the outer/middle nodes 322.
[00114] In order to transmit light from the second set of inner LEDs 310 to the
inner/middle nodes 320, the inner/middle nodes 320 may include a light transmitting
portion 326 that extends underneath the inner nodes 318. In most cases, the light
transmitting potions 326 are optically separated from the inner nodes 318 so that the
light does not bleed into the inner nodes 318 when the light is passing through the
light transmitting portions 326.
[00115] In order to transmit light from the second set of outer LEDs 314 to the
outer/middle nodes 322, the outer/middle nodes 322 ma.y include a light transmitting
portion 328 that extends underneath the outer nodes 324. In most cases, the light
transmitting potions 328 are optically separated from the outer nodes 324 so that the
light does not bleed into the outer nodes 324 when the light is passing through the
light transmitting portions 328. Alternatively, a light pipe may be used.
[00116] The light distribution panel 316 may be embodied in a variety of ways.
In one embodiment, the light distribution panel 316 includes an outer translucent ring

disposed over and around a stepped outer/middle translucent ring, and an inner
translucent ring disposed over and around a stepped inner/middle translucent ring that
is adjacent the outer/middle ring. A masking layer may be placed between the
various rings to prevent bleed through.
[00117] Although-the touch has been described with only four radial segments,
it should be noted that any number of radial segments may be used as well as any
number of angular segments to obtain the desired resolution.
[00118] Fig 10 is a diagram of an illuminated touch pad 350, in accordance
with one embodiment of the present invention. The touch pad 350 includes various
layers including a diffuser layer 352, an opaque electrode layer 354, a PCB 356 and a
light panel 358. The light panel 358 may be embodied as a light panel including
pixilated light sources such as the one described in Fig. 6 or a light panel including a
light distribution panel and side mounted light sources such as the one described in
Fig. 7.
[00119] To elaborate, the light diffuser layer 352 is disposed over the
electrode layer 354, the opaque electrode layer 354 is disposed on the PCB 356, and
the PCB 356 is disposed over the light panel 358. In order to illuminate the light
diffuser layer 352, the electrode layer 354 includes one or more openings 360 through
which light may pass when emitted from the light panel 358. The openings 360 may
be the gaps that would normally be formed between the spatially separated electrodes
or they may be predetermined gaps around which the spatially separated electrodes
are positioned when printed on the PCB 356. Furthermore, in order to allow light to
pass through the PCB 356, the PCB 356 either is formed from a translucent material
or it also includes one or more openings 362 that correspond with the openings 360 of
the electrode layer 354. Moreover, the light panel 358 is typically laid out similarly
to the openings 360/362 so that each of the openings 360/362 includes an individual
light source 364 of the light panel 358. During operation, the light emitted from each
of the light sources 364 travels through the PCB 356 and through the electrode layer
354 where it illuminates either the diffuser layer 352 or a button cap 366.
[00120] Fig. 11 is a diagram of a light panel 400 that can be used in an
illuminated touch pad, in accordance with another embodiment of the present
invention. The light panel may, for example, correspond to any of those described
above. The light panel 400 includes a diffuser panel 402, one or more light
multiplexers 404 and a controller. The diffuser panel 402 is configured to diffuse

light as described above. The light multiplexer 404, which is operatively coupled to
the controller 406 and in optical communication with the diffuser panel 402, is
configured to illuminate the diffuser panel 402 in a controlled and pixilated manner
via commands from the controller 406.
[100121] The light multiplexer 404 includes a single light emitting device 408
and a plurality of light pipes 410 with corresponding light switches 412. The light
pipe 410 may for example include one or more optical fibers, and the light emitting
device 408 may be embodied in many different forms including for example one or
more individual LEDs or one or more LED arrays.
[00122] The first end of each light pipe 410 is optically connected to a different
point, node or region of the diffuser panel 402. The light pipes 410 can therefore
form a pixilated pattern of illumination points, nodes or regions across the light
diffuser panel 402. By way of example, the position of the light pipes 410 may be
based on Cartesian coordinates, Polar coordinates, or some other coordinate system.
The second and opposite end of each of the light pipes 410 is optically connected to a
distinct light switch 412. The light switches 412 are therefore dedicated to a
particular illumination point, node or region of the diffuser panel 402. Further, the
light switches 412 are all in optical communication with the light emitting device 408.
ha some cases, the light emitting device 408 extends across the light switches 412. hi
other cases, the light emitted by the light emitting device 408 is focused onto the
various light switches 412 via a lens or light guide 414.
[00123] Furthermore, the light switches 412 and light emitting device 408 are
operatively coupled to the controller 406. During operation, the controller 406
selectively controls the light emitted by the light emitting device 408 (e.g., color and
intensity), and at the same time selectively controls the opening and closing of the
light switches 412. As such, the illumination provided at the diffuser panel 402 can
be controlled in a pixilated manner using a single light emitting device 408. Any
number of switches can be opened or closed at any particular point in time to provide
the desired illumination pattern (by opening and closing different light switches,
various patterns can be created). When the light is turned on and a light switch is
opened, light is allowed to pass through the light switch into the associated light pipe,
which carries the light from the light switch to a distinct illumination point node or
region of the diffuser panel. When the light is turned on and a light switch is

closed, light is blocked from entering the light pipe and therefore no illumination is
provided at the corresponding point, node or region of the diffuser panel.
[00124] It should be noted that the multiplexer can include any number of
switches, and the light panel can include any number of multiplexers to arrive at the
desired resolution of the light panel.
[00125] Fig. 12 is a method 450 of operating an illuminated touch pad, in
accordance with one embodiment of the present invention. The method includes at
least blocks 452 and 454. In block 452, an object is detected over an input surface of
the touch pad. This may for example be accomplished with a capacitance sensing
device. In 454, at least a portion of the input surface proximate the location of the
detected object is illuminated. This may be accomplished with a light panel disposed
above or below the touch pad. As a result, the user will be informed where the object
is located within the sensing plane at all times.
[00126] In one embodiment, the input surface is broken up into illumination
regions, and whichever region is closest to the detected object is illuminated. By way
of example, and referring to Fig. 13 A, if the user places their finger over a single
angular segment of the distribution panel that particular angular segment is
illuminated. If the user simultaneously places their finger over multiple segments one
of two things may occur. In one implementation, both segments are illuminated. In
another implementation, only one of the segments is illuminated. In the later case, a
decision may be made as to which segment is the intended segment.
[00127] In another embodiment, the input surface is broken up into
illumination nodes or points (pixilated), and those points contained within and/or
surrounding the detected object area are illuminated. In one implementation, at least
the area adjacent the object is illuminated. By way of example, and referring to Fig.
14A, if the user places their finger over the input surface, illumination points adjacent
and surrounding the location of the finger are illuminated. In some cases, the
illumination points are only those points next to the finger (e.g., halo). In other cases,
the illuminated points extend away from the finger as for example in a star like
configuration.
[00128] The method may additionally include blocks 456 and 458. In block
456, a second object is detected over the input surface at the same time as the first
object. This may for example be accomplished with a multipoint capacitance sensing
device. In block 458, at least a portion of the input surface proximate the location of

the second detected object is illuminated. As a result, the user will be informed where
distinct multiple objects are located within the sensing plane at all times.
[00129] In one embodiment, the input surface is broken up into illumination
regions, and the regions closest to the detected objects are illuminated. By way of
example, and referring to Fig. 13B, when two fingers are placed over the input-
surface, two illumination segments in the location of the fingers are illuminated at the
same time.
[00130] In another embodiment, the input surface is broken up into
illumination nodes or points (pixilated), and those points contained within and/or
surrounding the detected objects are illuminated. By way of example, and referring
to Fig. 14B, when two finger are placed over the input surface, the area around both
fingers are illuminated at the same time.
[00131] Fig. 15 is a method 500 of operating an illuminated touch pad, in
accordance with one embodiment of the present invention. The method 500 generally
begins at block 502 where object sensing is performed. This may for example be
accomplished with a capacitive touch sensing device. In block 504, at least a portion
of the input surface is illuminated when an object is detected. In most cases, the
portion of the illuminated surface that is illuminated is a localized area disposed near,
around, and/or underneath the location of the detected object. The illuminated
portion may for example be one or more illumination points, nodes or regions. In
most cases, the portion is sized similarly to the size of the object. In the case of a
finger for example the illumination portion may cover an area similar to the detected
area of the finger.
[00132] In block 506, a determination is made as to whether or not the object is
moving. If the object is not moving, the method proceeds to block 508 where a
determination is made as to whether or not the object is still detected. If the object is
still detected, the method proceeds back to block 504 where the same portion of the
input surface is illuminated. If the object is no longer detected, the method proceeds
to block 510 where the illumination is stopped. This may occur immediately after
determining that an object is no longer detected, or it may occur after a period of time
(e.g., time out). Furthermore, the illumination may be stopped using an illumination
effect such as fading out. Thereafter, the method proceeds back to block 502.
[00133] Referring back to block 506, if the object is moving across the input
surface, the method proceeds to block 512 where motion characteristics of the object

are determined. The motion characteristics-may for example include acceleration;
direction, and the like. Thereafter, in block 514, the characteristics of the
illumination are adjusted based on one or more motion characteristics. Following
blocks 514, the method proceeds back to block 506.
[00134] In one embodiment, block, 514. includes moving the illumination area
in accordance with the location of the moving object. That is, the illuminated portion
follows the finger as the finger is moved about the input surface (i.e., the illumination
tracks object movement). As a result, the user always knows where the object is
located relative to the input surface. In some cases, block 514 may further include
providing directional indicators around the illuminated portion in order to indicate
previous and/or future locations of the object based on the motion characteristics of
the moving object (e.g., acceleration, direction, etc.).
[00135] Figs. 16A-16D illustrate one implementation where the illuminated
portion follows the motion of the finger as it is moved across the surface. In this
illustration, illuminated segments are configured to follow the motion of the finger as
it is moved across the surface. Fig. 16A illustrates the state when no objects are
detected. Fig. 16B illustrates the state when an object is detected, and the segment
underneath the object is illuminated. Fig. 16C illustrates the state where the
illuminated segment follows the moving finger. Fig. 16D illustrates one
implementation where the illuminated segment further includes a leading edge, body
and trailing edge. The leading edge indicates the direction of the motion, the body
indicates the current location of the finger, and the trailing edge indicates where the
finger has been.
[00136] The leading edge, body and trailing edge may have different
illumination profiles. For example, the leading edge may have a high intensity level,
the body may have a medium intensity level and the trailing edge may have a low
intensity level. Alternatively, the leading edge may have a low intensity level, the
body may have a high intensity level, and the trailing edge may have a low intensity
level. Alternatively or additionally, the colors of these components may differ. For
example, the leading edge may be red, the body may be orange and the trailing edge
may be yellow. Furthermore, the trailing edge may include an illumination tail. For
example, the trailing edge may be segmented into regions that go from higher
intensity to lower intensity levels (e.g., fades outwardly from body).

[00137] Figs. 17A-17D illustrate another implementation where the illuminated
portion follows the motion of the finger as it is moved across the surface. In this
illustration, the area around the finger is illuminated and configured to follow the
motion of the finger as it is moved across the surface. Fig. 17A illustrates the state
when no objects are detected. Fig. 17B illustrates the state when an object is
detected, and the area around the object is illuminated (e.g., halo). Fig. 17C
illustrates the state where the illuminated area follows the moving finger. Fig. 17D
illustrates one implementation where the illuminated area includes a body and a tail
(e.g., comet). The body surrounds the finger with illumination and the tail tapers
away from the body to a point. The tail trails the body as the body moves around the
input surface. The tail therefore indicates the previous location of the object. The tail
typically has a lower intensity level than the body. The intensity of the tail may even
vary from higher to lower intensity levels as for example over its length or from its
core to its edge.
[00138] Fig. 18 is a method 550 of operating an illuminated touch pad, in
accordance with one embodiment of the present invention.. The method 500 generally
begins at block 552 where object sensing is performed. In block 554, the state of the
touch pad is determined. The states may for example be selected from a selection
state, tracking state or gesture state. In a selection state, the touch pad is set up for
receiving selection inputs from the user (e.g., acts like a button). In a tracking state,
the touch pad is set up to track a finger as it is moved about the input surface. In a
gesture state, the touch pad is set up to receive various gesture inputs from the user.
[00139] In block 556, the input surface is illuminated based on the state of the
touch pad. As a result, the user is alerted to the current state of the touch pad, and
therefore the type of inputs that can be made. By way of example, each state may
include a different illumination profile. An illumination profile defines the
illumination characteristics of the illumination to be provided. The illumination
characteristics include for example intensity and/or color and/or illumination effects
(e.g., fading, blinking, rastering, etc). In one example, a first state includes a first
illumination profile (e.g., first color), a second state includes a second illumination
profile (e.g., second color), and the third state includes a third illumination profile
(e.g., third color).
[00140] Referring to 19 A, one example of determining state will be described.
In blocks 570 and 574, one or more objects are detected. In block 572, if a single

object is detected, the touch pad is placed in a tracking slate such lhat object motion is
used to perform tracking operations. In block 576, if multiple objects are detected,
the touch pad is placed in a gesturing state such that object motion is used to perform
gesturing operations. By way of example, and referring to Figs. 20 A and 20B, when
a single finger is detected, the touch pad is placed in a first state, and when multiple
fingers are detected, the touch pad is placed in a second state.
[00141] Referring to 19B, one example of illuminating based on states will be
described. In blocks 580 and 582, a determination is made as to whether the touch
pad is in a tracking state or a gesturing state. In block 584, if the touch pad is in a
tracking state, the touch pad is illuminated with a first illumination profile. In block
586, if the touch pad is in a gesturing state, the touch pad is illuminated with a second
illumination profile that is different than the first illumination profile. The
illumination profiles contain illumination information such as color, intensity and
effects (e.g., blinking, fading, etc). By way of example, and referring to Figs. 20A
and 20B., when the touch pad is in a first state, the touch pad is illuminated with a first
color or intensity, and when the touch pad is in a second state, the touch pad is
illuminated with a second color or intensity.
[00142] In one embodiment, the method of changing illumination based on
states of the touch pad may be further developed. For example, the method may
include capturing a first touch image; determining the touch mode based on the first
touch image; and illuminating the touch surface based on the first touch mode. The
method may also include capturing a second touch image; determining the touch
mode based on the second touch image; determining if the touch mode changed
between the first and second touch images; if the touch mode stayed the same,
comparing the first and second touch images and performing a control function based
on the comparison; and if the touch mode changed, illuminating the touch surface
based on the second touch mode. The method additionally includes capturing a third
touch image; determining the touch mode based on the third touch image;
determining if the touch mode changed between the second and third touch images; if
the touch mode stayed the same, comparing the second and third touch images and
performing a control function based on the comparison; and if the touch mode
changed, illuminating the touch surface based on the third touch mode.
[00143] Fig. 21 is a method 600 of operating an illuminated touch pad, in
accordance with one embodiment of the present invention. The method generally

begins at block 602 where an object is detected. In block 604, a least a portion of the
input surface is illuminated when the object is detected. In block 606, z
characteristics of the object are determined. The z characteristics may include the
pressure being exerted on the input surface by the object, and/or the location of the
object in the z direction relative to the x-y input surface (e.g., how close the object is
to the x-y plane). In block 608, the illumination characteristics are adjusted based on
the z-characteristic. For example, the color and/or intensity of the illumination may
be adjusted based on the z height or pressure.
[00144] By way of example, and referring to Figs. 22A and 22B, the entire
touch pad may be illuminated when the object is detected, and the intensity of the
illumination may be increased when an object is closer or exerts increased pressure
relative to the touch surface, and the intensity may be decreased when an object is
further away or exerts decreased pressure relative to the touch surface. Alternatively,
only a portion of the touch pad may be illuminated (as for example a segment or the
area directly adjacent the finger) and the intensity of the illumination may be
increased when an object is closer or exerts increased pressure relative to the touch
surface, and the intensity may be decreased when an object is further away or exerts
decreased pressure relative to the touch surface.
[00145] Fig. 23 is a method 700 of operating an illuminated touch pad, in
accordance with one embodiment of the present invention. The method generally
begins at block 702 where an object sensing is performed. In block 704, at least a
portion of the input surface is illuminated in response to the sensed object. For
example, a segment or the area around a finger may be illuminated. Thereafter in
block 706, a rotational user input is detected over the input surface. For example, in
the case of a circular touch pad, the rotational user input may be the user swirling
their finger around the circular touch pad. In some cases, this may include
determining an acceleration value pertaining to the rotational user input. In block
708, the input surface is illuminated in accordance with the rotational user input. For
example, the region of the touch pad underneath the finger is illuminated as the user
rotates their finger around the circular touch pad. In some cases, this may include
moving through illumination segments, nodes or points, based on at least the
acceleration value, whereby the acceleration value specifies a degree of acceleration
associated with the rate at which said moving through illumination segments, nodes
or points is to be achieved.

[00146] Fig. 24 is a method 800 of operating an illuminated touch pad, in
accordance with one embodiment of the present invention. The method generally
begins at block 802 where at least a portion of the input surface is illuminated with a
first illumination profile when an object is detected proximate the input surface.
Following block 802, the method proceeds to block 804 where the illumination of
illuminated portion of the input surface changes when the object is moved. For
example, the intensity of the illumination may be varied based on the acceleration of
the moving object. For example, the intensity may be increased with increased
acceleration and the intensity may be decreased with decreased acceleration. In
another embodiment, thresholds are used. For example, a first intensity level may be
used for high acceleration, a second intensity level may be used for low acceleration,
and a third intensity level may be used for no acceleration (stationary).
[00147] By way of example and referring to Fig. 25, low intensity illumination
is provided when a touch is first detected, medium intensity illumination is provided
when the object is slowly moved around the input surface (e.g., low acceleration), and
high intensity illumination is provided when the object is quickly moved around the
input surface (e.g., high acceleration). Alternatively, the intensity may continuously
vary according to the acceleration of the object.
[00148] While this invention has been described in terms of several preferred
embodiments, there are alterations, permutations, and equivalents, which fall within
the scope of this invention.
[00149] For example, although the invention was primarily directed at touch
pads, it should be pointed out that this is not a limitation and that invention may be
applied to other touch sensing devices as for example touch sensitive housings and
touch sensing palm rests.
[00150] It should also be noted that there are many alternative ways of
implementing the methods and apparatuses of the present invention.
[00151] For example, different areas of the touch pad may be illuminated with
different illumination profiles at the same time. By way of example, the touch pad
may be segmented into illuminated quadrants that correspond to button functionality
of the touch pad such as menu, play/pause, forward and reverse.
[00152] Furthermore, the touch pad may be used as an indicator for a handheld
computing device such as a media player. For example, the touch pad may be

configured to ebb in and out when the device is in sleep mode or vary from high to
low intensity based on the battery level.
[00153] Moreover, the touch pad may be used as a timer or clock. In the case
of a clock, the touch pad may include segments corresponding to the position of a
clock, and the segments can be illuminated in a controlled manner to indicate the
current time. For example, to indicate 12:30, a 12 o'clock segment may be
illuminated with a first illumination profile and 6 o'clock segment may be illuminated
with a second illumination profile. In the case of a timer, the touch pad may be used
to show how much time is left in a playing media item such as a song. For example,
the entire touch pad may be illuminated when the song starts and consecutive
segments may be turned off as the song plays. When the song is over, the touch pad
is no longer illuminated. Alternatively, consecutive segments may be turned on as
the song plays until the song is over and the touch pad is fully illuminated. The may
be useful in a media player such as a music player.
[00154] In addition, the illumination of the touch pad may be further controlled
by a sensor such as a light sensor. The light sensor measures the ambient light level,
and the intensity of the illumination is adjusted based on the ambient light level.
[00155] It is therefore intended that the following appended claims be
interpreted as including all such alterations, permutations, and equivalents as fall
within the true spirit and scope of the present invention.

CLAIMS
1. An illuminated input device, the illuminated input device comprising:
an object sensing mechanism capable of sensing a user input over an input
surface; and
a visual feedback system configured to illuminate the input surface in
association with a user input.
2. The input device as recited in claim 1 further including a controller that is
operatively coupled to the object sensing mechanism and the visual feedback system,
the controller monitoring signals generated at the object sensing mechanism, the
controller selectively controlling the illumination provided by the visual feedback
system
3. The input device as recited in claim 1 wherein the object sensing mechanism
includes a touch panel capable of sensing touches in localized areas of the input
surface, and wherein the visual feedback system includes a light panel capable of
emitting light in localized areas of the input surface.
4. The input device as recited in claim 3 wherein the touch panel is optically
transmissive, and wherein the optically transmissive touch panel is disposed over the
light panel.
5. The input device as recited in claim 4 wherein the optically transmissive touch
panel is a touch screen.
6. The input device as recited in claim 5 wherein the light panel includes a
plurality of light sources and a light distribution panel for distributing the light from
the light sources to localized areas of the input surface.
7. The input device as recited in claim 5 wherein the light panel includes
pixilated light emitting points that are dispersed across the input surface.
8. The input device as recited in claim 3 wherein the touch panel is disposed
underneath the light panel, the light panel allowing touch sensing to occur
therethrough.
9. The input device as recited in claim 8 wherein the touch panel includes a
plurality of spatially separated electrodes positioned on a printed circuit board.
10. The input device as recited in claim 9 wherein the light panel includes
pixilated light emitting points that are dispersed across the input surface.

11. The input device as recited in claim 9 wherein the light panel includes a
plurality of light sources and a light distribution panel for distributing the light from
the light sources to localized areas of the input surface.
12. The input device as recited in claim 11 wherein the light panel includes a
plurality of side mounted light sources and a light distribution panel for distributing
the light from the side mounted light sources to localized areas of the input surface,
the side mounted light sources being disposed around the side of the light distribution
panel and configured to direct light into different portions of the light distribution
panel, the light distribution panel being configured to redirect the light made incident
thereon to the input surface.
13. The input device as recited in claim 12 wherein the light distribution panel is
broken up into a plurality of nodes, each node including its own side mounted light
source.
14. The input device as recited in claim 13 wherein the light distribution panel is
circular and wherein the nodes are embodied as distinct angular and radial segments.
15. The input device as recited in claim 3 wherein the light panel comprises:
a diffuser panel configured to diffuse light;
one or more light multiplexers configured to illuminate the diffuser panel in a
controlled manner; and
a controller operatively coupled to the one or more light multiplexers.
16. The input device as recited in claim 3 wherein the touch panel is based on
capacitance.
17. The input device as recited in claim 3 further including a light diffuser
configured to diffuse the light being emitted by the light panel.
18. The input device as recited in claim 1 further including an illuminated button
located at the center of the input surface.
19. A method of operating an input device, comprising:
sensing an object over an input surface; and
illuminating at least a portion of the input surface when an object is sensed.
20. The method as recited in claim 19 wherein a localized area of the input
surface is illuminated proximate the location of the object.
21. The method as recited in claim 20 wherein the localized area is a segmented
region of the input surface, the illuminated segmented region being the segmented
region closest to the object.

22. The method as recited in claim 20 wherein the localized area are a collection
of illumination points that at least surround the object.
23. The method as recited in claim 20 further comprising:
sensing a second object over the input surface at the same as the first object;
illuminating a second localized area of the input surface proximate the
location of the second object, the first and second localized areas being illuminated at
the same time.
24. The method as recited in claim 20 further comprising:
determining whether or not the object is moving across the input surface;
determining motion characteristics of the moving object; and
adjusting the illumination based on one or more motion characteristics.
25. The method as recited in claim 24 wherein the localized area moves in
accordance with the moving object.
26. The method as recited in claim 25 further including providing indicators
associated with the motion characteristics.
27. The method as recited in claim 25 wherein the indicators include directional
indicators.
28. The method as recited in claim 25 wherein the localized area includes a
leading edge, a body and a trailing edge.
29. The method as recited in claim 25 wherein the localized area includes a body
and a tail.

30. The method as recited in claim 20 further comprising:
detecting a rotational user input over the input surface; and
rotating the localized area in accordance with the rotational user input.
31. The method as recited in claim 19 further comprising:
terminating the illumination when an object is no longer sensed;
32. The method as recited in claim 19 further comprising:
if the sensed object is stationary, providing a first illumination;
if the sensed object is moving, providing a second illumination; and
if the object is no longer sensed, maintaining the illumination in the last
sensed location for a predetermined amount of time.
33. The method as recited in claim 19 further comprising:
determining z characteristics of the object; and
adjusting the illumination based on the z characteristics.

34. The method as recited in claim 33 wherein the intensity of the illumination is
increased when the object is moved towards the input surface, and wherein the
intensity of the illumination is decreased when the object is moved away from the
input surface.
35, The method as recitedin claim 19 wherein the entire input surface is
illuminated.
36. A method of operating an input device, comprising:
detecting a user input over the input surface;
determining an input state of the input device based on the user input;
illuminating the input surface based on the input state of the input device, each
input state having a different illumination profile.
37. The method as recited in claim 36 wherein said determining includes analyzing
the user input for characteristics indicative of tracking or gesturing and categorizing
the user input as a tracking or gesture input.
38. The method as recited in claim 36 wherein the input device is placed in a
tracking state when a single object is detected, and wherein the input device is placed
in a gesture state when multiple objects are detected.
39. The method as recited in claim 38 wherein the input device is illuminated with a
first illumination profile when the input device is placed in the tracking state, and '
wherein the input device is illuminated with a second illumination profile when the
input device is placed in the gesturing state.
40. The method as recited in claim 36 wherein the entire input surface is
illuminated.
41. The method as recited in claim 36 wherein only a localized area of the input
surface is illuminated.
42. A method of operating an input device, comprising: .
illuminating at least a portion of an input surface when an object is detected
over the input surface; and
adjusting the illumination when the object is moved over the input surface.
43. The method as recited in claim 42 wherein the illumination is varied based on
the acceleration of the object.
44. The method as recited in claim 43 wherein the illumination intensity is
increased with increased acceleration, and wherein illumination intensity is decreased
with decreased acceleration.

45. The method as recited in claim 43 wherein a first intensity level is used for high
acceleration, a second intensity level is used for low acceleration, and a third intensity
level is used for no acceleration.
46. The method as recited in claim 42 wherein the entire input surface is
illuminated.
47. The method as recited in claim 42 wherein only a localized area of the input
surface is illuminated.

Improved feedback
mechanisms for touch pads are
disclosed. One aspect relates to devices
capable of illuminating the touch
sensitive surface of the touch pad.
Another aspect relates to methods for
providing visual feedback at the touch
pad.

Documents:

02583-kolnp-2008-abstract.pdf

02583-kolnp-2008-claims.pdf

02583-kolnp-2008-correspondence others.pdf

02583-kolnp-2008-description complete.pdf

02583-kolnp-2008-drawings.pdf

02583-kolnp-2008-form 1.pdf

02583-kolnp-2008-form 3.pdf

02583-kolnp-2008-form 5.pdf

02583-kolnp-2008-international publication.pdf

02583-kolnp-2008-international search report.pdf

02583-kolnp-2008-pct request form.pdf

2583-KOLNP-2008-(09-05-2014)-ABSTRACT.pdf

2583-KOLNP-2008-(09-05-2014)-CLAIMS.pdf

2583-KOLNP-2008-(09-05-2014)-CORRESPONDENCE.pdf

2583-KOLNP-2008-(09-05-2014)-DESCRIPTION (COMPLETE).pdf

2583-KOLNP-2008-(09-05-2014)-DRAWINGS.pdf

2583-KOLNP-2008-(09-05-2014)-FORM-1.pdf

2583-KOLNP-2008-(09-05-2014)-FORM-2.pdf

2583-KOLNP-2008-(09-05-2014)-OTHERS.pdf

2583-KOLNP-2008-(09-05-2014)-PA.pdf

2583-KOLNP-2008-(13-01-2014)-CORRESPONDENCE.pdf

2583-KOLNP-2008-(19-09-2011)-CORRESPONDENCE.pdf

2583-KOLNP-2008-(19-09-2011)-PA.pdf

2583-KOLNP-2008-(22-05-2013)-CORRESPONDENCE.pdf

2583-KOLNP-2008-(22-05-2013)-OTHERS.pdf

2583-KOLNP-2008-(22-11-2013)-ABSTRACT.pdf

2583-KOLNP-2008-(22-11-2013)-ANNEXURE TO FORM 3.pdf

2583-KOLNP-2008-(22-11-2013)-CLAIMS.pdf

2583-KOLNP-2008-(22-11-2013)-CORRESPONDENCE.pdf

2583-KOLNP-2008-(22-11-2013)-DESCRIPTION (COMPLETE).pdf

2583-KOLNP-2008-(22-11-2013)-DRAWINGS.pdf

2583-KOLNP-2008-(22-11-2013)-FORM-1.pdf

2583-KOLNP-2008-(22-11-2013)-FORM-2.pdf

2583-KOLNP-2008-(22-11-2013)-GPA.pdf

2583-KOLNP-2008-(22-11-2013)-OTHERS.pdf

2583-KOLNP-2008-(22-11-2013)-PETITION UNDER RULE 137.pdf

2583-KOLNP-2008-CORRESPONDENCE 1.1.pdf

2583-KOLNP-2008-FORM 3 1.1.pdf

2583-kolnp-2008-form-18.pdf

abstract-2583-kolnp-08.jpg


Patent Number 264414
Indian Patent Application Number 2583/KOLNP/2008
PG Journal Number 01/2015
Publication Date 02-Jan-2015
Grant Date 29-Dec-2014
Date of Filing 25-Jun-2008
Name of Patentee APPLE INC.
Applicant Address 1 INFINITE LOOP CUPERTINO, CA 95014
Inventors:
# Inventor's Name Inventor's Address
1 NG. STANLEY C. 1380 MONTCLAIRE WAY, LOS ALTOS, CA 94024
2 KERR, DUNCAN ROBERT 2600 18TH STREET, #15, SAN FRANCISCO CA 94110
PCT International Classification Number G06F 3/033
PCT International Application Number PCT/US2006/045683
PCT International Filing date 2006-11-28
PCT Conventions:
# PCT Application Number Date of Convention Priority Country
1 60/755656 2005-12-30 U.S.A.
2 11/394493 2006-03-31 U.S.A.