Electronic device with haptic feedback

Information

  • Patent Grant
  • 8378979
  • Patent Number
    8,378,979
  • Date Filed
    Tuesday, January 27, 2009
    15 years ago
  • Date Issued
    Tuesday, February 19, 2013
    11 years ago
Abstract
Haptic feedback may be provided to a user of an electronic device, such as an electronic book reader device, to confirm receipt of user input or otherwise convey information to the user. The haptic feedback may be provided more quickly than a display update time of a display of the electronic device. Different patterns, durations, and/or intensities of haptic feedback may be used in response to different events.
Description
BACKGROUND

Portable electronic devices, such as gaming devices, mobile telephones, portable media players, portable digital assistants (PDAs), electronic book (eBook) reader devices, and the like, are becoming increasingly popular. These devices often include a screen for visually conveying information to a user. Users of such portable electronic devices can provide input via a user interface to, for example, select an object or text on the screen. This user input is typically manifest visually on the screen (e.g., by highlighting the text or object), thereby providing the user with feedback confirming that the user input was received and processed. In devices employing a cathode ray tube (CRT) display, a liquid crystal display (LCD), or other display with a relatively short display update time, the user receives this visual confirmation almost immediately after providing the input.


However, in devices employing a display with a longer display update time, such as an electronic paper display, there may be a perceptible delay between when a user provides input and when that input is visually displayed on the screen. This delay can be confusing to a user, because they are not sure whether the device registered their input. Some electronic devices have attempted to mitigate this uncertainty by providing an audible confirmation of the user input, such as by playing a sound effect in response to receipt of user input. However, users may not associate the sound with the action they just input. Also, sound may not be able to adequately convey prolonged inputs, such as selection of a string of text.





BRIEF DESCRIPTION OF THE DRAWINGS

The detailed description is set forth with reference to the accompanying figures. In the figures, the left-most digit(s) of a reference number identifies the figure in which the reference number first appears. The use of the same reference numbers in different figures indicates similar or identical items.



FIG. 1 is a schematic diagram of an illustrative electronic device, in the form of an electronic book (eBook) reader device, that may implement haptic feedback techniques described herein.



FIG. 2 is a schematic diagram of illustrative subsystems of the eBook reader device of FIG. 1.



FIG. 3 is a schematic diagram showing illustrative haptic feedback devices usable with the eBook reader device of FIG. 1 to implement the haptic feedback techniques described herein.



FIG. 4 is a schematic diagram showing a list of illustrative haptic profiles usable to provide different haptic feedback responses.



FIG. 5 is a schematic diagram graphically depicting four illustrative haptic signatures of different haptic feedback responses.



FIG. 6 is a flowchart of an illustrative method of providing haptic feedback using an electronic device.





DETAILED DESCRIPTION

Overview


As discussed above, for electronic devices employing a display with a relatively long display update time (e.g., update times of at least about 15 milliseconds), such as an electronic paper display, there may be a perceptible delay between when a user provides input and when that input is visually presented on the display. This delay can be confusing to a user, because they are unsure whether the device registered their input. As used herein “display update time” refers to the time it takes to update or refresh data displayed on a display screen. Display update time may include a full display frame update, in which all or substantially all of the data displayed on a display screen is refreshed, or a partial display frame update, in which only a portion (e.g., a menu or other localized region) of the display screen is refreshed.


This disclosure describes various illustrative techniques for providing haptic feedback to a user of an electronic device, such as an electronic book (eBook) reader device, to confirm receipt of user input, to notify the user of a system event, or otherwise convey information to the user. Haptic feedback refers to feedback imparted to the user via the sense of touch (e.g., by applying forces, vibrations and/or motions to the user). In at least some implementations, the haptic feedback may be provided more quickly than a display update time of the electronic device.


In some implementations, different patterns, durations, and/or intensities of haptic feedback may be used. These different patterns, durations, and/or intensities of haptic feedback are referred to herein as “haptic signatures.” Different haptic signatures may be applied to different events. The haptic signature used for each event may be chosen to simulate a real world activity. For example, user input event to highlight a string of text may be met by a haptic signature characterized by constant, low amplitude vibration, to simulate the sensation of dragging an actual highlighter across a piece of paper.


In some implementations, the haptic feedback may be accompanied by audible feedback. For example, in the case of highlighting a string of text, the electronic device may also play a sound representative of dragging a highlighter across a piece of paper, further enhancing the user experience.


Haptic feedback may be provided in response to, among other things, user input via a button, joystick, knob, thumb wheel, lever, touch screen, or other user interface mechanism. The haptic feedback may be provided in response to all or only a subset of user inputs. Additionally or alternatively, haptic feedback may be provided in response to other non-user input related events, such as system events and program events.


While haptic feedback techniques are described in the context of portable electronic devices having displays with relatively long display update times, such as electronic paper displays, the haptic feedback techniques described herein are also applicable to devices having displays with shorter display update times. In that case, the haptic feedback may be provided substantially contemporaneously with visual feedback. Moreover, while the techniques are described in the context of portable electronic devices, the techniques are also applicable to other electronic devices, such as laptop and desktop computers, set-top-boxes, game consoles or controls, televisions, appliances, audio/video equipment, and the like.


EXAMPLE ELECTRONIC DEVICE


FIG. 1 shows an example of one illustrative electronic device 100 capable of providing haptic feedback to a user. In the illustrated implementation, the device 100 is embodied as a handheld electronic book (eBook) reader device. However, virtually any other type of electronic device may be used and may be configured to provide haptic feedback to a user as described herein. By way of example and not limitation, other electronic devices that could be configured to employ the haptic feedback techniques described herein include gaming devices, mobile telephones, portable media players, portable digital assistants (PDAs), personal computers (desktop and laptop), netbooks, electronic book reader devices, set-top-boxes, televisions, appliances, audio/video equipment, and the like.


As shown in FIG. 1, the eBook reader device 100 has a body or housing 102, a display 104 for displaying information to a user, and a user interface. In this example, the display 104 comprises an electronic paper display, such as those made by eInk Corporation of Cambridge, Mass. The user interface comprises a variety of mechanism for users to interact with the device 100 including a keypad 106, an on/off slider 108, multiple buttons 110, and a user interface cluster 112 including a joystick 114 and a pivot button 116. In some implementations, the display 104 may also comprise a user interface mechanism in the form of a touch screen.


The user interface allows users to, among other actions, display and navigate through a collection of eBooks, web pages, audio files, video files, games, programs, and/or other electronic items. As used herein, the term eBook includes electronic copies of books, magazines, newspapers, maps, publications, and other at least partially text-based electronic documents. Users may also be able to select, highlight, and/or modify text or other items on the display. Users may be able to compose email or other items, perform searches of locally stored and/or remote information, browse the internet, or the like. Users may also be able to manage the content stored on the device to download or install new items and/or delete existing items. These and other actions are possible using the various user input mechanisms of the user interface.


In the illustrated eBook reader device 100, the keypad 106 includes a plurality of alphabetic, numeric, and/or function keys for entry of user input. The on/off slider 108, as its name suggests, allows users to selectively turn at least some systems of the device 100 on or off. However, power to some systems of the device may be controlled independently of the on/off switch. Buttons 110 allow users to turn pages forward and back within an eBook and go to a home page. The joystick 114 allows users to select text and items on the display. Pivot button 116 allows users to open menus and undo selections.


In other implementations, user interfaces of electronic devices may include any combination of these and other user input mechanisms. Other user inputs that may be present include, for example, scroll wheels, thumb wheels, thumb sticks, sensors that detect and register movement of a user's thumb or finger, tactile sensors, touch screens, or any other conventional user input mechanism.


Using an electronic paper screen as the display 104 reduces the power consumption of the electronic device 100, thereby extending battery life. Electronic paper displays also provide a very easy to read display that resembles an actual printed page of text. However, as mentioned above, electronic paper displays tend to have a longer display update time than other types of displays, such as cathode ray tube (CRT) displays and liquid crystal displays (LCDs). For example, electronic paper displays may have a display update time greater than about 15 milliseconds. Typically, electronic paper displays have display update times greater than about 100 milliseconds, and in some implementations, displays may have display update times greater than about 250 milliseconds.


With display update times greater than about 15 milliseconds, there is a perceptible delay between a time when a user requests an action via the user interface and a time when that action is displayed on the display 104. This delay may cause users to be uncertain whether the electronic device received an input. In some instances, users unsure about whether their input was received may reenter their input several times, possibly resulting in multiple and/or unintended actions being performed by the electronic device 100. Haptic feedback may be used to confirm the user's input more quickly, thereby removing any uncertainty in the mind of the user about whether an input was received. The haptic feedback may be provided more quickly than the display update times of the display 104; for example, in less than about 15 milliseconds in most cases.


Illustrative Electronic Device Subsystems


In addition to the exterior components shown in FIG. 1, the eBook reader device 100 also has various internal components, which are referred to generally as eBook reader subsystems 200. In one implementation, the subsystems 200 include memory 202 and a processing unit 204. The processing unit 204 interacts with the memory 202 to facilitate operation of the eBook reader device 100. The memory 202 may be used to store content 206, such as eBooks, audio and/or video media, or the like. The memory 202 also includes software programs or other executable modules 208 that may be executed by the processing unit 204. Examples of such programs or modules include indexing modules for indexing content, reader programs, control modules (e.g., power management), network connection software, an operating system, display drivers, sensor algorithms, page turn detectors, and the like.


The eBook reader subsystems 200 also include a haptic feedback device 210 to provide haptic feedback in response to user input or other events. In various implementations, the haptic feedback device 210 may be configured to apply haptic feedback in the form of vibration, force feedback, and/or motion, depending on the specific type of haptic feedback device used. An audio output 212 may also be provided as an eBook reader subsystem to provide audible feedback in response to user input or other events. The haptic feedback device 210 and the audio output 212 may be used separately or in concert to enhance users' experiences by providing feedback to a user of the device 100.


The memory 202 may also store one or more haptic profiles 214 that define different patterns, durations, and/or intensities of haptic feedback to be used in response to different user inputs or other events. These different patterns, durations, and/or intensities of haptic feedback are referred to herein as “haptic signatures.” The haptic signature used for each event may be chosen to simulate a real world activity to provide a realistic user experience.


A haptic device driver 216 is provided to drive the haptic feedback device 210 according to one or more of the haptic profiles 214. In one specific example, the haptic device driver 216 may comprise a simple drive circuit, with a single n-doped filed effect transistor (NFET) used to switch the haptic feedback device 210 on and off. The circuit may have a low-side switch configuration. A Schottky diode may be placed across the haptic feedback device 210, to protect against inductive kick-back.


While the haptic device driver 216 is illustrated and described as a drive circuit, in other implementations, the haptic device driver 216 could alternatively be configured as a software module stored in memory 202 or as firmware.


The memory 202 may include volatile memory (such as RAM), nonvolatile memory, removable memory, and/or non-removable memory, implemented in any method or technology for storage of information, such as computer-readable instructions, data structures, program modules, or other data. Also, the processing unit 204 may include onboard memory in addition to or instead of the memory 202. Some examples of storage media that may be included in the memory 202 and/or processing unit 204 include, but are not limited to, random access memory (RAM), read only memory (ROM), electrically erasable programmable read only memory (EEPROM), flash memory or other memory technology, compact disk (CD-ROM), digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by the eBook reader device 100. Any such computer-readable media may be part of the eBook reader device 100.


Various instructions, methods and techniques described herein may be considered in the general context of computer-executable instructions, such as program modules, executed by one or more computers or other devices. Generally, program modules include routines, programs, objects, components, data structures, etc. for performing particular tasks or implementing particular abstract data types. These program modules can be implemented as software modules that execute on the processing unit 204, as hardware, and/or as firmware. Typically, the functionality of the program modules may be combined or distributed as desired in various embodiments. An implementation of these modules and techniques may be stored on or transmitted across some form of computer-readable media.


EXAMPLE HAPTIC FEEDBACK MECHANISMS


FIG. 3 is a diagram of the eBook reader device 100 with the front cover removed to schematically show an illustrative haptic feedback device 300 usable to provide haptic feedback. In this implementation, one surface of the haptic feedback device 300 is affixed to a sub-frame of the back of the housing 102 of the electronic device 100 to impart the haptic feedback to the back of the housing 102. An opposite face and remaining surfaces of the haptic feedback device 300 may be encapsulated in rubber, or other damping material, to help isolate the vibrations to the housing 102. Thus, haptic feedback may be imparted to the housing 102 of the device 100 in response to user input or other events.


In other implementations, the haptic feedback device 300 may be coupled to a particular user input device of the user interface, such that the haptic feedback is localized to the particular user input device. For example, the haptic feedback device could be coupled to any of the display 104, keypad 106, on-off slider 108, buttons 110, joystick 114, and/or pivot button 116 of the user interface. Moreover, multiple haptic feedback devices could be provided, with different haptic feedback devices coupled to different user inputs. In this manner, haptic feedback could be applied independently to each of the user input mechanisms of the user interface.


Various different haptic feedback mechanisms may be used as the haptic feedback device 300. In one implementation, the haptic feedback device comprises an electric field generator 300A, a magnetic mass (not shown) disposed therein. The electric field generator 300A is selectively activatable to drive the mass to generate a vibration. More specifically, the electric field generator 300A comprises a two-lead cylindrical device. The leads are connected to opposite ends of a coil, similar to a voice coil found in a speaker. Applying a voltage across the coil generates an electromagnetic field that attracts or repels the magnetic mass inside, thus producing mechanical vibrations from electrical drive signals provided by the haptic feedback device driver 216. The electric field generator 300A haptic feedback device has a mechanical resonance (i.e., a specific frequency at which it produces strong vibrations). A roll-off from this peak resonance is sharp on both sides (in the frequency domain). Thus, an intensity of the haptic feedback provided can be adjusted by adjusting a frequency at which the electric field generator 300A is driven.


In another implementation, the haptic feedback device 300 comprises an unbalanced rotatable mass and a motor 300B to drive the unbalanced rotatable mass. As the motor 300B spins, the unbalanced mass creates a vibration that can be used to apply haptic feedback.


In yet another implementation, the haptic feedback device 300 comprises a piezoelectric actuator 300C. Piezoelectric actuators take advantage of the ability of some materials (notably crystals and certain ceramics) to generate an electric potential in response to applied mechanical stress, and conversely to deform when an electric field is applied. Thus, a piezoelectric actuator 300C may serve the dual purpose of registering user input by emitting a signal when the actuator is deformed (e.g., a button is depressed, a joystick is moved, or a wheel is turned), as well as providing haptic feedback by deforming when an electric field is applied thereto. In some implementations, the piezoelectric actuator 300C may be used in combination with an electric field generator, such as electric field generator 300A, to create the electric field to cause the piezoelectric actuator to deform. Additionally or alternatively, the piezoelectric actuator may take the form of a piezoelectric motor.


In some implementations, the haptic feedback device 300 may be configured to provide haptic feedback in the form of vibration applied to the housing 102 and/or one or more user input mechanisms (104-116). However, in other implementations, haptic feedback devices could be configured additionally or alternatively to apply force feedback (i.e., resistive force or pushback on one or more user input mechanisms) or motion feedback (i.e., physical motion of one or more user input mechanisms). For example, the motor 300B could be used to apply resistive force and/or motion feedback to one or more user mechanisms (e.g., a thumb wheel, dial, knob, or other rotatable user input mechanism) of the user interface. With a suitable mechanical linkage, the motor 300B could apply resistive force and/or motion feedback to non-rotatable user input mechanisms as well.


These and other types of haptic feedback devices 300N may be used alone or in combination with one or more force transmission mechanisms, such as one or more levers or linkages, a pin-rod matrix, or the like, to generate haptic feedback in accordance with this disclosure. Several examples of other haptic feedback devices 300N that may be used to implement the techniques described herein include shape memory alloys (SMAs) or shape memory polymers (SMPs). By way of example and not limitation, several illustrative examples of SMAs that could be used include nickel-titanium compositions, Bio Metal Fiber and Bio Metal Helix series of materials, available from Toki Corporation of Tokyo, Japan. Also by way of example and not limitation, several illustrative examples of SMPs that could be used include N-isopropylacrylamide (NIPAM), methacryloyl chloride (MACL) which contains poly(epsilon-caprolactone) (PCL) segments and poly(cyclohexyl methacrylate) (PCHMA) segments, or the like.


EXAMPLE HAPTIC PROFILES

As discussed above, in at least some implementations, haptic feedback is applied according to a haptic signature specified in a haptic profile 214 stored in memory 202 of the electronic device 100. The haptic signature specifies the different pattern, duration, and/or intensity of haptic feedback to apply for a given event. The haptic profiles 214 define which haptic signature to use for various events.



FIG. 4 is a schematic diagram showing an illustrative list 400 of haptic profiles 214 stored in memory 202 of the electronic device 100. A haptic profile 214 exists for at least some and possibly all possible user input events. By way of example and not limitation, user input events that may have a haptic profile include highlighting an object, selecting an object, turning a page of an electronic item, opening a menu, deleting an item from memory, invoking a search, clipping an object, activating a link, among other events. In addition to user input events, other events, such as system events, program events, and the like, may also have corresponding haptic profiles to define a haptic signature to use upon the occurrence of the respective event.


For each haptic profile 214, the list 400 includes an event entry 402 specifying an event, and a haptic signature entry 404 to use upon occurrence of the specified event. In some implementations, the list 400 may also include a sound entry 406, designating a sound to play upon the occurrence of the specified event.


The first entry in the list 400 corresponds to an event in which a user highlights an object, such as a string of text. The haptic profile 214 for that entry specifies that the electronic device 100 is to apply haptic signature 1 and play audio file 1. If the electronic device 100 is configured to simulate a corresponding real world action, haptic signature 1 in this example may be reminiscent of the sensation of dragging a highlighter across a piece of paper and audio file 1 may be a recording of the sound of dragging a highlighter across a piece of paper. The combination of haptic and audible feedback further enhances the user experience.


The second through mth entries in the list may also include haptic signatures and/or audio clips to play in response to given events.


EXAMPLE HAPTIC SIGNATURES


FIG. 5 is a graphical illustration of four example haptic signatures 400 that may be used in response to various events. As discussed above, the haptic signatures 400 may comprise vibration feedback, force feedback, and/or motion feedback. The haptic signatures 400 generally define a duration, amplitude, and pattern of haptic feedback provided.


A first haptic signature 402 is associated with user input to highlight an object, such as text. The highlighting haptic signature 402 is characterized by a relatively low amplitude, continuous vibration designed to simulate the physical act of dragging a highlighter across a piece of paper.


A second haptic signature 404 is associated with user input to delete an item from memory of the device. The delete haptic signature 404 is characterized by a relatively larger magnitude vibration of short duration designed to simulate a chopping sensation.


A third haptic signature 406 is associated with user input to select and/or move an item. The select/move haptic signature 406 is characterized by a positive force feedback of medium amplitude, designed to simulate the force necessary to lift or hold an object. The duration of the force corresponds to the length of time a user holds down a button, joystick or other user input mechanism to select or move the object.


A fourth haptic signature 408 is associated with user input to turn a page of a eBook, change a track of an album, or navigate within another digital item. The page-turn haptic signature 408 is characterized by a vibration pattern of three cycles of moderate amplitude for each page turn request. The page-turn haptic signature 408 is designed to simulate the sensation of flipping pages of a physical book. The page-turn haptic signature 408 shown in FIG. 5 is shown repeated a second time following a pause, indicating that the user has requested a second page turn.


The four foregoing haptic signatures are merely illustrative of haptic signatures that may be used for the corresponding events. However, other haptic signatures could be used in response to these events. Moreover, these or any number of other haptic signatures may be used in response to other events.


EXAMPLE OPERATION


FIG. 6 is a flowchart of an illustrative method 600 that may, but need not necessarily, be implemented using the electronic device of FIGS. 1-3. The method 600 may alternatively be implemented using different electronic devices. Moreover, the electronic devices described herein may be used to implement different haptic feedback methods.


The method 600 begins, at 602, with receipt of a signal from a user interface of an electronic device, such as the eBook reader 100. The received signal designates a specific user input (e.g., highlight an object, turn a page, etc.), system event (e.g., a low batter alert, error message, etc.), program event (e.g., returning search results, specifying a beginning or end of playback, etc.), or other event.


At 604, the processing unit 204 or another program module 208 in memory 202 of the electronic device 100 compares the signal received from the user interface to a haptic profile 214 stored in memory of the electronic device 100. The haptic profile 214 specifies a haptic signature corresponding to the event designed by the received signal. In some instances, the haptic profile 214 (or another profile stored in memory) may also specify an audio file corresponding to the event designated by the received signal.


At 606, the haptic device driver 216 or a program module 208 in memory 202, determines a haptic signature, such as one of haptic signatures 402-408, to use in response to the received signal from the user interface, based on the comparison at 604.


At 608, the haptic device driver 216 drives the haptic feedback device 210 of the electronic device 100 to provide haptic feedback according to the haptic signature determined to be used. In the case of generating vibration haptic feedback, driving the haptic feedback device 210 may comprise sending a square wave signal with pulses having amplitude, duration, and/or patterns as specified in the haptic signature, to the haptic feedback device 210. The square wave signal has a base frequency corresponding to the peak resonant frequency of the haptic feedback device 210. Three different illustrative drive techniques that may be used to modulate the strength of the vibration generated by the haptic feedback device 210 are described below. However, other drive techniques may additionally or alternatively be used to drive the haptic feedback device 210.


The first technique is to generate the square wave pulse with a pulse-width-modulated (PWM) signal at a much higher frequency than the resonant frequency of the haptic feedback device 210. In this technique, modulation of the vibration strength is achieved by controlling the duty cycle of the PWM signal. The idea behind this method is to make use of “mechanical filtering” of the haptic feedback device 210. The haptic feedback device 210 does not react to the individual pulses of the PWM signal (because they are much higher than the haptic device's peak resonant frequency), but instead reacts to the average of all the pulses. Increasing or decreasing the duty cycle of the PWM signal changes the amount of energy delivered to the haptic device, thus affecting the strength of its vibration.


The second drive technique simply involves increasing or decreasing the amplitude of the square wave pulses, which correspondingly affects the strength of the haptic feedback device 210 vibration.


The third drive technique takes advantage of the roll-off from the peak resonance frequency that the haptic feedback device's mechanical vibration exhibits. By generating square wave pulses with larger or shorter widths than the ideal peak resonance width, the same effect of vibration strength modulation can be achieved.


With each of these drive techniques, pulses that create vibrations of different strengths can be strung together, one after another. This effectively creates unique tactile sensations that define the haptic signatures. Unique haptic signatures are mapped to events in the haptic profiles, as described above.


If, at 604, the haptic profile 214 or another profile stored in memory 202 specifies an audio file corresponding to the event designed by the received signal, at 610, the audio output 212 plays the audio file.


Then, at 612, the electronic device updates a display of the electronic device to reflect the received signal from the user interface. In the case of an electronic device with an electronic paper display or other display with a relatively long display update time, this update of the display occurs perceptibly after the haptic and/or audio feedback response. However, in electronic devices with CRT displays, LCDs, or other displays with relatively short display update times, the display may be updated substantially contemporaneously with the haptic and/or audio feedback response.


Certain acts in method 600 need not be performed in the order described, may be modified and/or may be omitted entirely, depending on the circumstances. For example, acts 610 and/or 612 may be omitted.


Moreover, any of the acts of any of the methods described herein may be implemented by a processor or other computing device based on instructions stored on one or more computer-readable media. Computer-readable media can include any available media that can be accessed by the electronic device. By way of example, and not limitation, computer-readable media may comprise volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data. Some specific examples of computer-readable media include, without limitation, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, DVD, or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can accessed by the electronic device. Combinations of the any of the above should also be included within the scope of computer-readable media.


CONCLUSION

Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described. Rather, the specific features and acts are disclosed as illustrative forms of implementing the claims. For example, the methodological acts need not be performed in the order or combinations described herein, and may be performed in any combination of one or more acts.

Claims
  • 1. A portable electronic book reader device comprising: a housing;an electronic paper display screen disposed in the housing, the electronic paper display screen having a display update time;a user interface coupled to the housing and configured to receive user input to be rendered on the display; anda haptic feedback device coupled to the housing and configured to provide haptic feedback in response to receipt of input to the user interface, the haptic feedback device having a response time that is less than the display update time.
  • 2. The portable electronic book reader device of claim 1, further comprising: memory storing a plurality of haptic profiles, each haptic profile defining a haptic signature to be used in response to a given user input event; anda haptic feedback device driver configured to drive the haptic feedback device to provide haptic feedback according to one of the plurality of haptic profiles.
  • 3. The portable electronic book reader device of claim 2, wherein the haptic feedback device comprises an electric field generator and a magnetic mass, the electric field generator being actuatable to drive the mass to generate a vibration.
  • 4. An electronic device comprising: a display having a display update time;a user interface configured to receive user input; anda haptic feedback device configured to provide haptic feedback in response to user input received at the user interface, the haptic feedback device having a responsiveness that is faster than the display update time of the display.
  • 5. The electronic device of claim 4, wherein the display comprises an electronic paper display.
  • 6. The electronic device of claim 4, wherein the display has a display update time greater than about 15 milliseconds, and wherein the haptic feedback device has a response time of at most about 15 milliseconds.
  • 7. The electronic device of claim 4, wherein the haptic feedback device is configured to provide haptic feedback in response to one or more of the following user inputs: user input to highlight an object on the display;user input to select an object on the display;user input to turn a page presented on the display;user input to open a menu on the display;user input to delete an item from memory of the device;user input invoking a search function;user input to clip text on the display; anduser input to activate a link on the display.
  • 8. The electronic device of claim 4, further comprising memory storing a plurality of haptic profiles, each haptic profile defining a haptic signature to be used in response to a given user input.
  • 9. The electronic device of claim 8, wherein each haptic signature defines a duration, amplitude, and pattern of haptic feedback provided.
  • 10. The electronic device of claim 8, further comprising a haptic feedback device driver configured to drive the haptic feedback device to provide haptic feedback according to one of the plurality of haptic profiles.
  • 11. The electronic device of claim 4, wherein the haptic feedback device is configured to provide haptic feedback according to one or more haptic profiles.
  • 12. The electronic device of claim 4, wherein the haptic feedback device is configured to provide one or more of the following forms of feedback: vibration feedback, force feedback, motion feedback.
  • 13. The electronic device of claim 4, wherein the haptic feedback device comprises an electric field generator and a magnetic mass, the electric field generator being selectively activatable to drive the mass to generate a vibration.
  • 14. The electronic device of claim 4, wherein the haptic feedback device comprises an unbalanced rotatable mass and a motor to drive the unbalanced rotatable mass.
  • 15. The electronic device of claim 4, wherein the haptic feedback device comprises a piezoelectric actuator.
  • 16. The electronic device of claim 4, wherein the haptic feedback device comprises a shape memory alloy or a shape memory polymer.
  • 17. The electronic device of claim 4, wherein the haptic feedback device is fixed to a housing of the electronic device to impart the haptic feedback to the housing.
  • 18. The electronic device of claim 4, wherein the haptic feedback device is coupled to a particular user input device of the user interface, such that the haptic feedback is localized to the particular user input device.
  • 19. The electronic device of claim 18, wherein the haptic feedback device is coupled to a joystick of the electronic device to provide localized haptic feedback to the joystick.
  • 20. The electronic device of claim 18, further comprising an audio output device configured to output a sound in response to received user input.
  • 21. One or more computer-readable media storing computer-executable instructions that, when executed, are configured to perform acts comprising: receiving a signal from a user interface of an electronic device, the signal designating a specific user input;comparing the signal received from the user interface to a haptic profile stored in memory of the electronic device, the haptic profile comprising a haptic signature corresponding to the signal received from the user interface;determining the haptic signature to use in response to the received signal from the user interface, based on the comparison; anddriving a haptic feedback device of the electronic device to provide haptic feedback according to the haptic signature determined to be used.
  • 22. The one or more computer-readable media of claim 21, wherein the haptic signature defines a duration, amplitude, and pattern of haptic feedback provided.
  • 23. The one or more computer-readable media of claim 21, storing instructions configured to perform the further act of updating a display of the electronic device based on the signal received from the user interface, wherein the haptic feedback mechanism has a shorter response rate than a display update time of the display, such that driving the haptic feedback mechanism is configured to occur prior to updating the display.
  • 24. The one or more computer-readable media of claim 21, wherein the received signal designates one or more of the following user inputs: user input to highlight an object on the display;user input to select an object on the display;user input to turn a page presented on the display;user input to open a menu on the display;user input to delete an item from memory of the device;user input invoking a search function;user input to clip text on the display; anduser input to activate a link on the display.
  • 25. The one or more computer-readable media of claim 21, storing instructions configured to perform the further act of outputting a sound based on the signal received from the user interface.
US Referenced Citations (306)
Number Name Date Kind
4622627 Rodriguez et al. Nov 1986 A
5544305 Ohmaye et al. Aug 1996 A
5566098 Lucente et al. Oct 1996 A
5623260 Jones Apr 1997 A
5659742 Beattie et al. Aug 1997 A
5663748 Huffman et al. Sep 1997 A
5710922 Alley et al. Jan 1998 A
5742905 Pepe et al. Apr 1998 A
5761485 Munyan Jun 1998 A
5765168 Burrows Jun 1998 A
5813017 Morris Sep 1998 A
5845278 Kirsch et al. Dec 1998 A
5847698 Reavey et al. Dec 1998 A
5892900 Ginter et al. Apr 1999 A
5940846 Akiyama Aug 1999 A
5956048 Gaston Sep 1999 A
5960411 Hartman et al. Sep 1999 A
5973681 Tanigawa et al. Oct 1999 A
6018575 Gross et al. Jan 2000 A
6034839 Hamming Mar 2000 A
6037954 McMahon Mar 2000 A
6041335 Merritt et al. Mar 2000 A
6047189 Yun et al. Apr 2000 A
6049796 Siitonen et al. Apr 2000 A
6064980 Jacobi et al. May 2000 A
6073148 Rowe et al. Jun 2000 A
6148340 Bittinger et al. Nov 2000 A
6154757 Krause et al. Nov 2000 A
6164974 Carlile et al. Dec 2000 A
6201771 Otsuka et al. Mar 2001 B1
6233318 Picard et al. May 2001 B1
6300947 Kanevsky Oct 2001 B1
6308320 Burch Oct 2001 B1
6313828 Chombo Nov 2001 B1
6331867 Eberhard et al. Dec 2001 B1
6335678 Heutschi Jan 2002 B1
6351750 Duga et al. Feb 2002 B1
6385596 Wiser et al. May 2002 B1
6401086 Bruckner Jun 2002 B1
6401239 Miron Jun 2002 B1
6442651 Crow et al. Aug 2002 B2
6449627 Baer et al. Sep 2002 B1
6457030 Adams et al. Sep 2002 B1
6466951 Birkler et al. Oct 2002 B1
6493734 Sachs et al. Dec 2002 B1
6496803 Ho et al. Dec 2002 B1
6529920 Arons et al. Mar 2003 B1
6542874 Walker et al. Apr 2003 B1
6560699 Konkle May 2003 B1
6642947 Feierbach Nov 2003 B2
6658623 Schilit et al. Dec 2003 B1
6685482 Hopp et al. Feb 2004 B2
6687878 Eintracht et al. Feb 2004 B1
6704733 Clark et al. Mar 2004 B2
6721869 Senthil Apr 2004 B1
6724403 Santoro et al. Apr 2004 B1
6725227 Li Apr 2004 B1
6735583 Bjarnestam et al. May 2004 B1
6744891 Allen Jun 2004 B1
6847966 Sommer et al. Jan 2005 B1
6904449 Quinones Jun 2005 B1
6912398 Domnitz Jun 2005 B1
6933928 Lilienthal Aug 2005 B1
6938076 Meyer et al. Aug 2005 B2
6985932 Glaser et al. Jan 2006 B1
6999565 Delaney et al. Feb 2006 B1
7007015 Nayak Feb 2006 B1
7009596 Seet et al. Mar 2006 B2
7010500 Aarnio Mar 2006 B2
7020654 Najmi Mar 2006 B1
7054914 Suzuki et al. May 2006 B2
7057591 Hautanen et al. Jun 2006 B1
7062707 Knauft et al. Jun 2006 B1
7089292 Roderick et al. Aug 2006 B1
7092116 Calaway Aug 2006 B2
7103848 Barsness et al. Sep 2006 B2
7107533 Duncan et al. Sep 2006 B2
7130841 Goel et al. Oct 2006 B1
7133506 Smith Nov 2006 B1
7135932 Quadir et al. Nov 2006 B2
7149776 Roy et al. Dec 2006 B1
7165217 Kondo Jan 2007 B1
7188085 Pelletier Mar 2007 B2
7209888 Frid-Nielsen et al. Apr 2007 B2
7237123 LeVine et al. Jun 2007 B2
7246118 Chastain et al. Jul 2007 B2
7249060 Ling Jul 2007 B2
7249324 Nakamura et al. Jul 2007 B2
7257577 Fagin et al. Aug 2007 B2
7287068 Eriksson et al. Oct 2007 B1
7290285 McCurdy et al. Oct 2007 B2
7298851 Hendricks et al. Nov 2007 B1
7304635 Seet et al. Dec 2007 B2
7310629 Mendelson et al. Dec 2007 B1
7313759 Sinisi Dec 2007 B2
7350704 Barsness et al. Apr 2008 B2
7355591 Sugimoto Apr 2008 B2
7386480 Sarig Jun 2008 B2
7398244 Keith Jul 2008 B1
7401286 Hendricks et al. Jul 2008 B1
7454238 Vinayak et al. Nov 2008 B2
7496767 Evans Feb 2009 B2
7506246 Hollander et al. Mar 2009 B2
7509270 Hendricks et al. Mar 2009 B1
7533152 Stark et al. May 2009 B2
7539478 Herley et al. May 2009 B2
7562032 Abbosh et al. Jul 2009 B2
7574653 Croney et al. Aug 2009 B2
7631013 Parsons et al. Dec 2009 B2
7634429 Narin et al. Dec 2009 B2
7656127 Shutt et al. Feb 2010 B1
7657459 Anderson et al. Feb 2010 B2
7657831 Donahue Feb 2010 B2
7680849 Heller et al. Mar 2010 B2
7720892 Healey, Jr. et al. May 2010 B1
7760986 Beuque Jul 2010 B2
7788369 McAllen et al. Aug 2010 B2
7792756 Plastina et al. Sep 2010 B2
7835989 Hendricks et al. Nov 2010 B1
7849393 Hendricks et al. Dec 2010 B1
7865405 Hendricks et al. Jan 2011 B2
7865567 Hendricks et al. Jan 2011 B1
7870022 Bous et al. Jan 2011 B2
7900133 Cragun et al. Mar 2011 B2
7908628 Swart et al. Mar 2011 B2
8131647 Siegel et al. Mar 2012 B2
8165998 Semerdzhiev Apr 2012 B2
20010007980 Ishibashi et al. Jul 2001 A1
20010025302 Suzuki et al. Sep 2001 A1
20010027450 Shinoda et al. Oct 2001 A1
20010027478 Meier et al. Oct 2001 A1
20010037328 Pustejovsky et al. Nov 2001 A1
20010049623 Aggarwal et al. Dec 2001 A1
20010053975 Kurihara Dec 2001 A1
20020002540 DeMello et al. Jan 2002 A1
20020010707 Chang et al. Jan 2002 A1
20020010759 Hitson et al. Jan 2002 A1
20020012134 Calaway Jan 2002 A1
20020035697 McCurdy et al. Mar 2002 A1
20020054059 Schneiderman May 2002 A1
20020059415 Chang et al. May 2002 A1
20020069222 McNeely Jun 2002 A1
20020069312 Jones Jun 2002 A1
20020090934 Mitchelmore Jul 2002 A1
20020091584 Clark et al. Jul 2002 A1
20020092031 Dudkiewicz et al. Jul 2002 A1
20020101447 Carro Aug 2002 A1
20020103809 Starzl et al. Aug 2002 A1
20020120635 Joao Aug 2002 A1
20020123336 Kamada Sep 2002 A1
20020129012 Green Sep 2002 A1
20020138291 Vaidyanathan et al. Sep 2002 A1
20020143822 Brid et al. Oct 2002 A1
20020147724 Fries et al. Oct 2002 A1
20020184319 Willner et al. Dec 2002 A1
20020194474 Natsuno et al. Dec 2002 A1
20030009459 Chastain et al. Jan 2003 A1
20030018720 Chang et al. Jan 2003 A1
20030025731 Chastain et al. Feb 2003 A1
20030028395 Rodgers et al. Feb 2003 A1
20030040970 Miller Feb 2003 A1
20030046233 Ara et al. Mar 2003 A1
20030052928 Williams Mar 2003 A1
20030058265 Robinson et al. Mar 2003 A1
20030065642 Zee Apr 2003 A1
20030069812 Yuen et al. Apr 2003 A1
20030093312 Ukita et al. May 2003 A1
20030093382 Himeno et al. May 2003 A1
20030097354 Finlay et al. May 2003 A1
20030105679 Krishnan et al. Jun 2003 A1
20030110503 Perkes Jun 2003 A1
20030129963 Nurcahya Jul 2003 A1
20030135495 Vagnozzi Jul 2003 A1
20030152894 Townshend Aug 2003 A1
20030163399 Harper et al. Aug 2003 A1
20030164844 Kravitz et al. Sep 2003 A1
20030182551 Frantz et al. Sep 2003 A1
20030190145 Copperman et al. Oct 2003 A1
20030191737 Steele et al. Oct 2003 A1
20030204496 Ray et al. Oct 2003 A1
20030212613 Sarig Nov 2003 A1
20040002943 Merrill et al. Jan 2004 A1
20040003398 Donian et al. Jan 2004 A1
20040015467 Fano Jan 2004 A1
20040030686 Cardno et al. Feb 2004 A1
20040044723 Bell et al. Mar 2004 A1
20040054499 Starzyk et al. Mar 2004 A1
20040078757 Golovchinsky et al. Apr 2004 A1
20040098350 Labrou et al. May 2004 A1
20040117189 Bennett Jun 2004 A1
20040128359 Horvitz et al. Jul 2004 A1
20040139400 Allam et al. Jul 2004 A1
20040167822 Chasen et al. Aug 2004 A1
20040201633 Barsness et al. Oct 2004 A1
20040205457 Bent et al. Oct 2004 A1
20040212941 Haas et al. Oct 2004 A1
20040237033 Woolf et al. Nov 2004 A1
20040243613 Pourheidari Dec 2004 A1
20040267552 Gilliam et al. Dec 2004 A1
20040268253 DeMello et al. Dec 2004 A1
20050021464 Lindauer et al. Jan 2005 A1
20050022113 Hanlon Jan 2005 A1
20050044148 Son et al. Feb 2005 A1
20050044224 Jun et al. Feb 2005 A1
20050066219 Hoffman et al. Mar 2005 A1
20050069225 Schneider et al. Mar 2005 A1
20050088410 Chaudhri Apr 2005 A1
20050102618 Naito May 2005 A1
20050125222 Brown et al. Jun 2005 A1
20050138007 Amitay Jun 2005 A1
20050138428 McAllen et al. Jun 2005 A1
20050144895 Grimes et al. Jul 2005 A1
20050176438 Li Aug 2005 A1
20050177567 Hughes et al. Aug 2005 A1
20050193330 Peters Sep 2005 A1
20050195975 Kawakita Sep 2005 A1
20050198070 Lowry Sep 2005 A1
20050222977 Zhou et al. Oct 2005 A1
20050250439 Leslie Nov 2005 A1
20050256822 Hollingsworth Nov 2005 A1
20060004840 Senda Jan 2006 A1
20060020469 Rast Jan 2006 A1
20060031316 Forstadius Feb 2006 A1
20060047830 Nair et al. Mar 2006 A1
20060048184 Poslinski et al. Mar 2006 A1
20060057960 Tran Mar 2006 A1
20060061595 Goede et al. Mar 2006 A1
20060071754 Tofts et al. Apr 2006 A1
20060075205 Martin et al. Apr 2006 A1
20060080261 Christal Apr 2006 A1
20060098900 King et al. May 2006 A1
20060101328 Albornoz et al. May 2006 A1
20060123053 Scannell Jun 2006 A1
20060129618 Maier Jun 2006 A1
20060143558 Albornoz et al. Jun 2006 A1
20060156222 Chi et al. Jul 2006 A1
20060161635 Lamkin et al. Jul 2006 A1
20060175983 Crouse et al. Aug 2006 A1
20060179137 Jennings et al. Aug 2006 A1
20060190489 Vohariwatt et al. Aug 2006 A1
20060190568 Patterson Aug 2006 A1
20060236240 Lebow Oct 2006 A1
20060250994 Sasaki et al. Nov 2006 A1
20060253441 Nelson Nov 2006 A1
20060253461 de Bonet Nov 2006 A1
20060256083 Rosenberg Nov 2006 A1
20060265518 Owens et al. Nov 2006 A1
20060281058 Mangoaela Dec 2006 A1
20070014404 Cha Jan 2007 A1
20070025704 Tsukazaki et al. Feb 2007 A1
20070039023 Kataoka Feb 2007 A1
20070050346 Goel et al. Mar 2007 A1
20070061335 Ramer et al. Mar 2007 A1
20070073596 Alexander et al. Mar 2007 A1
20070079383 Gopalakrishnan Apr 2007 A1
20070094285 Agichtein et al. Apr 2007 A1
20070094351 Kalish et al. Apr 2007 A1
20070105536 Tingo May 2007 A1
20070112817 Danninger May 2007 A1
20070118533 Ramer et al. May 2007 A1
20070136660 Gurcan et al. Jun 2007 A1
20070136679 Yang Jun 2007 A1
20070150456 Lian et al. Jun 2007 A1
20070174545 Okada et al. Jul 2007 A1
20070185865 Budzik et al. Aug 2007 A1
20070189719 Furumachi et al. Aug 2007 A1
20070219983 Fish Sep 2007 A1
20070233562 Lidwell et al. Oct 2007 A1
20070233692 Lisa et al. Oct 2007 A1
20070234209 Williams Oct 2007 A1
20070282809 Hoeber et al. Dec 2007 A1
20070283173 Webb et al. Dec 2007 A1
20070288853 Neil Dec 2007 A1
20080005097 Kleewein et al. Jan 2008 A1
20080005664 Chandra Jan 2008 A1
20080027933 Hussam Jan 2008 A1
20080031595 Cho Feb 2008 A1
20080040233 Wildman et al. Feb 2008 A1
20080082518 Loftesness Apr 2008 A1
20080082911 Sorotokin et al. Apr 2008 A1
20080089665 Thambiratnam et al. Apr 2008 A1
20080113614 Rosenblatt May 2008 A1
20080120280 Iijima et al. May 2008 A1
20080133479 Zelevinsky et al. Jun 2008 A1
20080163039 Ryan et al. Jul 2008 A1
20080164304 Narasimhan et al. Jul 2008 A1
20080168073 Siegel et al. Jul 2008 A1
20080208833 Basmov Aug 2008 A1
20080222552 Batarseh et al. Sep 2008 A1
20080235351 Banga et al. Sep 2008 A1
20080243788 Reztlaff, II et al. Oct 2008 A1
20080243828 Reztlaff, II et al. Oct 2008 A1
20080259057 Brons Oct 2008 A1
20080293450 Ryan et al. Nov 2008 A1
20080294674 Reztlaff, II et al. Nov 2008 A1
20080295039 Nguyen et al. Nov 2008 A1
20090094528 Gray et al. Apr 2009 A1
20090094540 Gray et al. Apr 2009 A1
20090181649 Bull et al. Jul 2009 A1
20090228774 Matheny et al. Sep 2009 A1
20090263777 Kohn Oct 2009 A1
20090319482 Norlander et al. Dec 2009 A1
20100023259 Krumm et al. Jan 2010 A1
20100081120 Nanjiani et al. Apr 2010 A1
20100188327 Frid et al. Jul 2010 A1
20120041941 King et al. Feb 2012 A1
Foreign Referenced Citations (11)
Number Date Country
1362682 Aug 2002 CN
101120358 Feb 2008 CN
1842150 Oct 2007 EP
2002099739 Apr 2002 JP
WO9720274 Jun 1997 WO
WO0045588 Aug 2000 WO
WO0056055 Sep 2000 WO
WO0075840 Dec 2000 WO
WO0239206 May 2002 WO
WO2004055647 Jul 2004 WO
WO2006078728 Jul 2006 WO
Non-Patent Literature Citations (113)
Entry
Card et al., “3Book: A 3D Electronic Smart Book,” AVI'04, May 25-28, 2004, Hallipoli, Italy, ACM 2004, pp. 303-307.
Translated Japanese Office Action mailed Oct. 25, 2011 for Japanese patent application No. 2007-552235, a counterpart foreign application of U.S. Appl. No. 11/039,645, 3 pages.
Mercier et al., “Sphere of influence Model in Information retrieval,” IEEE 2005 International Conference on Fuzzy Systems, pp. 120-125.
Final Office Action for U.S. Appl. No. 11/537,484, mailed on Jan. 24, 2012, Thomas A. Ryan, “Expedited Acquisition of a Digital Item Following a Sample Presentation of the Item,” 22 pages.
Final Office Action for U.S. Appl. No. 12/414,914, mailed on Jan. 4, 2012, Agarwal et al., “Questions on Highlighted Passages,” 41 pages.
Office action for U.S. Appl. No. 11/763,357, mailed on Dec. 21, 2011, Retzlaff et al., “Obtaining and Verifying Search Indices,” 14 pages.
Final Office Action for U.S. Appl. No. 11/763,363, mailed on Dec. 23, 2011, James R. Retzaff II et al., “Search Results Generation and Sorting,” 10 pages.
Non-Final Office Action for U.S. Appl. No. 11/763,369, mailed on Dec. 29, 2011, James R. Retzlaff II et al., “Search of Multiple Content Sources on a User Device,” 21 pages.
Final Office Action for U.S. Appl. No. 11/693,685, dated Dec. 8, 2011, John Lattyak et al., “Relative Progress and Event Indicators,” 23 pages.
PCT Search Report for PCT Application No. PCT/US10/22060, mailed Mar. 8, 2010 (7 pages).
Cavanaugh, “EBooks and Accommodations,” Teaching Exceptional Children, vol. 35, No. 2, Copyright 2002 CEC, 6 pages.
Final Office Action for U.S. Appl. No. 11/763,314, mailed on Jun. 13, 2011, Craig S. Griffin, “Display Dependent Markup Language,” 26 pages.
Non-Final Office Action for U.S. Appl. No. 11/693,682, mailed on Jun. 9, 2011, Hilliard B. Siegel, “Providing Annotations of a Digital Work,” 12 pages.
Office action for U.S. Appl. No. 12/333,215, mailed on Jul. 18, 2011, Zagorie et al., “Device-Specific Presentation Control for Electronic Book Reader Devices”, 22 pages.
Non-Final Office Action for U.S. Appl. No. 11/693,685, mailed on Jul. 6, 2011, John Lattyak, “Relative Progress and Event Indicators,” 22 pages.
Office action for U.S. Appl. No. 12/414,914, mailed on Aug. 4, 2011, Agarwal et al., “Questions on Highlighted Passages,” 39 pages.
BarnesandNoble.com, “Barnes and Noble Homepage,” retrieved on Aug. 2, 2011 at <<http://web.archive.org/web/19981202183957/http://www.barnesandnoble.com/>>, Dec. 2, 1998, 2 pages.
Translated Chinese Office Action mailed Jul. 14, 2011 for Japanese patent application No. 20078004873.9, a counterpart foreign application of US patent No. 7,865,817, 6 pages.
Translated Chinese Office Action mailed Aug. 25, 2011 for Chinese patent application No. 200880024964.2, a counterpart foreign application of US patent application No. 11/763,358, 6 pages.
Non-final Office Action for U.S. Appl. No. 11/537,484, mailed on Aug. 19, 2011, Thomas A. Ryan, “Expedited Acquisition of a Digital Item Following a Sample Presentation of the Item,” 13 pages.
Non-final Office Action for U.S. Appl. No. 11/763,363, mailed on Aug. 26, 2011, James R. Retzlaff II, “Search Results Generation and Sorting,” 10 pages.
Non-Final Office Action for U.S. Appl. No. 11/763,358, mailed on Sep. 12, 2011, James R. Retzlaff II, “Managing Status of Search Index Generation,” 11 pages.
Cafesoft.com, “Security Glossary”, dated Oct. 13, 2003, retrieved from the Wayback Machine on Jul. 2, 2009 at <<http://web.archive.org/web/20031013022218/http://cafesoft.com/support/security-glossary.html>>, 6 pages.
Translated Chinese Office Action mailed Oct. 10, 2011 for Chinese patent application No. 200880017259.X, a counterpart foreign application of U.S. Appl. No. 11/693,682, 7 pages.
Translated Chinese Office Action mailed Sep. 26, 2011 for Chinese patent application No. 200880017589.9, a counterpart foreign application of U.S. Appl. No. 11/763,369, 9 pages.
Translated Chinese Office Action mailed Sep. 30, 2011 for Chinese patent application No. 200880025056.5, a counterpart foreign application of U.S. Appl. No. 11/763,374, 9 pages.
Office action for U.S. Appl. No. 12/360,089, mailed on Oct. 5, 2011, Killalea et al., “Aggregation of Highlights,” 75 pages.
Final Office Action for U.S. Appl. No. 11/763,392, mailed on Oct. 14, 2011, Thomas Ryan, “Administrative Tasks a Media Consumption System,” 38 pages.
Office action for U.S. Appl. No. 11/763,390, mailed on Oct. 24, 2011, Bajaj et al., “Providing User-Supplied Items to a User Device,” 11 pages.
Final Office Action for U.S. Appl. No. 12/759,828, mailed on Nov. 10, 2011, James R. Retzalff II, “Search and Indexing on a User Device,” 16 pages.
Final Office Action for U.S. Appl. No. 11/537,518, mailed on Nov. 25, 2011, John Lattyak, “Acquisition of an Item Based on a Catalog Presentation of Items,” 8 pages.
Non-Final Office Action for U.S. Appl. No. 11/763,386, mailed on Nov. 8, 2011, Thomas Ryan, “Handling of Subscription-Related Issues in a Media Consumption System,” 10 pages.
Non-Final Office Action for U.S. Appl. No. 11/763,395, mailed on Nov. 9, 2011, Thomas Ryan, “Providing Supplemental Information Based on Hints in a Media Consumption System,” 10 pages.
Final Office Action for U.S. Appl. No. 11/963,618, mailed on Sep. 26, 2011, Michael Rykov, “Dissemination of Periodical Samples,” 15 pages.
Non-Final Office Action for U.S. Appl. No. 11/763,374, mailed on Sep. 27, 2011, Thomas Ryan, “Consumption of Items via a User Device,” 17 pages.
Beigbeder, et al., “An Information Retrieval Model Using the Fuzzy Proximity Degree of Term Occurences”, 2005 ACM Symposium on Applied Computing, pp. 1018-1022.
Biskup, et al, “Towards a Credential-Based Implementation of Compound Access Control Policies” SACMAT '04, Proceedings of the ninth ACM symposium on Access control models and technologies, Jun. 4, 2004, NY, retrieved from the internet http://portal.acm.org/citation.cfm?id=990036.990042>>, retrieved Nov. 9, 2010, pp. 1-10.
Breu, et al., “The Medoc Distributed Electronic Library: Accounting and Security Aspects”, Electronic Publishing, New Models and Opportunities, Proceedings of an ICCC/IFIP Conference, Apr. 14, 1997, pp. 237-249.
Gladney, “Access Control for Large Collections”, NY, vol. 15, No. 2, Apr. 1, 1997, pp. 154-194.
Henke, “Survey on Electronic Book Features”, Open eBook Forum, online, Mar. 20, 2002, pp. 1-14, retrieved from the internet http://www.openebook.org/doc—library/surveys/IDPF—eBook—Features—2002.pdf>> retrieved Nov. 8, 2010.
Navarro, et al., “Modern Information Retrieval, Chapter 8: Indexing and Searching”, Jan. 1, 1999, Modern Information Retrieval, ACM Press, New York, pp. 191-228.
Ziviani, et al., “Modern Information Retrieval, Text Operations”, Jan. 1, 1999, Modern Information Retrieval, ACM Press, NY, pp. 163-190.
Zobel, et al., “Inverted Files for Text Search Engines” ACM Computing Surveys, vol. 38, No. 2, Jul. 1, 2006, pp. 1-56, NY, NY.
Final Office Action for U.S. Appl. No. 11/763,374, mailed on Feb. 13, 2012, Thomas Ryan et al., “Consumption of Items via a User Device”, 14 pages.
Non-Final Office Action for U.S. Appl. No. 11/763,393, mailed on Feb. 16, 2012, John Lattyak et al., “Incremental Updates of Items”, 24 pages.
Final Office Action for U.S. Appl. No. 12/360,089, mailed on Mar. 28, 2012, Tom Killalea et al., “Aggregation of Highlights”, 17 pages.
Non-Final Office Action for U.S. Appl. No. 12/366,941, mailed on Mar. 30, 2012, Scott Dixon et al., “Bundled Digital Content”, 12 pages.
Non-Final Office action for U.S. Appl. No. 11/693,682, mailed on Apr. 23, 2012, Siegel et al., “Providing Annotations of a Digital Work”, 12 pages.
Final Office Action for U.S. Appl. No. 11/763,386, mailed on Apr. 26, 2012, Thomas Ryan et al., “Handling of Subscription-Related Issues in a Media Consumption System”, 14 pages.
Non-Final Office Action for U.S. Appl. No. 13/083,445, mailed on May 4, 2012, Hilliard B. Siegel et al., “Method and System for Providing Annotations of a Digital Work”, 20 pages.
Desmoulins et al., “Pattern-Based Annotations on E-books: From Personal to Shared Didactic Content”, Proceedings of the IEEE International Workshop on Wireless and Mobile Techniques in Education, 2002, 4 pages.
Canadian Office Action mailed Jul. 6, 2012 for Canadian patent application No. 2594573, a counterpart foreign application of U.S. Appl. No. 8,131,647, 5 pages.
Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 11/763,378, mailed on Mar. 16, 2010, 16 pgs.
Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 11/763,375, mailed on Jan. 19, 2010, 31 pgs.
Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 11/763,378, mailed on Oct. 15, 2009, 31 pgs.
Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 11/763,375, mailed on Aug. 6, 2010, 17 pgs.
Bradley, “Plastic Shape Shifter”, retreived on May 7, 2009 at <<http://www.reactivereports.com/61/61—3.html>>, Chemistry WebMagazine, Issue No. 61, Dec. 2006, 2 pgs.
Cavanaugh “EBooks and Accommodations”, Teaching Expectional Children vol. 35 No. 2 p. 56-61 Copyright 2002 CEC, pp. 56-61.
Chi et al. “eBooks with Indexes that Reorganize Conceptually”, CHI2004, Apr. 24-29, 2004, Vienna, Austria ACM 1-58113-703-6/04/0004, pp. 1223-1226.
Davison et al. “The Use of eBooks and Interactive Multimedia, as Alternative Forms of Technical Documentation” SIGDOC'05, Sep. 21-23, 2005, Conventry, United Kingdom, Copyright 2005 ACM 1-59593-175-9/5/0009, pp. 108-115.
Elspass, et al., “Portable Haptic Interface with Active Functional Design”, In Proceedings SPIE Conference on Smart Structures and Integrated Systems, Newport Beach, California, vol. 3668, Mar. 1999, 926-932.
“Haptic History—Machine Haptics (Expansion)” retrieved on May 7, 2009 at <<http://hapticshistory.chc61.uci.cu/haptic/site/pages/Machine-Haptics-Became—5.php.>> from Google's cache, text-only version as webpage appeared on Apr. 16, 2009, 8 pgs.
Jones, et al., “Development of a Tactile Vest”, IEEE Computer Society, In the Proceedings of the 12th International Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems, Mar. 27-28, 2004, pp. 82-89.
Leutwyler, “Shape-shifting Polymer Gels”, retrieved on May 7, 2009 at <<http://www.scientificamerican.com/article.cfm?id=shape-shifting-polymer-ge&print=true>>, Scientific American, Nov. 9, 2000, 1 pg.
Nakatani, et al., “3D Form Display with Shape Memory Alloy”, In Proceedings of 13th International Conference on Artificial Reality and Teleexistence (ICAT), 2003, pp. 179-184.
“Shape Memory Polymer”, retrieved on May 7, 2009 at <<http://en.wikipedia.org/wiki/Shape—Memory—Polymer>>, Wikipedia, 8 pgs.
Sohn et al. “Development of a Standard Format for eBooks”, SAC2002, Madrid, Spain, 2002 ACM 1-58113-445-2/02/0, pp. 535-540.
Wellman, et al., “Mechanical Design and Control of a High-Bandwidth Shape Memory Alloy Tactile Display”, Springer-Verlag, In the Proceedings of the International Symposium on Experimental Robotics, Barcelona, Spain, Jun. 1997, pp. 56-66, 12 pgs.
Yoshikawa, et al., “Vertical Drive Micro Actuator for Haptic Display Using Shape Memory Alloy Thin Film”, IEE Japan, Papers of Technical Meeting on Micromachine and Sensor System, Journal Code L2898B, vol. MSS-05, No. 21-44, 2005, pp. 103-108.
Cleveland, Jr. et al., “Questions and Answers about Biological Effects and Potential Hazards of Radiofrequency Electromagnetic Fields” OET Bulletin 56, Fourth Edition, Aug. 1999, 38 pages.
Cleveland, Jr., et al, “Evaluating Compliance with FCC Guidelines for Human Exposure to Radiofrequency Electromagnetic Fields” OET Bulletin 65, Edition 97-01, Aug. 1997, 84 pages.
Means, et al., “Evaluating Compliance with FCC Guidelines for Human Exposure to Radiofrequency Electromagnetic Fields”, OET Bulletin 65 Edition 97-01, Jun. 2001, 57 pages.
OQO “A Full PC That Fits in Your Pocket” Retrieved on Sep. 22, 2008 at <<http://www.oqo.com/support/documentation.html>>.
PCT Search Report for International Application No. PCT/US 07/89105, mailed Aug. 18, 2008 (4 pages).
PCT Search Report for International Application No. PCT/US 08/57829, mailed Aug. 15, 2008 (2 pages).
PCT Search Report and Written Opinion for International Application No. PCT/US 08/64387, mailed Sep. 9, 2008 (14 pages).
PCT Search Report for International Application No. PCT/US 08/64368, mailed Sep. 22, 2008 (2 pages).
PCT Search Report and Written Opinion for International Application No. PCT/US 08/64389, mailed Jan. 28, 2009 (7 pages).
PCT Search Report and Written Opinion for International Application No. PCT/US 08/57848, mailed Jul. 7, 2008 (9 pages).
Translated Chinese Office Action mailed May 17, 2012 for Chinese patent application No. 20078004873.9, a counterpart foreign application of U.S. Appl. No. 7,865,817, 5 pages.
Translated Chinese Office Action mailed May 21, 2012 for Chinese patent application No. 200880017589.9, a counterpart foreign application of U.S. Appl. No. 11/763,369, 9 pages.
Final Office Action for U.S. Appl. No. 11/763,395, mailed May 9, 2012, Thomas Ryan et al., “Providing Supplemental Information Based on Hints in a Media Consumption System ”, 12 pages.
Office action for U.S. Appl. No. 11/763,390, mailed on Jun. 27, 2012, Bajaj et al., “Providing User-Supplied Items to a User Device”, 7 pages.
Office action for U.S. Appl. No. 11/763,392, mailed on Jun. 27, 2012, Ryan et al., “Administrative Tasks in a Media Consumption System ”, 47 pages.
Non-Final Office Action for U.S. Appl. No. 12/943,211, mailed on Jun. 6, 2012, James. R. Retzlaff II et al., “Obtaining and Verifying Search Indices”, 10 pages.
Non-Final Office Action for U.S. Appl. No. 11/763,369 mailed on Jun. 7, 2012, James R. Reztlaff II et al., “Search of Multiple Content Sources on a User Device”, 20 pages.
“Annotation Engine,” Berkman Center for Internet & Society at Harvard Law School, retrieved on Jan. 30, 2004 at <<http://cyber.law .harvard. edulproj ectsl annotate.html>>, 3 pages.
“Annotator Instructions,” Berkman Center for Internet & Society at Harvard Law School retrieved on Jan. 30, 2004 at <<http://cyber.law.harvard.edu/annotate/instructions.html>>, also found at <<http://cyber.law.harvard.edu/cite/instructions.html>>, 1 page.
“Annotator Wishlist,” Berkman Center for Internet & Society at Harvard Law School retrieved on Jan. 30, 2004 at <<http://cyber.law.harvard.edulcite/annotate.cgi ?ction=print&markup ;center=; view=http%3A%2F%2Fcy . . . >>, 1 page.
Roscheisen et al., “Beyond Browsing: Shared Comments, SOAPs, Trails, and On-Line Communities,” Computer Networks and ISDN Systems 27:739- 749, 1995, 12 pages.
Canadian Office Action mailed Apr. 14, 2009 for Canadian Patent Application No. 2594573, a counterpart foreign application of U.S. Appl. No. 11/039,645, 3 pages.
Translated Chinese Office Action mailed May 9, 2008 for Chinese Patent Application No. 200680002606.2, a counterpart foreign application of U.S. Appl. No. 11/039,645, 22 pages.
Translated Chinese Second Office Action mailed Jun. 5, 2009 for Chinese Patent Application No. 200680002606.2, a counterpart foreign application of U.S. Appl. No. 11/039,645, 20 pages.
Translated Chinese Third Office Action mailed Nov. 27, 2009 for Chinese Patent Application No. 200680002606.2, a counterpart foreign application of U.S. Appl. No. 11/039,645, 15 pages.
Carter et al., “Digital Graffiti: Public Annotation of Multimedia Content,” Proceedings of the CHI2004, Vienna, Austria, Apr. 24-29, 2004, pp. 1207-1210.
Extended European Search Report mailed Dec. 22, 2009, issued in corresponding European Patent Application No. EP 06 71 8773.2, filed Jan. 18, 2006, 9 pages.
European Office Action mailed Mar. 26, 2010 for European Patent Application No. 06718773.2, a counterpart foreign application of U.S. Appl. No. 11/039,645, 1 page.
Translated Japanese Office Action mailed Jan. 25, 2011 for Japanese Patent Application No. 2007-552235, a counterpart foreign application of U.S. Appl. No. 11/039,645, 6 pages.
Non-Final Office Action for U.S. Appl. No. 11/537,518, mailed on Apr. 28, 2011, John Lattyak, “Acquisition of an Item Based on a Catalog Presentation of Items,” 26 pages.
Office action for U.S. Appl. No. 11/763,357, mailed on May 26, 2011, Retzlaff, “Obtaining and Verifying Search Indices,” 21 pages.
International Search Report mailed Sep. 9, 2008, in International Application No. PCT/US/08/64387, filed May 21, 2008, 1 page.
International Search Report mailed Aug. 15, 2008, in corresponding International Application No. PCT/US08/57829, filed Mar. 21, 2008, 1 page.
International Search Report mailed Jul. 7, 2008, in International Application No. PCT/US08/57848, filed Mar. 31, 2008, 2 pages.
PCT International Search Report and the Written Opinion for Application No. PCT/US 08/64389, mailed on Jan. 28, 2009, 7 pages.
“Say NO to Third Voice,” Worldzone.net, 1999-2004, retrieved on Jan. 30, 2004 at <<http://worldzone.netiinternetipixelsnttv/index.html>>, 3 pages.
Marshall, “The Future of Annotation in a Digital (Paper) World,” Proceedings of the 35th Annual GSLIS Clinic, University of Illinois at Urbana Champaign, Urbana, 11, Mar. 22-24, 1998, pp. 1-19.
Kumar, “Third Voice Trails off . . . ,” Wired News, 2004, retrieved on Jan. 30, 2004 at <<http://www.wired.comlnews/printIO. 1294,42803 ,00.html>>, 3 pages.
“A Universally Unique IDentifier (UUID) URN Namespace”, Jul. 2005, IETF, 32 pages. Retrieved on Apr. 21, 2010 at http://tools.iettorg/pdf/r1c4122.pdf.
Bellwood, et al., “UDDI Version 2.04 API Specification UDDI Committee Specification, Jul. 19, 2002”, Oasis, 95 pages. Retrieved on Apr. 21, 2010 via Wayback Machine at http://web.archive.org/web/20050314033213/www.oasis-open.org/committees/uddi-spec/doc/tcspecs.htm.
“Universal Unique Identifier”, dated Dec. 16, 2002. The Open Group, 9 pages. Retrieved on Apr. 21, 2010 via Wayback Machine at http://web.archive.org/web/20021216070918/http://www.opengroup.org/onlinepubs/9629399/apdxa.htm.
“Web Services Architecture: W3C Working Group Note Feb. 11, 2004”, Feb. 11, 2004, W3C, 100 pages. Retrieved on Apr. 21, 2010 via Wayback Machine at http://web.archive.org/web/2004040205185/http://www.w3.org/TR/ws-arch/.
Translated Chinese Office Action mailed Aug. 3, 2012 for Chinese patent application No. 200880025056.5, a counterpart foreign application of U.S. Appl. No. 11/763,374, 17 pages.
Final Office Action from the U.S. Patent and Trademark Office for U.S. Appl. No. 12/414,914, mailed on Jan. 4, 2012, 16 pgs.
Related Publications (1)
Number Date Country
20100188327 A1 Jul 2010 US