Techniques for interactively coupling electronic content with printed media

Information

  • Patent Grant
  • 8261967
  • Patent Number
    8,261,967
  • Date Filed
    Wednesday, July 19, 2006
    17 years ago
  • Date Issued
    Tuesday, September 11, 2012
    11 years ago
Abstract
A technique for interactively coupling electronic content to printed or written content, in accordance with one embodiment of the present invention, includes use of a computing device and a pre-printed control panel. The control panel includes graphic representations overlaid on unique coded patterns. Each unique coded pattern is associated with a particular function or datum. The control panel is used to input an identifier of a written or printed content to the computing device. In response thereto, the computing device interactively delivers the electronic content related to the particular written or printed content.
Description
BACKGROUND OF THE INVENTION

A number of techniques for interacting with content printed on encoded media have been developed. In one technique, content is printed or written on paper containing an electro-optically readable coding pattern. The coding pattern is utilized to determine position information which can be utilized to interact with the content thereon. However, there is a substantial amount of content that has not been disseminated on non-encoded paper. Without being able to determine position information, it is difficult to deliver interactive electronic content related to such printed or written content. It is also undesirable to reproduce the substantial volume of existing literary work on encoded paper. Therefore, what is needed is a technique for delivering electronic content in an interactive session with related content printed on non-encoded media.


SUMMARY OF THE INVENTION

Embodiments of the present invention are directed toward systems and methods for interactively coupling electronic content with printed or written content. In one embodiment, the techniques include receiving an identifier of a particular written or printed content utilizing a control panel that is printed on encoded media. Electronic content related to the written or printed content may then be interactively delivered.


In another embodiment, a pre-printed control panel includes a plurality of graphics, wherein each graphics overlays a unique coded pattern. Each coded pattern is associated with a particular function or datum. A given graphic may represent a button, key, alphabetic character, numeric character, symbol, figure or the like for invoking the function or input the datum associated with the coded pattern that it overlays.


In yet another embodiment, a computing device includes an optical tracking interface adapted to receive one or more encoded patterns from a printed control panel. A processor of the computing device determines an identifier of a written or printed content from the received encoded patterns. The processor interactively delivers electronic content related to the identified written or printed content.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments of the present invention are illustrated by way of example and not by way of limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar elements and in which:



FIG. 1 shows a block diagram of a computing device upon which embodiments of the present invention can be implemented.



FIG. 2 shows an exemplary media provided with a coding pattern upon which embodiments of the present invention can be implemented.



FIG. 3 illustrates an exemplary use of a control panel printed on encoded media that enables interaction between a computing device and conventional media, in accordance with one embodiment of the present invention.



FIG. 4 illustrates a first sub-control panel of an exemplary control panel, in accordance with one embodiment of the present invention.



FIG. 5 illustrates a second and third sub-control panel of an exemplary control panel, in accordance with one embodiment of the present invention.



FIG. 6 a flow diagram of a method of using a control panel printed on one encoded media to interactively couple electronic media with printed media, in accordance with one embodiment of the present invention.





DETAILED DESCRIPTION OF THE INVENTION

Reference will now be made in detail to the embodiments of the invention, examples of which are illustrated in the accompanying drawings. While the invention will be described in conjunction with these embodiments, it will be understood that they are not intended to limit the invention to these embodiments. On the contrary, the invention is intended to cover alternatives, modifications and equivalents, which may be included within the scope of the invention as defined by the appended claims. Furthermore, in the following detailed description of the present invention, numerous specific details are set forth in order to provide a thorough understanding of the present invention. However, it is understood that the present invention may be practiced without these specific details. In other instances, well-known methods, procedures, components, and circuits have not been described in detail as not to unnecessarily obscure aspects of the present invention.



FIG. 1 shows a block diagram of a computing device 100 upon which embodiments of the invention can be implemented. In general, computing device 100 may have a form factor similar to a pen, stylus or the like. The computing device 100 includes system memory 105, a processor 110, an input/output interface 115, an optical tracking interface 120, one or more buses 125, and a pointing and/or writing instrument 130. The system memory 105, processor 110, input/output interface 115 and optical tracking interface 120 are communicatively coupled to each other by the one or more buses 125.


The memory 105 may include one or more well known computer-readable media, such as static or dynamic read only memory (ROM), random access memory (RAM), flash memory, magnetic disk, optical disk and/or the like. The memory 105 may be used to store one or more sets of instructions and data that, when executed by the processor 110, cause the computing device 100 to perform the techniques for interactively coupling electronic media with printed or written media as described herein.


The computing device 100 may further include an external memory controller 135 for removably coupling an external memory 140 to the one or more busses 125. The external memory may be used to store various electronic content for use by the computing device 100. The computing device 100 may also include one or more communication ports 145 communicatively coupled to the one or more busses 125. The one or more communication ports can be used to communicatively couple the computing device 100 to one or more other devices 150. The computing device 110 may be communicatively coupled to other devices 150 by a wired communication link and/or a wireless communication link 155. Furthermore, the communication link may be a point-to-point connection and/or a network connection. The communications ports may be used to receive various electronic content from the other devices 150.


The input/output interface 115 may include one or more electro-mechanical switches operable to receive commands and/or data from a user. The input/output interface 115 may also include one or more audio devices, such as a speaker, a microphone, and/or one or more audio jacks for removably coupling an earphone, headphone, external speaker and/or external microphone. The audio device is operable to output audio content and information and/or receiving audio content, information and/or instructions from a user. The input/output interface 115 may optionally include video devices, such as an LCD display for displaying alphanumeric and/or graphical information and/or a touch screen display for displaying and/or receiving alphanumeric and/or graphical information.


The optical tracking interface 120 includes a light source or optical emitter and a light sensor or optical detector. The optical emitter may be a light emitting diode (LED) and the optical detector may be a charge coupled device (CCD) or complementary metal-oxide semiconductor (CMOS) imager array, for example. The optical emitter illuminates a surface of a media or a portion thereof and light reflected from the surface is received at the optical detector.


The surface of the media may contain a pattern detectable by the optical detector. Referring now to FIG. 2, an exemplary media 210 provided with a coding pattern, upon which embodiments of the present invention can be implemented, is shown. In one implementation, the media 210 is provided with a coding pattern in the form of electro-optically readable position code that consists of a pattern of dots 220. The dots 220 may not be easily discernible by the human visual system, and may appear as grayscale on the sheet of paper. In one implementation, the pattern comprises a unique pattern of dots 220 printed on paper or other such media. The dots 220, for example, may have a nominal spacing of 0.3 millimeters, with 669,845,157,114,773,458,169 dots encompassing an area in excess of 4.6 million square kilometers, which corresponds to about 73 trillion letter-size pages. The “encoded paper” may be subdivided into unique sections. Different sections can be used (e.g., licensed) by different entities.


As the pointing and/or writing instrument 130 and the optical tracking interface 120 move together relative to the surface, successive images are captured. The optical detector, for example, may take snapshots of the surface 100 times a second or more. By analyzing the images, position on the surface and movement relative to the surface of the media can be tracked. In one implementation, the optical detector fits the dots to a reference system in the form of a raster with raster lines 230, 240 that intersect at raster points 250. Each of the dots 220 is associated with a raster point. For example, the dot 220 is associated with raster point 250. For the dots 220 in an image, the displacement of a dot 220 from the raster point 250 associated with the dot 220 is determined. Using these displacements, the pattern in the image is compared to patterns in the reference system. Each pattern in the reference system is associated with a particular location on the surface. Thus, by matching the pattern in the image with a pattern in the reference system, the position of the writing instrument 130 relative to the surface can be determined.


By interpreting the positions of the dots 220 captured in each snapshot, the operating system and/or one or more applications executing on the computing device 100 can precisely determine the position (e.g., position in a unique region) of the computing device 100 in two dimensions. That is, for example, an x-coordinate and a y-coordinate corresponding to the position of the pointing and/or writing instrument 130 relative to the page (e.g., surface) can be determined. Furthermore, as the pointing and/or writing instrument 130 and the optical detector move together relative to the surface, the direction and distance of each movement can be determined from successive position data.


In addition, different regions of the pattern can be assigned different function, and software programs and applications may assign functionality to the various patterns of dots within a respective region. Therefore, by placing the optical detector in a particular position on the surface and performing some type of actuating event, a specific instruction, command, function, data or the like associated with the position can be entered and/or executed. For example, the pointing and/or writing instrument 130 may be mechanically coupled to an electro-mechanical switch of the input/output interface 115. Therefore, double tapping substantially the same position is detected by the electro-mechanical switch in combination with the position determined by the optical detector as an input that can cause data or a function associated with the particular position to be received or executed, respectively.


The writing element 130 can be, for example, a pen, pencil, stylus, marker or the like, and may or may not be retractable. In one or more instances, a user can use the writing element 130 to make strokes on the surface, including letters, numbers, symbols, figures and the like. These user-produced strokes can be captured (e.g., imaged and/or tracked) and interpreted by the computing device 100 according to their position on the surface on the encoded media. The position of the strokes can be determined using the pattern of dots on the surface. The strokes can also be interpreted by the computing device 100 using optical character recognition (OCR), handwriting recognition and/or handwriting understanding techniques that recognize handwritten characters (or ‘written signals’).


A user, in one implementation, uses the writing element 130 to create a character “M” at a given position on the encoded media. The user may or may not create the character in response to a prompt from the computing device 100. In one implementation, when the user creates the character, device 100 records the pattern of dots that are uniquely present at the position where the character is created. The computing device 100 associates the pattern of dots with the character just captured. When computing device 100 is subsequently positioned over the “M,” the computing device 100 recognizes the particular pattern of dots associated therewith and recognizes the position as being associated with an “M.” In effect, the computing device 100 recognized the presence of the character using the pattern of markings at the position where the character is located, rather then by recognizing the character itself.


In another implementation, the computing device 100 analyzes the pattern of dots that are uniquely present at the position where the character is created (e.g., stroke data). Using a character recognition application, the stroke data captured by analyzing the pattern of dots can be read and translated into the character “M.”


In another implementation, a character is associated with a particular command. For example, a user can write a character composed of a circled “M” that identifies a particular command, and can invoke that command repeatedly by simply positioning the optical detector over the written character. In other words, the user does not have to write the character for a command each time the command is to be invoked; instead, the user can write the character for a command one time and invoke the command repeatedly using the same written character.


In another implementation, the encoded paper may be preprinted with one or more graphics at various locations in the pattern of dots. For example, the graphic may be a preprinted graphical representation of a button. The graphics lies over a pattern of dots that is unique to the position of the graphic. By placing the optical detector over the graphic, the pattern of dots underlying the graphics are read (e.g., scanned) and interpreted, and a command, instruction, function or the like associated with that pattern of dots is implemented by the computing device 100 or data associated with that pattern is received as an input by the computing device 100. Furthermore, some sort of actuating event or movement may be performed using the computing device 100 in order to indicate that the user intends to invoke the command, instruction, function or the like associated with the graphic and/or receive the data as an input.


In yet another implementation, the optical tracking interface 120 may be used to read a barcode. By analyzing the pattern of the barcode, the optical detector in combination with the processor 110, executing an applicable application, can determine one or more data and/or one or more commands associated with the particular barcode. For example, the bar code on the back of the book may be decoded to determine what book it is.


Referring now to FIG. 3, an exemplary use of a control panel printed on encoded media 305 that enables interaction between a computing device 310 and conventional media 315, in accordance with one embodiment of the present invention, is illustrated. The conventional media 315 may include printed or written content, including text and graphics, utilized to disseminate literary work.


The control panel printed on encoded media 305 may include one or more graphics at various regions in the coding pattern. The pattern of dots underlying each graphic may be associated with a given datum or function. The graphics may include graphical representations of one or more bottoms, one or more icons, one or more keys, one or more alphabetic characters, one or more numeric characters, one or more designs, one or more symbols, one or more figures and/or the like. For example, the control panel 305 may include graphical representations of “start,” “enter,” and “repeat” buttons 330. The underlying pattern of dots corresponding to the particular graphical representation of the “enter” button is pre-associated with an input for invoking an enter function in one or more application running on the pen-shaped computing device. Accordingly, when a user taps on the “enter” button an enter function is performed. The control panel 305 may also, for example, include a plurality of alphabetic characters 325. The underlying pattern of dots proximate a given alphabetic character is pre-associated with the particular letter. Accordingly, when a user taps on the “J” graphic, a computing device-readable representation of the letter J is received as an input.


It is appreciated that the illustrated control panel on encoded media 305 is for illustration only and is not intended to limit the scope of the invention. Instead, the control panel may include graphical representation of any number and/or combination of pre-associated functions, data, and open input regions. Furthermore, the pattern of dots visible in the exemplary control panel 305 is for illustrative purposes. Normally, the dot pattern would not be easily discernable to a user.


The control panel 305 may be produced on encoded media, such as paper, cardboard, plastic or the like. The coding pattern on the media is electro-optically readable by the computing device 310. The control panel may have a form factor of a card, sheet, book mark, poster, folder, pamphlet, leaflet, folio, journal or the like. The form fact of the control panel may have dimensions the same as or similar to a business card, postcard, book marker, standard page size, book cover. In one implementation, a first set of controls (e.g., graphics) may be printed on a first surface of the control panel 305 and a second set of controls (e.g., graphics) may be printed on a second surface. In another implementation, the controls may be printed in a first orientation (e.g., portrait) on a first surface and the same controls may be printed in a second orientation (e.g., landscape) on a second surface.


The user may hold the computing device 310 in a similar manner to holding a pen, and therefore it may be referred to as a pen-shaped computing device 310. In particular, as the user points, taps, writes or the like with the pen-shaped computing device 310, the Operating System (OS) or another application collects position information associated with the user's actions on the encoded control panel. The optical tracking interface 120 of the pen-shaped computing device 310, for example, may take snapshots of the surface of the encoded control panel 305 approximately 10-1000 times per second. By analyzing the images of the position code, movement relative to the surface can be tracked. The relative movement can be used to capture the selection of functions and/or input of data that are pre-associated with particular regions of the encoded pattern, and/or stroke data (e.g., writing) in open input regions.


In one example, a student may enter the international standard book number (ISBN) of a work book by tapping on the corresponding alphanumeric keys 325, 330 printed on the control panel 305. The student may then tap enter 335 on the control panel 305 using the pen-shaped computing device 310. The student may enter the international standard book number (ISBN) in response to an audio prompt 340. The prompt 340 may be generated when a particular application is launched on the pen-shaped computing device 310 by the student. The pen-shaped computing device 310 utilizes the ISBN to determine which book is being used by the student. The student may also enter the page number 345 and then tap enter 335 to begin at a given page within the book. The student may then receive interactive audio content corresponding to the given page in the book from the pen-shaped computing device 310. For example, the portion of the content corresponding to the indicated page could be output by the pen-shaped computing device 310 as an audio file 340.


In another implementation, the student may enter the tile of the book using the graphical representation of the alphabetic characters 3325, instead of the ISBN. In yet another implementation, the student may use the optical tracking interface 120 to scan a barcode on the book. An application executing on the processor 110 may determine the identity of the particular book from the barcode.


Another exemplary control panel is illustrated in FIGS. 4 and 5. The illustrated control panel 410 may be in the shape of a pamphlet. A first sub-control panel is printed on the first surface (e.g., front) of the pamphlet, as depicted in FIG. 4. A second sub-control panel is printed on the left inside panel and a third sub-control panel is printed on the right inside panel of the pamphlet, as depicted in FIG. 5. The first sub-control panel may include common functions represented by graphical buttons, such as start, enter and repeat buttons. The first sub-control panel may also include general alphabetic and numeric data entry buttons. The first sub-control panel may also include additional graphical regions for invoking other functions and data, such as activity selection buttons 415 and the like. The other sub-control panels may include one or more of the same graphical buttons that are on the first sub-control panel and also additional graphics for invoking other functions and/or entering other data. Any of the sub-panels may also include open input regions 510 for capturing hand written inputs from the user.


A user, such as a student, may use the control panel 410 to interactively couple electronic content with printed or written content. In particular, the student may receive audio content 340 from the pen-shaped computing device 310 related to the content in a given book 315. In one example, the student may take a quiz about the subject matter on a given page of the book using the pen-shaped computing device. The quiz question may be generated based upon the content of the book or may be those present in the book. For instance, the student may enter the page number 345 and question number 350 of a question 355 appearing in the book 315. The question may be output as an audio prompt and the student may use one or more graphical buttons, such as true and false buttons, multiple choice buttons or the like, to answer the question. The questions may be framed within an engaging theme, such as a quick quiz game, a word storm game, a memory match game or the like. The pen-shaped computing device 310 may also provide audio scaffolded hints for answering the questions, an indication of whether the answer provided by the student via the control panel 410 is correct or not, and other pedagogy.


In yet another example, a language button may be provided on the control panel. Selection of the language button may then provide interactive content in the native language of the user in accordance with a language selected during a configuration of the pen-shaped computing device or in response to an additional prompt when the language button is taped by the student. In addition, the control panel 410 may include audio control graphical buttons, such as up volume, down volume, mute and the like, to adjusting the audio output.


Referring now to FIG. 6, a method of using a control panel printed on encoded media to interactively couple electronic content with printed or written media, in accordance with one embodiment of the present invention is shown. The method includes receiving an identifier of written or printed content utilizing a control panel printed on encoded media, at 610. In one implementation, a user may input a title of the written or printed content by performing actuating events relative to one or more graphics overlaying particular position coded regions each of which is pre-associated with a particular function or datum. For instance, a user may tap on one or more graphical representations of alphanumeric keys to enter the title of a particular book. In another implementation, the user may input the ISBN of the book. In yet another implementation, the user may input a barcode identifying the book. In other implementations, the user may identify the content any number of other ways.


Optionally, the method may also include receiving one or more identifiers of a particular portion of the written or printed content, at 620. For instance, the user may enter a particular page number of the book, a range of page numbers, a chapter number, section identifier, a subsection identifier, a heading, a volume number, an issue number or the like, by tapping on the appropriate graphical representations of alphanumeric keys.


At 630, interactive electronic content related to the written or printed content is delivered. Furthermore, if a identifier of a particular portion of the written or printed content is received, a particular portion of the electronic content related to the particular portion of the written or printed content may be selectively delivered. In addition, if one or more identifiers of additional portions of the written or printed content are received, the additional portions of the electronic content related to the identified portions of the written or printed content may be delivered in response thereto.


The interactive electronic content may provide directed user activities and/or group activities. For example, the interactive content may include one or more contextually-relevant teaching points and sound effect that engage a student in the content of the conventional printed or written media. The teaching points may include factoids, quick reading comprehension suggestions and/or the like. Sound effects may be utilized to put the reader “in the world” that they are reading about and provide additional “entry points” for neural net recall.


The interactive content may also include touch diagram and other visual aids to provide help with strategies that may be challenging when a user is first asked to go it alone. The visual aids may include Venn diagrams, sequence charts, main idea/details, thinking aloud, cause-effect charts and/or the like. Use of the strategies are directly related to the content or a particular portion of the printed or written media.


The interactive content may also include engaging interactions such as game models, such as word storm, quick quiz, memory match and/or the like. For example, in word storm two or more users are given a category. Each user is given a period to generate a list of words that relate to some aspect of the content of the written or printed media. The user may write their list on encoded paper or plain paper. After the period ends the users compare lists and score based on the correctness, uniqueness and/or the like of their responses. In quick quiz, one or more users may answer true/false, multiple choice, free-entry 1-2 word answers in reading comprehension quizzes. In memory match a user may write audio hints in individual regions and then match pairs of audio clues. Such activities may encourage careful listening, group discussion, resolution of conflicts and ambiguities, compare and contrast analysis, creation of visual mnemonics and/or the like.


In one implementation, the interactive content provided by the computing device may be stored in the memory 105. In another implementation, the interactive content provided by the computing device may be received via an external memory 140 such as a memory cartridge that can be removable coupled to the computing device. In yet another implementation, the interactive content may be received via a communication link with another computing device 150.


Optionally, the method may also include reporting parameters concerning the interactive session, at 640. In particular, logged information collected by the computing device 100 during one or more interactive sessions may be reported.


Embodiments advantageous enable a user to interact with written and printed media using a computing device. In particular, the control panel for non-encoded content can be used to enable interaction between a pen-shaped computing device and printed books, pamphlets and the like. Thus, content on conventional written and printed media is no longer static. Instead, the usefulness of content on non-encoded media is increased by use of the control panel printed on encoded media. The control penal also benefits publishers and users by reducing or eliminating the cost of re-publishing content on encoded media.


The foregoing descriptions of specific embodiments of the present invention have been presented for purposes of illustration and description. They are not intended to be exhaustive or to limit the invention to the precise forms disclosed, and obviously many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, to thereby enable others skilled in the art to best utilize the invention and various embodiments with various modifications as are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the Claims appended hereto and their equivalents.

Claims
  • 1. A method of interactively coupling electronic content with printed media: receiving, by a computing device, an identifier of a written or printed content on non-encoded media, utilizing a control panel printed on location encoded media, wherein the control panel is separate from the written or printed content; andinteractively delivering, by the computing device, electronic content related to the written or printed content.
  • 2. The method according to claim 1, wherein receiving the identifier of the written or printed content comprises receiving an international standard book number (ISBN) or a title.
  • 3. The method according to claim 1, wherein interactively delivering electronic content comprises outputting an audio file of the written or printed content, a quiz related to the written or printed content, a hint or an answer, an audio file of the written or printed content in a native language of a user, a directed user activity, a group activity, a sound effect, a touch diagram, a visual aid or a game.
  • 4. The method according to claim 1, further comprising: receiving an identifier, of a particular portion of the written or printed content, utilizing the control panel; andinteractively delivering a particular portion of the electronic content related to the particular portion of the written or printed content.
  • 5. The method according to claim 4, wherein receiving the identifier of the particular portion of the written or printed content comprises receiving a page number, a range of page numbers, a chapter number, a section identifier, a subsection identifier, a heading, a volume number or an issue number.
  • 6. The method according to claim 4, further comprising: receiving an identifier, of an additional portion of the written or printed content, utilizing the control panel; andinteractively delivering an additional portion of the electronic content related to the additional portion of the written or printed content.
  • 7. The method according to claim 1, further comprising reporting parameters concerning an interactive session.
  • 8. An article of manufacture comprising a printed control panel to control interactive delivery of electronic content by a computing device related to written or printed content on non-encoded media, the control panel including a plurality of graphical representations, wherein each graphical representation corresponds to a particular function or datum and wherein each graphical representation overlays a particular region of a position encoding pattern associated with the particular function or datum, wherein the position encoding pattern is adapted to be electro-optically read by the computing device, and the printed control panel is separate from the written or printed content.
  • 9. The pre-printed control panel of claim 8, wherein at least one graphical representation comprises a button, a key, an icon, an alphabetic character, a numeric character, a symbol or a figure.
  • 10. The pre-printed control panel of claim 8, wherein the graphical representations are printed or written on media selected from a group consisting of paper, cardboard and plastic.
  • 11. The pre-printed control panel of claim 10, wherein the media has a form factor selected from a group consisting of a card, a business card, a postcard, a book marker, a poster, a sheet, a folder, a book cover, a pamphlet, a leaflet, a folio and a journal.
  • 12. The pre-printed control panel of claim 11, wherein the form factor comprises: a first surface having a first portion of the plurality of graphics printed thereon; anda second surface having a second portion of the plurality of graphics printed thereon.
  • 13. The pre-printed control panel of claim 11, wherein the form factor comprises: a first surface having the plurality of graphical representations printed in a first orientation; anda second surface having the plurality of graphical representations printed in a second orientation.
  • 14. A computing device comprising: an optical tracking interface adapted to receive one or more location encoding patterns from a printed control panel that includes a plurality of graphical representations, wherein each graphical representation corresponds to a particular function or datum and wherein each graphical representation overlays a unique location encoding pattern associated with the particular function or datum; anda processor, communicatively coupled to the optical tracking interface, for determining an identifier of written or printed content on non-encoded media from one or more functions and data associated with the one or more location encoding patterns received by the optical tracking interface from the printed control panel and interactively delivering electronic content related to the written or printed content, wherein the written or printed content is separate from the printed control panel.
  • 15. The computing device of claim 14, further comprising a computing device-readable media, communicatively coupled to the processor, for storing the electronic content.
  • 16. The computing device of claim 15, wherein the computing device-readable media comprises a memory cartridge.
  • 17. The computing device of claim 14, further comprising a speaker, communicatively coupled to the processor, for interactively outputting the electronic content in an audio format.
  • 18. The computing device of claim 14, further comprising a pointing and or writing instrument mechanically coupled to the optical tracking interface.
  • 19. The computing device of claim 14, further comprising a communication port, communicatively coupled to the processor, for receiving the electronic content.
US Referenced Citations (409)
Number Name Date Kind
2182334 Crespo Dec 1939 A
2932907 Stieber et al. Apr 1960 A
3292489 Johnson et al. Dec 1966 A
3304612 Proctor et al. Feb 1967 A
3530241 Ellis Sep 1970 A
3591718 Asano et al. Jul 1971 A
3657812 Lee Apr 1972 A
3782734 Krainin Jan 1974 A
3798370 Hurst Mar 1974 A
3888311 Cooke, Jr. Jun 1975 A
3911215 Hurst et al. Oct 1975 A
3921165 Dym Nov 1975 A
4079194 Kley Mar 1978 A
4220815 Gibson et al. Sep 1980 A
4318096 Thornburg et al. Mar 1982 A
4337375 Freeman Jun 1982 A
4375058 Bouma et al. Feb 1983 A
4425099 Naden Jan 1984 A
4464118 Scott et al. Aug 1984 A
4492819 Rodgers et al. Jan 1985 A
4570149 Thornburg et al. Feb 1986 A
4603231 Reiffel et al. Jul 1986 A
4604058 Fisher et al. Aug 1986 A
4604065 Frazer et al. Aug 1986 A
4619539 Kageyama Oct 1986 A
4627819 Burrows Dec 1986 A
4630209 Saito et al. Dec 1986 A
4650926 Nakamura et al. Mar 1987 A
4686332 Greanias et al. Aug 1987 A
4706090 Hashiguchi et al. Nov 1987 A
4739299 Eventoff et al. Apr 1988 A
4748318 Bearden et al. May 1988 A
4787040 Ames et al. Nov 1988 A
4793810 Beasley, Jr. Dec 1988 A
4839634 More et al. Jun 1989 A
4841387 Rindfuss Jun 1989 A
4853494 Suzuki Aug 1989 A
4853498 Meadows et al. Aug 1989 A
4853499 Watson Aug 1989 A
4880968 Kwang-Chien Nov 1989 A
4913463 Tlapek et al. Apr 1990 A
4922061 Meadows et al. May 1990 A
4924387 Jeppesen May 1990 A
4964167 Kunizawa et al. Oct 1990 A
4972496 Sklarew Nov 1990 A
4990093 Frazer et al. Feb 1991 A
4991987 Holloway et al. Feb 1991 A
5007085 Greanias et al. Apr 1991 A
5030117 Delorme Jul 1991 A
5053585 Yaniger Oct 1991 A
5057024 Sprott et al. Oct 1991 A
5059126 Kimball Oct 1991 A
5113178 Yasuda et al. May 1992 A
5117071 Greanias et al. May 1992 A
5128525 Stearns et al. Jul 1992 A
5149919 Greanias et al. Sep 1992 A
5157384 Greanias et al. Oct 1992 A
5168147 Bloomberg Dec 1992 A
5184003 McMillin et al. Feb 1993 A
5194852 More et al. Mar 1993 A
5209665 Billings et al. May 1993 A
5217376 Gosselin Jun 1993 A
5217378 Donovan Jun 1993 A
5220136 Kent Jun 1993 A
5220649 Forcier Jun 1993 A
5221833 Hecht Jun 1993 A
5250930 Yoshida et al. Oct 1993 A
5260697 Barrett et al. Nov 1993 A
5294792 Lewis et al. Mar 1994 A
5301243 Olschafskie et al. Apr 1994 A
5314336 Diamond et al. May 1994 A
5356296 Pierce et al. Oct 1994 A
5401916 Crooks Mar 1995 A
5406307 Hirayama et al. Apr 1995 A
5409381 Sundberg et al. Apr 1995 A
5413486 Burrows et al. May 1995 A
5417575 McTaggart May 1995 A
5438168 Wolfe et al. Aug 1995 A
5438662 Randall Aug 1995 A
5466158 Smith, III Nov 1995 A
5474457 Bromley Dec 1995 A
5480306 Liu Jan 1996 A
5484292 McTaggart Jan 1996 A
5485176 Ohara et al. Jan 1996 A
5509087 Nagamine Apr 1996 A
5510606 Worthington et al. Apr 1996 A
5517579 Baron et al. May 1996 A
5520544 Manico et al. May 1996 A
5561446 Montlick Oct 1996 A
5572651 Weber et al. Nov 1996 A
5574519 Manico et al. Nov 1996 A
5574804 Olschafskie et al. Nov 1996 A
5575659 King et al. Nov 1996 A
5596698 Morgan Jan 1997 A
5604517 Filo Feb 1997 A
5624265 Redford et al. Apr 1997 A
5629499 Flickinger et al. May 1997 A
5635726 Zavislan et al. Jun 1997 A
5636995 Sharpe, III et al. Jun 1997 A
5640193 Wellner Jun 1997 A
5649023 Barbara et al. Jul 1997 A
5652412 Lazzouni et al. Jul 1997 A
5652714 Peterson et al. Jul 1997 A
5661506 Lazzouni et al. Aug 1997 A
5663748 Huffman et al. Sep 1997 A
5666214 MacKinlay et al. Sep 1997 A
5686705 Conroy et al. Nov 1997 A
5689667 Kurtenbach Nov 1997 A
5694102 Hecht Dec 1997 A
5697793 Huffman et al. Dec 1997 A
5698822 Haneda et al. Dec 1997 A
5717939 Bricklin et al. Feb 1998 A
5730602 Gierhart et al. Mar 1998 A
5739814 Ohara et al. Apr 1998 A
5757361 Hirshik May 1998 A
5760773 Berman et al. Jun 1998 A
5767457 Gerpheide et al. Jun 1998 A
5788508 Lee et al. Aug 1998 A
5790114 Geaghan et al. Aug 1998 A
5801687 Peterson et al. Sep 1998 A
5805167 van Cruyningen Sep 1998 A
5835726 Shwed et al. Nov 1998 A
5844483 Boley Dec 1998 A
5847698 Reavey et al. Dec 1998 A
5852434 Sekendur Dec 1998 A
5855483 Collins et al. Jan 1999 A
5877458 Flowers Mar 1999 A
5889506 Lopresti et al. Mar 1999 A
5896403 Nagasaki et al. Apr 1999 A
5902968 Sato et al. May 1999 A
5903729 Reber et al. May 1999 A
5910009 Leff et al. Jun 1999 A
5913629 Hazzard Jun 1999 A
5914707 Kono Jun 1999 A
5932863 Rathus et al. Aug 1999 A
5933829 Durst et al. Aug 1999 A
5945656 Lemelson et al. Aug 1999 A
5951298 Werzberger Sep 1999 A
5957697 Iggulden et al. Sep 1999 A
5960124 Taguchi et al. Sep 1999 A
5963199 Kato et al. Oct 1999 A
5963208 Dolan et al. Oct 1999 A
5973420 Kaiserman et al. Oct 1999 A
5974558 Cortopassi et al. Oct 1999 A
5978773 Hudetz et al. Nov 1999 A
5992817 Klitsner et al. Nov 1999 A
5997309 Metheny et al. Dec 1999 A
6000613 Hecht et al. Dec 1999 A
6000621 Hecht et al. Dec 1999 A
6002387 Ronkka et al. Dec 1999 A
6008799 Van Kleeck Dec 1999 A
6009393 Sasaki Dec 1999 A
6011949 Shimomukai Jan 2000 A
6018656 Shirai Jan 2000 A
6020895 Azami Feb 2000 A
6021306 McTaggart Feb 2000 A
6041215 Maddrell et al. Mar 2000 A
6050735 Hazzard Apr 2000 A
6052117 Ohara et al. Apr 2000 A
6064855 Ho May 2000 A
6072476 Harada et al. Jun 2000 A
6076734 Dougherty et al. Jun 2000 A
6076738 Bloomberg et al. Jun 2000 A
6081261 Wolff et al. Jun 2000 A
6088023 Louis et al. Jul 2000 A
6089943 Lo Jul 2000 A
6094197 Buxton et al. Jul 2000 A
6100877 Chery et al. Aug 2000 A
6104387 Chery et al. Aug 2000 A
6104388 Nagai et al. Aug 2000 A
6119944 Mulla et al. Sep 2000 A
6124851 Jacobson Sep 2000 A
6130666 Persidsky Oct 2000 A
6144371 Clary et al. Nov 2000 A
6148173 Bell Nov 2000 A
6164534 Rathus et al. Dec 2000 A
6164541 Dougherty et al. Dec 2000 A
6181329 Stork et al. Jan 2001 B1
6183262 Tseng Feb 2001 B1
6188983 Hanson Feb 2001 B1
6199042 Kurzweil Mar 2001 B1
6199048 Hudetz et al. Mar 2001 B1
6201903 Wolff et al. Mar 2001 B1
6201947 Hur et al. Mar 2001 B1
6208771 Jared et al. Mar 2001 B1
6215476 Depew et al. Apr 2001 B1
6215901 Schwartz Apr 2001 B1
6218964 Ellis Apr 2001 B1
6239792 Yanagisawa et al. May 2001 B1
6241528 Myers Jun 2001 B1
6252564 Albert et al. Jun 2001 B1
6256638 Dougherty et al. Jul 2001 B1
6262711 Cohen et al. Jul 2001 B1
6262719 Bi et al. Jul 2001 B1
6275301 Bobrow et al. Aug 2001 B1
6295439 Bejar et al. Sep 2001 B1
6297812 Ohara et al. Oct 2001 B1
6297824 Hearst et al. Oct 2001 B1
6304667 Reitano Oct 2001 B1
6304898 Shiigi Oct 2001 B1
6304989 Kraus et al. Oct 2001 B1
6309122 Wang Oct 2001 B1
6313828 Chombo Nov 2001 B1
6322369 Patterson et al. Nov 2001 B1
6330976 Dymetman et al. Dec 2001 B1
6331865 Sachs et al. Dec 2001 B1
6331867 Eberhard et al. Dec 2001 B1
6335727 Morishita et al. Jan 2002 B1
6349194 Nozaki et al. Feb 2002 B1
6363239 Tutt et al. Mar 2002 B1
6388681 Nozaki May 2002 B1
6392632 Lee May 2002 B1
6396481 Challa et al. May 2002 B1
6405167 Cogliano Jun 2002 B1
6415108 Kamishima et al. Jul 2002 B1
6418326 Heinonen et al. Jul 2002 B1
6421524 Padgett Jul 2002 B1
6431439 Suer et al. Aug 2002 B1
6434561 Durst, Jr. et al. Aug 2002 B1
6441807 Yamaguchi Aug 2002 B1
6442350 Stephany et al. Aug 2002 B1
6456749 Kasabach et al. Sep 2002 B1
6460155 Nagasaki et al. Oct 2002 B1
6473072 Comiskey et al. Oct 2002 B1
6476834 Doval et al. Nov 2002 B1
6493734 Sachs et al. Dec 2002 B1
6502756 Fahraeus Jan 2003 B1
6509893 Akhlagi et al. Jan 2003 B1
6516181 Kirwan Feb 2003 B1
6529920 Arons et al. Mar 2003 B1
6532314 Plain et al. Mar 2003 B1
6535799 Levanoni et al. Mar 2003 B2
6556188 Cordner Apr 2003 B1
6564249 Shiigi May 2003 B2
6577299 Schiller et al. Jun 2003 B1
6584249 Gu et al. Jun 2003 B1
6587859 Dougherty et al. Jul 2003 B2
6592039 Smith et al. Jul 2003 B1
6593908 Borgstrom et al. Jul 2003 B1
6608618 Wood et al. Aug 2003 B2
6609653 Lapstun et al. Aug 2003 B1
6627870 Lapstun et al. Sep 2003 B1
6628847 Kasabach et al. Sep 2003 B1
6641401 Wood et al. Nov 2003 B2
6644545 Lapstun et al. Nov 2003 B1
6647369 Silverbrook et al. Nov 2003 B1
6651879 Lapstun et al. Nov 2003 B2
6661405 Flowers Dec 2003 B1
6663008 Pettersson et al. Dec 2003 B1
6665490 Copperman et al. Dec 2003 B2
6668156 Lynch et al. Dec 2003 B2
6676411 Rehkemper et al. Jan 2004 B2
6678499 Silverbrook et al. Jan 2004 B1
6689966 Wiebe Feb 2004 B2
6724373 O'Neill, Jr. et al. Apr 2004 B1
6724374 Lapstun et al. Apr 2004 B1
6732927 Olsson et al. May 2004 B2
6738050 Comiskey et al. May 2004 B2
6738053 Borgstrom et al. May 2004 B1
6750978 Marggraff et al. Jun 2004 B1
6752557 Hsieh Jun 2004 B1
6755584 O'Brien et al. Jun 2004 B2
6763995 Song Jul 2004 B1
6771283 Carro Aug 2004 B2
6773185 Hsieh Aug 2004 B1
6798403 Kitada et al. Sep 2004 B2
6816702 Kuntz et al. Nov 2004 B2
6831632 Vardi Dec 2004 B2
6847883 Walmsley et al. Jan 2005 B1
6853293 Swartz et al. Feb 2005 B2
6874883 Shigemura et al. Apr 2005 B1
6885878 Borgstrom et al. Apr 2005 B1
6886036 Santamaki et al. Apr 2005 B1
6915103 Blume Jul 2005 B2
6933928 Lilienthal Aug 2005 B1
6938222 Hullender et al. Aug 2005 B2
6940491 Carro Sep 2005 B2
6947027 Lapstun et al. Sep 2005 B2
6956562 O'Hara et al. Oct 2005 B1
6965454 Silverbrook et al. Nov 2005 B1
6966495 Lynggaard et al. Nov 2005 B2
6966777 Robotham Nov 2005 B2
6982703 Lapstun et al. Jan 2006 B2
6985138 Charlier Jan 2006 B2
6989816 Dougherty et al. Jan 2006 B1
7006116 Meyers et al. Feb 2006 B1
7035583 Ferrigno et al. Apr 2006 B2
7068860 Kasabach et al. Jun 2006 B2
7080103 Womack Jul 2006 B2
7099019 Silverbrook et al. Aug 2006 B2
7134606 Chou Nov 2006 B2
7155382 Boys Dec 2006 B2
7184592 Iga et al. Feb 2007 B2
7193618 Morehouse Mar 2007 B2
7202861 Lynggaard Apr 2007 B2
7239306 Fahraeus et al. Jul 2007 B2
7289110 Hansson Oct 2007 B2
7295193 Fahraeus Nov 2007 B2
7350996 Bielecki et al. Apr 2008 B2
7409089 Simmons et al. Aug 2008 B2
7421439 Wang et al. Sep 2008 B2
7453447 Marggraff et al. Nov 2008 B2
7922099 Schmidt et al. Apr 2011 B1
7936339 Marggraff et al. May 2011 B2
20010015721 Byun et al. Aug 2001 A1
20010024193 Fahraeus Sep 2001 A1
20010051329 Lynch et al. Dec 2001 A1
20020000468 Bansal Jan 2002 A1
20020001418 Fahraeus et al. Jan 2002 A1
20020011989 Ericson et al. Jan 2002 A1
20020021284 Wiebe Feb 2002 A1
20020023957 Michaelis et al. Feb 2002 A1
20020029146 Nir Mar 2002 A1
20020041290 LeKuch et al. Apr 2002 A1
20020044134 Ericson et al. Apr 2002 A1
20020060665 Sekiguchi et al. May 2002 A1
20020076683 Chen Jun 2002 A1
20020077902 Marcus Jun 2002 A1
20020083101 Card et al. Jun 2002 A1
20020087598 Carro Jul 2002 A1
20020113802 Card et al. Aug 2002 A1
20020113823 Card et al. Aug 2002 A1
20020118230 Card et al. Aug 2002 A1
20020120854 LeVine et al. Aug 2002 A1
20020193975 Zimmerman Dec 2002 A1
20020197589 Wood et al. Dec 2002 A1
20030001020 Kardach Jan 2003 A1
20030013073 Duncan et al. Jan 2003 A1
20030013483 Ausems et al. Jan 2003 A1
20030014615 Lynggaard Jan 2003 A1
20030016210 Soto et al. Jan 2003 A1
20030016212 Lynggaard Jan 2003 A1
20030020629 Swartz et al. Jan 2003 A1
20030024975 Rajasekharan Feb 2003 A1
20030025951 Pollard et al. Feb 2003 A1
20030028451 Ananian Feb 2003 A1
20030029919 Lynggaard et al. Feb 2003 A1
20030040310 Barakat et al. Feb 2003 A1
20030046256 Hugosson et al. Mar 2003 A1
20030052900 Card et al. Mar 2003 A1
20030067427 Comiskey et al. Apr 2003 A1
20030071850 Geidl Apr 2003 A1
20030080948 Lapstun et al. May 2003 A1
20030087219 Berger et al. May 2003 A1
20030089777 Rajasekharan et al. May 2003 A1
20030090477 Lapstun et al. May 2003 A1
20030095098 Paul et al. May 2003 A1
20030112220 Yang et al. Jun 2003 A1
20030133164 Tsai Jul 2003 A1
20030134257 Morsy et al. Jul 2003 A1
20030162162 Marggraff Aug 2003 A1
20030173405 Wilz, Sr. et al. Sep 2003 A1
20030195820 Silverbrook et al. Oct 2003 A1
20030208410 Silverbrook et al. Nov 2003 A1
20030218604 Wood et al. Nov 2003 A1
20030234824 Litwiller Dec 2003 A1
20040012198 Brotzell et al. Jan 2004 A1
20040022454 Kasabach et al. Feb 2004 A1
20040023200 Blume Feb 2004 A1
20040029092 Orr et al. Feb 2004 A1
20040039750 Anderson et al. Feb 2004 A1
20040043365 Kelley et al. Mar 2004 A1
20040043371 Ernst et al. Mar 2004 A1
20040084190 Hill et al. May 2004 A1
20040091842 Carro May 2004 A1
20040104890 Caldwell et al. Jun 2004 A1
20040121298 Creamer et al. Jun 2004 A1
20040140966 Marggraff et al. Jul 2004 A1
20040164975 Ho et al. Aug 2004 A1
20040167895 Carro Aug 2004 A1
20040169695 Forman Sep 2004 A1
20040202987 Scheuring et al. Oct 2004 A1
20040219501 Small et al. Nov 2004 A1
20040229195 Marggraff et al. Nov 2004 A1
20040259067 Cody et al. Dec 2004 A1
20050002053 Meador et al. Jan 2005 A1
20050005246 Card et al. Jan 2005 A1
20050013487 Clary et al. Jan 2005 A1
20050022130 Fabritius Jan 2005 A1
20050024322 Kupka Feb 2005 A1
20050024346 Dupraz et al. Feb 2005 A1
20050055628 Chen et al. Mar 2005 A1
20050060644 Patterson Mar 2005 A1
20050082359 Marggraff et al. Apr 2005 A1
20050083316 Brian et al. Apr 2005 A1
20050106547 Chiu May 2005 A1
20050131803 Lapstun et al. Jun 2005 A1
20050134926 Takezaki et al. Jun 2005 A1
20050135678 Wecker et al. Jun 2005 A1
20050138541 Euchner et al. Jun 2005 A1
20050165663 Razumov Jul 2005 A1
20050188306 Mackenzie Aug 2005 A1
20050198208 Nystrom Sep 2005 A1
20050208458 Smith et al. Sep 2005 A1
20050211783 Chou Sep 2005 A1
20050214730 Rines Sep 2005 A1
20060033725 Marggraff et al. Feb 2006 A1
20060067576 Marggraff et al. Mar 2006 A1
20060067577 Marggraff et al. Mar 2006 A1
20060080609 Marggraff Apr 2006 A1
20060125805 Marggraff Jun 2006 A1
20060126105 Sedky et al. Jun 2006 A1
20060127872 Marggraff Jun 2006 A1
20060146029 Diercks Jul 2006 A1
20060159345 Clary et al. Jul 2006 A1
20060168261 Serval et al. Jul 2006 A1
20060242562 Wang et al. Oct 2006 A1
20060269168 Kasabach et al. Nov 2006 A1
20070003168 Oliver Jan 2007 A1
Foreign Referenced Citations (58)
Number Date Country
1142471 Jun 1999 CN
1520542 Aug 2004 CN
1655184 Aug 2005 CN
0495618 Jul 1992 EP
0519714 Dec 1992 EP
0539053 Apr 1993 EP
0697780 Feb 1996 EP
0866397 Sep 1998 EP
0973314 Jan 2000 EP
1256090 Nov 2002 EP
1256091 Nov 2002 EP
1315085 May 2003 EP
1416426 May 2004 EP
1681624 Jul 2006 EP
2811130 Jan 2002 FR
2202664 Sep 1988 GB
57238486 Aug 1980 JP
5137846 Jun 1993 JP
5217688 Jun 1993 JP
6146516 May 1994 JP
H06231466 Aug 1994 JP
H08036452 Feb 1996 JP
1011639 Apr 1998 JP
11119790 Apr 1999 JP
2000247074 Sep 2000 JP
2000293303 Oct 2000 JP
2001184291 Jul 2001 JP
2002297308 Oct 2002 JP
2003528402 Sep 2003 JP
2004503840 Feb 2004 JP
2007296387 Nov 2007 JP
2002009615 Feb 2002 KR
20020033775 May 2002 KR
9957648 Nov 1999 WO
0073983 Dec 2000 WO
0101670 Jan 2001 WO
0116691 Mar 2001 WO
0126032 Apr 2001 WO
0148685 Jul 2001 WO
0161455 Aug 2001 WO
0171475 Sep 2001 WO
0167222 Sep 2001 WO
0169917 Sep 2001 WO
0171653 Sep 2001 WO
0171743 Sep 2001 WO
0175723 Oct 2001 WO
0175773 Oct 2001 WO
0175780 Oct 2001 WO
0183213 Nov 2001 WO
0186612 Nov 2001 WO
0195559 Dec 2001 WO
0242894 May 2002 WO
03001357 Jan 2003 WO
03001475 Jan 2003 WO
03067553 Aug 2003 WO
03083763 Oct 2003 WO
03094489 Nov 2003 WO
2004084190 Sep 2004 WO