Method and apparatus for accessing electronic data via a familiar printed medium

Information

  • Patent Grant
  • 8485445
  • Patent Number
    8,485,445
  • Date Filed
    Thursday, September 6, 2012
    12 years ago
  • Date Issued
    Tuesday, July 16, 2013
    11 years ago
Abstract
The disclosed methods and apparatus relate generally to the electronics media industry, such as cable television (CATV), home shopping services, on-line computer services and computer memory applications. These methods and apparatus allow a user to access and make use of electronic media input and output devices by reference to and/or utilization of standard printed matter, such as magazines, textbooks, or any other printed matter that can be correlated to electronic media.
Description
FIELD OF THE INVENTION

The present invention relates generally to the electronic media industry—such as cable television (CATV), home shopping services, on-line computer services, personal computer applications, and CD-ROM-based multi-media applications—and, more particularly, to a method and apparatus for allowing a user to access and make use of such electronic media services via a device that makes use of a standard book, magazine or other printed work.


BACKGROUND OF THE INVENTION

It is a well-known fact that a large fraction of the population is unfamiliar with the operation or use of modern computer devices. This remains true despite significant advances in user-interface technology, such as mouses, windows, menus and the like. It is commonly said that such technology makes computers “user friendly.” However, the modern mouse/window/menu-based user-interface technology has been widely available for several years, and yet it still appears that this technology is not a panacea for computer-phobia. Indeed, studies have shown that a majority of VCR owners cannot operate the simple menu-based interface used to program their VCRs. Thus, there exists a great need for improved user-interface technology, if computer-based electronic media is to become as widely used and accepted as television or radio.


This need for improved user-interface technology will become even greater as technology such as the “information superhighway,” continues to be developed and implemented throughout the world. The “information superhighway” permits delivery of high-bandwidth (i.e., full-motion video rate) digital data into millions of homes via fiber optic, cable, Radio Frequency (RF), microwave or satellite links. As such technology continues to improve, in concept, it will permit instant delivery of a virtually limitless selection of commercial, informational, educational and entertainment programming at a user's request. However, without better user-interface technology, the average user may not be capable of enjoying the vast capabilities of such a system. Indeed, it is clear that the number of programming choices available to the user of such technology will be far greater than the number of choices involved in programming a VCR—a task already demonstrated to be too complicated for the average user of present-day user-interface technology.


In contrast to the difficulty many people encounter in using remote controls and other prior art computer interfaces, printed matter—such as books and magazines—represents an almost universally familiar and non-intimidating medium by which a user can acquire desired information. Even illiterate individuals incapable of reading text can nonetheless peruse pages of printed matter and appreciate the substance of flashy advertisements and the like. Thus, it would be highly desirable to provide a system with the information accessing capabilities of a modern CD-ROM, USB storage device, personal computer, or on-line computer system, and the user-interface simplicity of printed matter.


The prior art includes a class of devices known as “talking books”—see, e.g., U.S. Pat. Nos. 4,636,881 entitled “Talking Book With an Infrared Detector Used to Detect Page Turning”; 4,702,573 entitled “Visual and Audible Activated Work and Method of Forming Same”; 4,778,391 entitled “Sound-Producing Amusement or Educational Devices”; 4,809,246 entitled “Sound Illustrated Book Having Page Indicator Circuit”; 4,990,092 entitled “Talking Book”; and 5,209,665 entitled “Interactive Audio Visual Work”—all of which are incorporated herein by reference. Typically, these “talking books” consist of a book with various sensors which—when activated by touching, page turning, etc.—cause a sound generating means (also embedded within the book) to produce or replay particular sounds. Talking books thus provide an interface for allowing an unsophisticated user (i.e., a child) to access a very primitive computer (i.e., the sound generating means embedded within the book) via familiar printed matter (i.e., the book with embedded sensors). Importantly, however, talking books do not provide a means for interfacing with modern electronic media—such as cable television (CATV), home shopping services, on-line computer services, CD-ROM-based multi-media applications, interactive TV or home computer applications.


One approach to interfacing with these modern electronic media is the “simulated book”—see, e.g., U.S. Pat. No. 4,855,725, entitled “Microprocessor Based Simulated Book,” incorporated herein by reference. The “simulated book” is in essence a book-size intelligent graphics terminal. Unlike the talking books, the simulated book is not a self-contained system, but rather transmits commands to and receives data from a CD-ROM equipped personal computer via a wireless link. Thus, the programming that the simulated book can access is not limited to that which can be stored in embedded memory devices, as with the talking books. Importantly, however, the user-interface provided by the simulated book is essentially the conventional computer interface—i.e., keys, pointer, menus, etc. Therefore, a computer-phobic user will likely still find the simulated book intimidating and inaccessible.


A disadvantage of both the talking book and simulated book technologies is that both include relatively costly electronics—i.e., microprocessors, memory, display devices, etc.—as a part of the “book.” Thus, these technologies cannot be effectively used to create a “throw-away” interactive magazine, newspaper or advertising brochure.


Thus, there remains a need for a method and apparatus for accessing the vast resources of electronic media using a device as familiar and non-intimidating as printed matter. There remains a further need for such a method and apparatus which utilizes a low cost, throw-away printed matter.


Therefore, in the case of a user accessing electronic data from a remote server, it would be desirable to produce a method by which said user can tailor the retrieval of electronic data without uploading personal profile information or registering a user profile.


SUMMARY OF THE INVENTION

One object of the present invention is a method and apparatus for allowing a user to access electronic media via a printed matter.


Another object of the invention is a method and apparatus for allowing a user to access electronic media relating to, or expanding upon, material presented in the printed matter.


Another object of the invention is a low cost, throw-away printed matter useful in connection with other objects of the invention.


Still another object of the invention is an improved method of providing electronic media services.


Yet another object of the invention is an intelligent controller for use in connection with the invention.


Still another object of the invention is to individualize the media retrieved by recognizing a machine-recognizable feature in printed matter.


Yet another object of the invention is a method and apparatus for allowing a user to select desired content associated with a printed matter.


In accordance with one embodiment, the invention comprises: (i) a printed matter having at least one sensor and a transmitter associated therewith; and (ii) an intelligent controller having a receiver and a means for accessing programming material. A user triggers said sensor through interaction with said printed matter, for example, by touching a particular spot on a page or by turning a page. In response to the triggering of said sensor, the transmitter sends a signal indicative of said sensor. The receiver receives said signal and, in response thereto, the intelligent controller executes a pre-programmed command related to accessing or controlling electronic media or programming. For example, when the user triggers a sensor associated with an advertisement in the printed matter, the intelligent controller may, in response, send a signal via a telephone line, cable connection, wireless modem or cellular link to a remote video server, and thereby cause a promotional program to appear on the user's television.


In accordance with another embodiment, the invention comprises: (i) a printed matter having at least one machine-recognizable feature—such as a bar code or magnetic strip (or any commonly used printed indicia, such as a printed character, symbol or pictorial icon), (ii) a feature recognition unit having a means for recognizing said feature and a transmitter, and (iii) an intelligent controller having a receiver. The user directs the feature recognition unit to a feature on said printed matter. In response, the recognition unit transmits a signal indicative of the identity of the particular feature. The receiver receives the signal and the intelligent controller, in response thereto, executes an appropriate pre-programmed command.


In accordance with another embodiment, the invention comprises: (i) a printed matter having at least one machine-recognizable feature—such as a bar code or magnetic strip (or any commonly used printed indicia, such as a printed character, symbol or pictorial icon), which contains a plurality of distinct data portions, whereby each data portion is embedded with a unique command or link, (ii) a feature recognition unit having a machine recognizing device configured to recognize the feature, a display and a transmitter, and (iii) an intelligent controller having a receiver. The user directs the feature recognition unit to a feature on said printed matter. In response, the recognition unit decodes the plurality of data portions and displays a menu to the user. The menu is preferably comprised of a plurality of different links for the user to select from, wherein each link had been previously encoded in its respective data portion within the machine-recognizable feature. Once the user selects the desired link, the recognition unit transmits a signal indicative of the identity of the particular printed matter and the selection made by the user. The receiver receives the signal and the intelligent controller, in response thereto, executes an appropriate pre-programmed command that causes the desired data to be retrieved and displayed to the user.


In accordance with another embodiment, the invention comprises: (i) a printed matter having at least one machine-recognizable feature—such as a bar code or magnetic strip (or any commonly used printed indicia, such as a printed character, symbol or pictorial icon), (ii) a feature recognition unit having a machine recognizing device configured to recognize the feature, a display, a memory to store individual preferences and a transmitter, and (iii) an intelligent controller having a receiver. The user directs the feature recognition unit to a feature on said printed matter. In response, the recognition unit decodes the machine-recognizable feature and transmits a signal indicative of the identity of the particular printed matter and the individual's preferences. The receiver receives the signal and the intelligent controller, in response thereto, executes an appropriate pre-programmed command that causes data to be retrieved and displayed to the user.


In accordance with an embodiment of the invention, the encoding of a plurality of URLs reflecting a plurality of web sites and/or web pages in the machine recognizable feature and providing said user with the opportunity to select which web site and/or web pages containing user preferred information he or she would like to access before transmitting a command to the processing means associated with said remote server. In accordance with another embodiment, the invention comprises: (i) a printed matter having at least one machine-recognizable feature and (ii) an intelligent feature recognition unit having means for recognizing said feature, means for associating said recognized feature with a command, and means for issuing said command over a wireless link. In use, the user directs said intelligent feature recognition unit at a feature on said printed matter. In response, said intelligent recognition unit associates said feature with a pre-programmed command and issues a pre-programmed command sequence over a wireless data link to control or access electronic media services. The command may, for example, be transmitted via an infrared (IR) or ultrasound link to a CATV control box in the same room, or via a cellular or satellite link to the CATV company office.


In accordance with another embodiment, the invention comprises a printed matter having: (i) at least one sensor, (ii) a control module and (iii) a transmitter associated therewith. In response to the triggering of said sensor, said control module directs the transmitter to transmit a command related to accessing or controlling an electronic media service.


Other aspects of the invention relate to methods of providing, accessing or utilizing electronic media services. In accordance with one such aspect, the invention involves: (i) providing a printed matter having at least one sensor associated therewith, (ii) providing an intelligent controller which, in response to the triggering of said sensor, performs a pre-programmed command, and (iii) executing said pre-programmed command to access or control an electronic media.


Another aspect of the invention involves a method of providing electronic media services, which includes the steps of: (i) providing printed matter to a potential customer and (ii) pre-programming an intelligent controller to access or control an electronic media service in response to an event wherein the customer interacts with the printed matter in a particular manner. Advantageously, said printed matter comprises a low cost, throw-away publication.


In accordance with another aspect of the invention, an improved method of providing shop-at-home services includes the steps of: (i) providing to the customer a printed catalogue having at least one sensor or machine-recognizable feature associated therewith, (ii) programming a controller to execute a pre-programmed command in response to an event wherein the customer interacts with said sensor or feature, and (iii) providing a service—e.g., displaying promotional programming on the customer's television, contacting the customer by telephone, establishing a computer “chat” link, etc.—by telephone, cable, or wireless link in response to the execution of said command.


Another aspect of the present invention relates to an improved method of instruction, including the steps of: (i) providing an instructional printed matter—such as a textbook, cookbook, children's book or manual—having at least one sensor or machine-recognizable feature associated therewith, (ii) providing a means, distinct from said textbook or other printed matter, for executing a pre-programmed command in response to an event wherein a reader interacts with said sensor or feature, and (iii) in response to said command, causing or controlling: (a) the electronic delivery or presentation of information related to that in the textbook or other printed matter; and/or (b) the establishment of a communication link to a live tutor or consultant familiar with the subject matter contained in the instructional printed matter.


Another aspect of the invention relates to a low cost, throw-away printed matter including at least one machine-recognizable feature adapted for use in connection with the invention.


It is an object of this invention to provide a means for the encoding of a plurality of URLs reflecting a plurality of web sites and/or web pages in the machine recognizable feature and providing said user with the opportunity to select which web site and/or web pages containing user preferred information he or she would like to access before transmitting a command to the processing means associated with said remote server.


It is another object of the present invention to disclose means of individualizing the information retrieved by scanning machine recognizable features in printed matter and other objects or materials for purposes of delivering the most appropriate content.


It is another object of the present invention to disclose means by which a user device interacts with a machine recognizable feature in a printed matter to permit a user to select desired content retrieved from a remote server associated with the particular printed matter.


It is another object of the present invention to disclose means by which individual preferences can be stored and later accessed for the purpose of retrieving specific information relevant to the interests of particular users.


It is another object of the present invention to provide means for automatically linking stored user preferences with tailored programming material comprising desired information.





BRIEF DESCRIPTION OF THE DRAWINGS

The above—as well as other—aspects, objects and features of the present invention will be described in the Detailed Description below, which is intended to be read in conjunction with the following set of drawings, in which:



FIG. 1 depicts an embodiment wherein the display unit is embedded within the printed matter;



FIG. 2 depicts an embodiment of the invention wherein electronic media is presented on a user's TV set;



FIG. 3 depicts an embodiment of the invention wherein programming material is accessed from a remote source;



FIG. 4 depicts an embodiment of the invention which includes a feature recognition unit;



FIG. 4
a depicts an embodiment of the invention wherein the feature recognition unit provides an interface between the display unit and a remote source of programming material;



FIG. 4
b depicts an embodiment of the invention wherein the display unit is embedded within feature recognition unit;



FIG. 4
c further depicts an embodiment of the invention adapted for presentation of material selecting portions of the information encoded in the machine recognizable features;



FIG. 5 depicts an embodiment of the invention wherein the display unit comprises a personal computer;



FIG. 6 depicts an embodiment of the invention adapted for commercial and shop-at-home applications;



FIG. 7 depicts a flow chart for the embodiment where the user selects portions of the information encoded in the machine recognizable features.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

In this section, the various preferred embodiments of the invention are described from two general perspectives. The first, a “functional” perspective, focuses on the contemplated interactions between the user and the various components—i.e., the printed matter, controller, display unit, etc.—of the invention. This functional description provides the insight needed to implement the software or firmware used in connection with the invention. The second perspective, the “apparatus” view, describes the various technologies that can be used to implement the individual components of the present invention.


The Functional Perspective

Reference is now made to FIG. 1, which depicts an embodiment of the invention comprising a printed matter 1 in communication (preferably wireless) with a data server 2. Printed matter 1 includes at least one sensor 3 and a controller, which preferably includes a microprocessor 4. A stiff or flexible page 5 (any page within a book) holds a display screen 6. Printed matter 1 can take the form of a book, magazine, manual, musical score, catalog, advertisement, newspaper, telephone or electronic service directory, or other like means. The controller—including microprocessor 4—is preferably embedded within the spine or any other page of printed matter 1. Display screen 6 can be an LED display, a passive or active matrix LCD display or other like means, and may also have an audio transducer associated therewith. Sensor 3 is preferably touch sensitive, but can also be a page sensor or a combination of touch and page sensor, as described below. In response to a user's touch, microprocessor 4 causes programming material to be retrieved from data server 2 and displayed on screen 6. Data communication between server 2 and microprocessor 4 may operate via RF cellular, microwave, IR, optical, conductive, telephonic or CATV links, or any combination of these or other like means.


Reference is now made to FIG. 2, which depicts an alternative embodiment of the invention comprising a printed matter 21, an intelligent controller 24 and display unit 25. Printed matter 21 includes at least one sensor 22 and a transmitter 23. In response to a user's actuation of sensor 22, transmitter 23 transmits a coded signal indicative of the identity of the actuated sensor. A receiver 26 in intelligent controller 24 receives the coded signal. Controller 24 then identifies the actuated sensor, and initiates display of appropriate programming material on display unit 25 (which can be a TV set or any other means for audio or audiovisual presentation, including but not limited to a personal computer). Programming material preferably derives from a mass storage device—e.g., a magnetic disk, CD-ROM, ROM, flash RAM, PCMCIA card or other memory means—associated with intelligent controller 24 (or with display unit 25). As used herein, the term “memory means” shall also include future storage technologies, such as the recently announced multi-layer CD-ROMs being developed by IBM. See “New I.B.M. Laser Method Stacks Data on Disks,” New York Times, May 13, 1994.


Reference is now made to FIG. 3, which depicts an embodiment of the invention in which intelligent controller 31 derives programming material from a remote server 30. Controller 31 includes means for accessing a remote server 30 of programming material, and preferably further includes means for decompressing compressed programming material received from server 30. In response to a coded signal received by receiver 26, intelligent controller 31 sends an appropriate command to server 30 to select and initiate transfer of appropriate programming material. Controller 31 then receives programming material from remote server 30 and prepares (e.g., decompresses, if necessary) the material for presentation on display unit 25. As with conventional pay-per-view CATV services and on-line computer services, the user is billed according to the volume and/or nature of programming material requested.


Reference is now made to FIG. 4, which depicts an embodiment of the invention including a printed matter 40, a feature recognition unit 42, an intelligent controller 31, a server 30 and a display unit 25. Printed matter 40 includes at least one machine-recognizable feature 41 in the printed work, such as a bar code, invisible bar code, magnetic code, printed character, symbol or pictorial icon, or other feature. Feature recognition unit 42 is a hand-held device and includes means 44 for recognizing feature 41, a transmitter 45 and an optional user actuation switch 43. The user directs recognition unit 42 at a feature and depresses actuation switch 43. In response thereto, means 44 for recognizing “reads” (i.e., for a printed feature, scans and identifies) the feature 41, and transmitter 45 transmits a coded signal indicative of the identity of feature 41. The operation of controller 31, server 30 and display unit 25 is otherwise identical or similar to the FIG. 3 embodiment.


Still referring to FIG. 4, use of embedded features 41—as opposed to sensors—lowers the fabrication cost of printed matter 40. Advantageously, printed matter 40 can be a low cost, throw-away publication.


An identification code generated either by recognition unit 42 or intelligent controller 31 allows the user to be billed for his/her use of transmitted material by server 30 and, for commercial applications, allows the advertiser to identify the potential customer.


Reference is now made to FIG. 4a, which depicts an embodiment of the invention in which the functions of the intelligent controller are integrated into a recognition/control unit 46. Unit 46 includes means 44 for recognizing a feature 41 and means 48 for accessing—preferably via a cellular RF link—programming material associated with a server 30. In addition, the recognition/control unit 46 optionally includes a microprocessor. Either the recognition/control unit 46 or the display unit 47 may include means for decompressing compressed programming material.


Reference is now made to FIG. 4b, which depicts an embodiment of the invention adapted for selecting portions of the information encoded in the machine recognizable features. The embodiment includes the feature recognition unit which is integrated into a recognition/display unit 46. The unit 46 includes means 44 for recognizing a feature 41 and means 48 for accessing programming material associated with a server 30. In addition, the unit 46 includes means 80 for displaying the programming material that was received via the unit 46 or intelligent controller. Additionally, the recognition/display unit 46 optionally includes a microprocessor. The recognition/display unit 46 or intelligent controller may include means for decompressing compressed programming material. A plurality of URLs are displayed on the display 80. The plurality of URLs reflect a plurality of web sites and/or web pages in the machine recognizable feature. The plurality of URLs provide the user with the opportunity to select which web site and/or web pages containing user preferred information he or she would like to access before transmitting a command to the processing means associated with said remote server.


Reference is now made to FIG. 4c, which shows display 80 in two different stages. The plurality of URLs are presented on the left. After the URL that the user is interested in is selected (in this case ‘Family Resorts’) another list of URLs is shown on the right. This list only contains URLs that are included in the ‘Family Resorts’ group of URLs. Therefore, not all of the URLs need to be downloaded.


Reference is now made to FIG. 5, which depicts an embodiment of the invention utilizing a personal computer to access and display electronic programming material. In this embodiment, portions of the personal computer provide the functions of the previously-described intelligent controller, while other portions of the computer provide the functions of the display unit. The scope of this invention should not be limited to the traditional notions of Personal Computers. The present invention will find use in conjunction with more non-traditional applications such as WEB-TV®, Video Phones, as well as other chip based technologies. An interface unit 60 includes a receiver 26 for receiving a coded signal from a transmitter 23, representative of an actuated sensor 22 (or, in an alternative embodiment which includes a feature recognition unit, from the feature recognition unit indicative of the selected feature) on printed matter 21. Interface unit 60 connects to personal computer 61 by any conventional means, such as an I/O port, card slot, etc. Personal computer 61 monitors the interface unit 60, and displays (or controls the display of) appropriate programming material selected by the user through his/her interaction with printed matter 21.


Programming material can be stored on personal computer 61—on a ROM, CD-ROM, flash RAM, PCMCIA card, or other disk/card supplied along with printed matter 21—or accessed from a remote data server 62.


Reference is now made to FIG. 6, which depicts an embodiment of the invention adapted for commercial and shop-at-home applications. The customer views and interacts with a printed matter 70 (via sensors or a feature recognition unit, both as previously described). Printed matter 70 is preferably a throw-away catalog or advertising brochure listing commercial items, such as programming choices, merchandise, travel or event schedules, or television and/or radio programming schedules. The user interacts with printed matter 70 (i.e., with certain features and/or sensors therein) to cause a coded signal 71 indicative of the customer's selection to be transmitted (either by a transmitter embedded in the printed matter or within a feature recognition unit) to controller 73. In response thereto, controller 73 utilizes a modem/telephone (or other wired or wireless communication) link 74-75 to communicate the customer's selection to a commercial provider's remote office 78 (or to a centralized data bank or information superhighway).


The commercial provider preferably uses a “caller ID” function to identify the customer and can respond in a number of ways: (i) have a representative contact the customer; (ii) send the ordered merchandise or tickets to the customer; (iii) direct a data server 72 to provide programming or additional promotional material via a CATV line 76 to controller 73, which replays the material on the customer's TV set; or (iv) download and execute an interactive merchandise selection program on the customer's personal computer or any other electronic media input, output or sensory stimulating device.


Reference is now made to FIG. 7, which shows a flow chart of an embodiment of the invention wherein a user: (i) selects programming material associated with a feature on a printed matter; and (ii) controls or interacts with a personal computer during the presentation and/or execution of the programming material. The recognition device may be comprised of one of many possible devices already in use within the computer and electronics industry.


After starting in step 722, the user captures information from printed matter in step 724. In step 726, menu options are retrieved by the device and displayed to the user. In step 728, the user selects the desired option to be retrieved, e.g., by scrolling through a list displayed on his cell phone display unit and pressing down on one of them. Then in step 730, the device transmits the user selection to the content provider. This is done by transmitting a command sequence to the processor associated with the remote server. Then, in step 732, the content provider transmits electronic data to the device in response to the user's selection. To do this, the command sequence instructs the processor which electronic data to access and also instructs the processor where to send said electronic data. Then in step 734, the device displays the selected content to the user that is related to the printed matter.


In one example shown in FIGS. 4b and 4c, a user sees an advertisement for “Vacation Packages” in a printed matter 40, such as a newspaper or magazine, that contains a 2D barcode 41 or other printed indicia, such as an icon or symbol. The user may have an interest in a family vacation package to a resort that offers a waterpark. The user engages a digital camera 44 housed in a cell phone 46 to capture a barcode 41 or other indicia in the printed matter 40, whereupon through a software processing application housed in the cell phone 46 containing the scanning means, a response menu appears on the cell phone's display 80 showing a range of retrievable content, such as “Honeymoon Packages,” “Beachfront Resorts,” “Family Resorts,” “Family Resorts with Waterparks,” etc. The user then selects one or more of these response options, initiating a preprogrammed command sequences that causes the desired electronic data to be retrieved from an online source and displayed on the cell phone display.


In another example similar to that shown in FIGS. 4b and 4c, a user sees an advertisement for a movie of interest, or a movie poster, containing a 2D barcode or other printed indicia, such as an icon or symbol. By using the camera in the user's cell phone to capture the barcode or other indicia, the user is presented with a set of response options on the device's display, such as “See Movie Trailer,” “Read Reviews,” “Find Nearby Theater Locations,” and “Order Tickets.” The user then selects “See Movie Trailer” and the desired content is displayed on the cell phone display. A prompt may then appear indicating the availability of additional information, such as “Buy Tickets Now.” Upon selecting that option, the user is directed to an online ticket purchasing vendor.


The Apparatus Perspective

The various technologies used to implement the individual functional components that comprise a part of this invention are described below.


Touch Pad Technology


Referring now to FIG. 2, a plurality of touch sensors 22 are embedded within the back binding of the printed matter 21. Alternatively, touch sensors 22 can be positioned along inner or outer margins of the binding, or of particular pages, of printed matter 21 (or along an edge of a mounting tray in which a disposable printed matter is placed). Electronics mounted within printed matter 21 (or within a mounting tray) responds to the actuation of a touch sensor 22 and transmits a coded signal indicative of the identity of the actuated sensor.


Advantageously, touch sensors can be combined with page sensors—as described in U.S. Pat. Nos. 5,209,665; 4,990,092; 4,636,881 and 4,809,246, all previously incorporated herein—to provide a larger number of “effective sensors.” In this embodiment, each time a sensor is actuated, the electronics within the printed matter also checks the page sensor(s) to determine which page is currently being viewed. The electronics then generates a coded signal that identifies both the sensor actuated and the currently visible page. Thus, each (page, sensor) combination represents, in essence, a single “effective sensor.”


Hybrid embodiments combining touch pads and page sensors are also possible. For example, a printed matter might contain a sheet of touch sensors every so many pages, with the page interval being dependent on the stiffness of the pages and the responsiveness of the sensors. That is, if the maximum number of pages through which a user's touch could be reliably detected by a touch sensor was X, then the printed matter would preferably include a sheet of touch sensors every X pages of text.


Intelligent Controllers and Data/Communication Servers


In accordance with several embodiments of the invention, an intelligent controller controls the selection of programming in a video “file server” system. File servers are well known in the art. Generally, as used herein, the term intelligent controller can refer to computer equipment having either: (a) some type of mass storage device, typically a magnetic or magneto-optical disk, that is connected to a network and utilized as central storage for multiple users via the network; and/or (b) some type of network interface which allows establishment of a communication link with other user(s).


Referring to a “file server” type of intelligent controller, the information that is typically stored on such a system consists of binary computer data such as executable programs and corresponding data. Exemplary of the types of corresponding data stored are numbers for spreadsheets, ASCII characters and codes for word processors, and drawing commands for drawing and CAD programs. These or similar types of data do not generally have a time element associated with them, nor do they generally require any extra processing beyond the usual interpretation that is accomplished by the computer program with which they are intended to be used.


In contrast, the playback of audio/video information has a time element associated with its use. Furthermore, the amount of data required to store audio/video program information in “raw” digital form is often prohibitive. Thus, it is common to employ audio/video compression techniques to compress audio/video program information before it is stored. As a result, further processing in the form of decompression and digital-to-analog conversion is required before the real-time viewing of an audio/video program can take place. Decompression may be accomplished through the use of known hardware or software techniques. Digital-to-analog conversion may be required, depending on the type of equipment used for viewing the audio/video program. A known audio/video file server apparatus manufactured and marketed by Protocomm of Trevose, Pa. permits real-time playback of compressed audio/video program information.


Generally, modern audio/video file server systems include: a storage unit for centrally storing compressed digital audio/video program information, a transceiver for receiving compressed digital audio/video program information from an external source over a network (or wireless communication link) to update the stored audio/video program information, a plurality of playback units, each associated with an external playback line and an external playback station, each including a decompression unit, for receiving selected compressed digital audio/video program information from the storage unit, for decompressing the selected compressed digital audio/video program information received from the storage unit, and for playing it in real time over the associated playback line to the associated playback station, a network interface unit for receiving playback requests from the playback units, and a processor for controlling the storage and playback units to play the decompressed selected audio/video program information in real time for viewing by users at selected playback stations.


Such a system is described, for example, in U.S. Pat. No. 5,262,875 entitled “Audio/Video File Server Including Decompression/Playback Means,” which is incorporated herein by reference. The '875 system can be adapted to operate in accordance with the present invention by coupling intelligent controller of the present invention with the '875 system's network interface unit (which provides the playback requests to the playback unit). In the '875 system: “Playback may be initiated through a user request from one of the playback stations or through direct interaction with audio/video file server 5. In the latter case, a simple user interface, such as a menu interface, may be employed for selection of the desired program information and the playback station. For example, a menu of the available items of audio/video program information, as named by the user or by the system, may be presented. The user may make a selection through cursor control keys on a keyboard.”


In contrast, the present invention permits the user to select and control the presentation of audio/video programming through a familiar printed matter interface, as has been previously described. The present invention thus offers the advantage of nearly universal accessibility.


In accordance with the present invention, multimedia programming is preferably delivered from the file server(s) to the replay unit via an Integrated Service Digital Network (ISDN). U.S. Pat. No. 5,206,859 entitled “ISDN Multimedia Communications System,” incorporated herein by reference, describes such a system.


In accordance with the invention, a variety of means can be employed to communicate selection instructions to the video server and to communicate programming material from the server to the user's display unit. Since the selection instructions are very compact (i.e., low bandwidth), these instructions can be communicated via a standard telephone link using an inexpensive, low speed modem. This is similar, for example, to the method cable TV viewers use to select pay-per-view programming from local CATV companies—i.e., the viewer calls a particular toll-free number associated with the desired program, and the CATV office uses a “caller ID” function to determine which customer is to receive the program.


Communication of the programming material from the server to the display unit may require a higher-bandwidth means. No doubt, certain material—such as pricing information for a shop-at-home service or scheduling information for a travel booking service—could be easily transmitted via an inexpensive telephone/modem link. In addition, using the preferred compression/decompression techniques, many multi-media applications could also operate over a conventional telephone link. High quality audio/video programming, however, will likely require a higher bandwidth transmission medium, such as a CATV line, microwave link, DSB link, optical fiber link, cellular radio link, or enhanced bandwidth telephone connection.


Although the invention has been described above with reference to several presently preferred embodiments, such embodiments are merely exemplary and are not intended to define the scope of, or exhaustively enumerate the features of, the present invention. Accordingly, the scope of the invention shall be defined by the following claims. Where a feature or limitation of a preferred embodiment is omitted in a claim, it is the inventors' intent that such claim not be construed to impliedly require the omitted feature or limitation.

Claims
  • 1. A method for displaying programming material on a user device by means of a machine recognizable feature within a printed matter, comprising the steps of: recognizing the machine recognizable feature of the printed matter using a feature recognizing device coupled to a user device;identifying encoded data associated with the machine recognizable feature;transmitting, using the user device, the encoded data associated with the machine recognizable feature to a remote server;receiving the programming material from the remote server, wherein said programming material comprises a plurality of portions of encoded data;displaying said plurality of portions of encoded data on a display of the user device in a humanly recognizable form;selecting at least one of the plurality of portions of encoded data; anddisplaying content associated with the selected portion of encoded data.
  • 2. The method of claim 1, wherein each portion of the plurality of portions of encoded data comprises a link for retrieving content.
  • 3. The method of claim 2, wherein the link is selected from the group consisting of an image data link, a video data link, an audio data link, a programming data link, an online data link, a pictorial data link, an electronic data link, an electronic media link, an “information superhighway” data link, Internet link, a computer network link, a local area network link, a wide area network link, a wireless network link, an Ethernet network link, a Radio Frequency cellular network link, an Integrated Services Digital Network link, a telephone line link, a cable television line link, and any combination thereof.
  • 4. The method of claim 1, wherein said content is retrieved from a web site or a web page associated with the selected portion of encoded data.
  • 5. The method of claim 1, wherein the machine recognizable feature is selected from the group consisting of a printed matter, a code, a barcode, a two-dimensional barcode, a QR code, a magnetic strip, an audible audio emitting device, a biometric sample, a watermark, a digital watermark, a transponder, a radio frequency identification device tag, a touch sensor, an infrared tag, a quantum dot, and any combination thereof.
  • 6. The method of claim 1, wherein the printed matter is selected from the group consisting of a book, a textbook, a newspaper, an editorial, a letter-to-the-editor, an advertisement, a magazine, a passport, a boarding pass, a real estate listing, an identification card, a smart card, a student identification card, a license, a registration, a receipt, a ticket, a business card, a credit card, a cash card, and a document.
  • 7. The method of claim 1, wherein the user device is selected from the group consisting of a personal computer, a hand-held communications device, a cellular telephone, a camera-enabled cellular telephone, a scanner, a personal digital assistant, an Internet-enabled telephone, an Internet-enabled device, an intelligent terminal, a television, a pager, a simulated book, a laptop computer, and a wireless communications device.
  • 8. The method of claim 1, wherein the feature recognizing device is selected from the group consisting of a scanner, a camera, a barcode scanner, a code reader, a magnetic stripe reader, a transceiver, a biometric reading device, a watermark reader, a receiver, a radio frequency identification device reader, an infrared reader, a quantum dot reader, an optical reader, and an optical mouse.
  • 9. The method of claim 1, wherein the remote server comprises at least one selected from the group consisting of a local database, a remote database, a network of databases, and any combinations thereof.
  • 10. The method of claim 1, wherein said at least one of the plurality of portions of encoded data is associated with user location data.
  • 11. A system for displaying programming material on a user device, the system comprising: a printed matter having a machine recognizable feature;the user device including: a feature recognition device configured to recognize said machine recognizable feature;a transmitter configured to transmit a coded signal upon recognition of said machine recognizable feature;a receiver configured to receive said programming material, wherein said programming material comprises a plurality of portions of encoded data;a display configured to display said plurality of portions of encoded data in a humanly recognizable form, and upon a selection of at least one of the plurality of portions of encoded data to display content associated with the selected portion of encoded data; anda remote server including: a receiver configured to receive said coded signal from said user device;a processor configured to access said programming material associated with said machine recognizable feature; anda transmitter configured to transmit said programming material to said user device.
  • 12. The system according to claim 11, wherein each portion of the plurality of portions of encoded data comprises a link for retrieving content.
  • 13. The system according to claim 12, wherein the link is selected from the group consisting of an image data link, a video data link, an audio data link, a programming data link, an online data link, a pictorial data link, an electronic data link, an electronic media link, an “information superhighway” data link, Internet link, a computer network link, a local area network link, a wide area network link, a wireless network link, an Ethernet network link, a Radio Frequency cellular network link, an Integrated Services Digital Network link, a telephone line link, a cable television line link, and any combination thereof.
  • 14. The system according to claim 11, wherein said content is retrieved from a web site or a web page associated with the selected portion of encoded data.
  • 15. The system according to claim 11, wherein said printed matter is selected from the group consisting of a book, a textbook, a newspaper, an editorial, a letter-to-the-editor, an advertisement, a magazine, a license, a registration, a passport, a ticket of admission, a boarding pass, a travel ticket, a business card, an identification card, a real estate listing, a commercial document, and an educational document.
  • 16. The system according to claim 11, wherein said machine recognizable feature is selected from the group consisting of a printed matter, a code, a bar code, a two-dimensional bar code, a magnetic strip, a biometric sample, a watermark, a digital watermark, and a transponder.
  • 17. The system according to claim 11, wherein said feature recognition device is selected from the group consisting of a scanner, a camera, a bar code scanner, a code reader, a magnetic stripe reader, a biometric reading device, a watermark reader, a receiver, and a transceiver.
  • 18. The system according to claim 11, wherein the remote server comprises at least one selected from the group consisting of a local database, a remote database, a network of databases, and any combinations thereof.
  • 19. The system according to claim 11, wherein said at least one of the plurality of portions of encoded data is associated with user location data.
  • 20. A method for displaying programming material on a user device by means of a machine recognizable feature within a printed matter, comprising the steps of: recognizing the machine recognizable feature of the printed matter using a feature recognizing device coupled to a user device;identifying encoded data associated with the machine recognizable feature, wherein said encoded data comprises a preprogrammed command;executing said preprogrammed command, and thereby accessing programming material stored on a database, wherein said programming material comprises a plurality of data links;displaying said plurality of data links on a display of the user device;selecting at least one of the plurality of data links; anddisplaying content associated with the selected data link.
  • 21. The method according to claim 20, wherein each of said data link is selected from the group consisting of an image data link, a video data link, an audio data link, a programming data link, an online data link, a pictorial data link, an electronic data link, an electronic media link, an “information superhighway” data link, Internet link, a computer network link, a local area network link, a wide area network link, a wireless network link, an Ethernet network link, a Radio Frequency cellular network link, an Integrated Services Digital Network link, a telephone line link, a cable television line link, and any combination thereof.
  • 22. The method according to claim 20, wherein said content is retrieved from a web site or a web page associated with the selected data link.
  • 23. A method for providing selected programming material on a user device by means of recognizing a machine recognizable feature associated with a printed matter, comprising the steps of: receiving encoded data associated with the machine recognizable feature of the printed matter obtained by a feature recognizing device;accessing programming materials associated with the machine recognizable feature based on the encoded data, wherein said programming material comprises a plurality of portions of encoded data;transmitting the programming material to the user device;receiving from the user device a portion of encoded data selected from the plurality of portions of encoded data;transmitting content associated with said selected portion of encoded data to the user device.
  • 24. The method of claim 23, wherein said feature recognizing device is coupled to the user device, a third party device, or any combination thereof.
  • 25. The method of claim 23, wherein each portion of the plurality of portions of encoded data comprises a link for retrieving content.
  • 26. The method of claim 25, wherein the link is selected from the group consisting of an image data link, a video data link, an audio data link, a programming data link, an online data link, a pictorial data link, an electronic data link, an electronic media link, an “information superhighway” data link, Internet link, a computer network link, a local area network link, a wide area network link, a wireless network link, an Ethernet network link, a Radio Frequency cellular network link, an Integrated Services Digital Network link, a telephone line link, a cable television line link, and any combination thereof.
  • 27. The method of claim 23, wherein said content is retrieved from a web site or a web page associated with the selected portion of encoded data.
  • 28. The method of claim 23, wherein the machine recognizable feature is selected from the group consisting of a printed matter, a code, a barcode, a two-dimensional barcode, a QR code, a magnetic strip, an audible audio emitting device, a biometric sample, a watermark, a digital watermark, a transponder, a radio frequency identification device tag, a touch sensor, an infrared tag, a quantum dot, and any combination thereof.
  • 29. The method of claim 23, wherein the printed matter is selected from the group consisting of a book, a textbook, a newspaper, an editorial, a letter-to-the-editor, an advertisement, a magazine, a passport, a boarding pass, a real estate listing, an identification card, a smart card, a student identification card, a license, a registration, a receipt, a ticket, a business card, a credit card, a cash card, and a document.
  • 30. The method of claim 23, wherein the user device is selected from the group consisting of a personal computer, a hand-held communications device, a cellular telephone, a camera-enabled cellular telephone, a scanner, a personal digital assistant, an Internet-enabled telephone, an Internet-enabled device, an intelligent terminal, a television, a pager, a simulated book, a laptop computer, and a wireless communications device.
  • 31. The method of claim 23, wherein the feature recognizing device is selected from the group consisting of a scanner, a camera, a barcode scanner, a code reader, a magnetic stripe reader, a transceiver, a biometric reading device, a watermark reader, a receiver, a radio frequency identification device reader, an infrared reader, a quantum dot reader, an optical reader, and an optical mouse.
  • 32. A system for displaying programming material on a user device, the system comprising: a printed matter having a machine recognizable feature;a first device including: a feature recognition device configured to recognize said machine recognizable feature; anda transmitter configured to transmit a coded signal upon recognition of said machine recognizable feature;a remote server including: a receiver configured to receive said coded signal from said user device;a processor configured to access said programming material associated with said machine recognizable feature; anda transmitter configured to transmit said programming material to said user devicethe user device including: a receiver configured to receive said programming material, wherein said programming material comprises a plurality of portions of encoded data;a display configured to display said plurality of portions of encoded data in a humanly recognizable form, and upon a selection of at least one of the plurality of portions of encoded data to display content associated with the selected portion of encoded data.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of application Ser. No. 12/590,557, filed Nov. 10, 2009, which is a continuation-in-part of application Ser. No. 11/786,540, now U.S. Pat. No. 7,712,668, filed Apr. 11, 2007 and application Ser. No. 11/786,704, now U.S. Pat. No. 7,703,683, filed Apr. 12, 2007, both of which are continuations-in-part of application Ser. No. 10/943,798, now U.S. Pat. No. 7,523,868, filed Sep. 17, 2004, which is a continuation-in-part of application Ser. No. 09/365,961, now U.S. Pat. No. 6,866,196, filed Aug. 2, 1999, which is a continuation-in-part of application Ser. No. 08/628,246, now U.S. Pat. No. 5,932,863, filed Apr. 4, 1996, which is a division of application Ser. No. 08/250,799, filed May 25, 1994, now abandoned.

US Referenced Citations (392)
Number Name Date Kind
3673416 Berler Jun 1972 A
3705384 Wahlberg Dec 1972 A
3735350 Lemelson May 1973 A
3826900 Moellering Jul 1974 A
3976995 Sebestyen Aug 1976 A
3989929 Trieber Nov 1976 A
3991299 Chadima, Jr. et al. Nov 1976 A
4005388 Morley et al. Jan 1977 A
4117542 Klausner et al. Sep 1978 A
4118687 McWaters et al. Oct 1978 A
4141078 Bridges et al. Feb 1979 A
4144656 Podkopaev et al. Mar 1979 A
4201887 Burns May 1980 A
4247908 Lockhart, Jr. et al. Jan 1981 A
4251798 Swartz et al. Feb 1981 A
4270182 Asija May 1981 A
4283621 Pembroke Aug 1981 A
4286145 Palmer Aug 1981 A
4286146 Uno et al. Aug 1981 A
4335302 Robillard Jun 1982 A
4359633 Bianco Nov 1982 A
4360798 Swartz et al. Nov 1982 A
4365148 Whitney Dec 1982 A
4418277 Tremmel et al. Nov 1983 A
4419573 von Geldern Dec 1983 A
4437127 Hirose Mar 1984 A
4438432 Hurcum Mar 1984 A
4460120 Shepard et al. Jul 1984 A
4463970 Kaule et al. Aug 1984 A
4465926 Apitz et al. Aug 1984 A
4475153 Kihara et al. Oct 1984 A
4481412 Fields Nov 1984 A
4488035 Withnall et al. Dec 1984 A
4492164 Cassanas et al. Jan 1985 A
4500880 Gomersall et al. Feb 1985 A
4517410 Williams et al. May 1985 A
4528643 Freeny, Jr. Jul 1985 A
4588881 Pejas et al. May 1986 A
4601573 Utsugi Jul 1986 A
4609283 Murata et al. Sep 1986 A
4609358 Sangster Sep 1986 A
4628193 Blum Dec 1986 A
4639606 Boles et al. Jan 1987 A
4654482 DeAngelis Mar 1987 A
4654514 Watson et al. Mar 1987 A
4659213 Matsumoto Apr 1987 A
4672658 Kavehrad et al. Jun 1987 A
4682014 Iwama Jul 1987 A
4711994 Greenberg Dec 1987 A
4723212 Mindrum et al. Feb 1988 A
4734858 Schlafly Mar 1988 A
4751500 Minasy et al. Jun 1988 A
4757348 Rourke et al. Jul 1988 A
4780599 Baus Oct 1988 A
4789983 Acampora et al. Dec 1988 A
4793813 Bitzer et al. Dec 1988 A
4797911 Szlam et al. Jan 1989 A
4812879 Suzuki Mar 1989 A
4813350 Drillick Mar 1989 A
4820167 Nobles et al. Apr 1989 A
4825058 Poland Apr 1989 A
4829540 Waggener, Sr. et al. May 1989 A
4831610 Hoda et al. May 1989 A
4833308 Humble May 1989 A
4839507 May Jun 1989 A
4841132 Kajitani et al. Jun 1989 A
4850009 Zook et al. Jul 1989 A
4855725 Fernandez Aug 1989 A
4866756 Crane et al. Sep 1989 A
4868375 Blanford Sep 1989 A
4897865 Canuel Jan 1990 A
4901073 Kibrick Feb 1990 A
4910672 Off et al. Mar 1990 A
4916441 Gombrich Apr 1990 A
4924078 Sant'Anselmo et al. May 1990 A
4933538 Heiman et al. Jun 1990 A
4937853 Brule et al. Jun 1990 A
4943995 Daudelin et al. Jul 1990 A
4947028 Gorog Aug 1990 A
4952785 Kikuda Aug 1990 A
4954699 Coffey et al. Sep 1990 A
4959530 O'Connor Sep 1990 A
4984155 Geier et al. Jan 1991 A
4992824 Plummer Feb 1991 A
4992940 Dworkin Feb 1991 A
5002491 Abrahamson et al. Mar 1991 A
5003577 Ertz et al. Mar 1991 A
5021984 Meade et al. Jun 1991 A
5027196 Ono et al. Jun 1991 A
5029183 Tymes Jul 1991 A
5031098 Miller et al. Jul 1991 A
5038023 Saliga Aug 1991 A
5047614 Bianco Sep 1991 A
5051779 Hikawa Sep 1991 A
5080399 Olson Jan 1992 A
5083271 Thacher et al. Jan 1992 A
5093873 Takahashi Mar 1992 A
5108052 Malewicki et al. Apr 1992 A
5111196 Hunt May 1992 A
5115326 Burgess et al. May 1992 A
5119188 McCalley et al. Jun 1992 A
5126547 Norman, Jr. Jun 1992 A
5142662 Gump et al. Aug 1992 A
5148297 Ishii et al. Sep 1992 A
5151687 Younger Sep 1992 A
5157687 Tymes Oct 1992 A
5158310 Tannehill et al. Oct 1992 A
5163007 Slilaty Nov 1992 A
5168303 Ikenoue et al. Dec 1992 A
5173851 Off et al. Dec 1992 A
5180192 Herbert Jan 1993 A
5181062 Kazumi Jan 1993 A
5185857 Rozmanith et al. Feb 1993 A
5189291 Siemiatkowski Feb 1993 A
5198644 Pfeiffer et al. Mar 1993 A
5204813 Samph et al. Apr 1993 A
5211287 Weisburn et al. May 1993 A
5222157 Yoneda et al. Jun 1993 A
5222624 Burr Jun 1993 A
5229584 Erickson Jul 1993 A
5235509 Mueller et al. Aug 1993 A
5237156 Konishi et al. Aug 1993 A
5237157 Kaplan Aug 1993 A
5239487 Horejsi et al. Aug 1993 A
5243174 Veeneman et al. Sep 1993 A
5250789 Johnsen Oct 1993 A
5250790 Melitsky et al. Oct 1993 A
5260553 Rockstein et al. Nov 1993 A
5262860 Fitzpatrick et al. Nov 1993 A
5280498 Tymes et al. Jan 1994 A
5280609 MacPhail Jan 1994 A
5284164 Andrews et al. Feb 1994 A
5287181 Holman Feb 1994 A
5287266 Malec et al. Feb 1994 A
5288976 Citron et al. Feb 1994 A
5288977 Amendolia et al. Feb 1994 A
5292004 Cesarini Mar 1994 A
5295064 Malec et al. Mar 1994 A
5295836 Ryu et al. Mar 1994 A
5296688 Hamilton et al. Mar 1994 A
5302990 Satoh et al. Apr 1994 A
5304786 Pavlidis et al. Apr 1994 A
5310349 Daniels et al. May 1994 A
5313052 Watanabe May 1994 A
5315098 Tow May 1994 A
5319454 Schutte Jun 1994 A
5319542 King et al. Jun 1994 A
5324922 Roberts Jun 1994 A
5329106 Hone et al. Jul 1994 A
5331137 Swartz Jul 1994 A
5331547 Laszlo Jul 1994 A
5334821 Campo et al. Aug 1994 A
5337361 Wang et al. Aug 1994 A
5339412 Fueki Aug 1994 A
5340966 Morimoto Aug 1994 A
5340971 Rockstein et al. Aug 1994 A
5340973 Knowles et al. Aug 1994 A
5345071 Dumont Sep 1994 A
5351285 Katz Sep 1994 A
5353219 Mueller et al. Oct 1994 A
5368129 Von Kohorn Nov 1994 A
5369571 Metts Nov 1994 A
5382779 Gupta Jan 1995 A
5383029 Kojima Jan 1995 A
5385371 Izawa Jan 1995 A
5385475 Sudman et al. Jan 1995 A
5386298 Bronnenberg et al. Jan 1995 A
5387783 Mihm et al. Feb 1995 A
5397156 Schach et al. Mar 1995 A
5397410 Handly Mar 1995 A
5398336 Tantry et al. Mar 1995 A
5401944 Bravman Mar 1995 A
5404001 Bard et al. Apr 1995 A
5410642 Hakamatsuka et al. Apr 1995 A
5412191 Baitz et al. May 1995 A
5414252 Shinoda et al. May 1995 A
5418713 Allen May 1995 A
5420403 Allum et al. May 1995 A
5420943 Mak May 1995 A
5424524 Ruppert et al. Jun 1995 A
5430555 Sawada et al. Jul 1995 A
5450491 McNair Sep 1995 A
5451998 Hamrick Sep 1995 A
5452379 Poor Sep 1995 A
5459304 Eisenmann Oct 1995 A
5465291 Barrus et al. Nov 1995 A
5469371 Bass Nov 1995 A
5477042 Wang Dec 1995 A
5478990 Montanari et al. Dec 1995 A
5483052 Smith, III et al. Jan 1996 A
5484992 Wilz et al. Jan 1996 A
5485176 Ohara et al. Jan 1996 A
5490217 Wang et al. Feb 1996 A
5493355 Kazami Feb 1996 A
5500681 Jones Mar 1996 A
5505494 Belluci et al. Apr 1996 A
5506697 Li et al. Apr 1996 A
5513264 Wang et al. Apr 1996 A
5541394 Kouchi et al. Jul 1996 A
5551011 Danby et al. Aug 1996 A
5551021 Harada et al. Aug 1996 A
5555496 Tackbary et al. Sep 1996 A
5559714 Banks et al. Sep 1996 A
5563955 Bass et al. Oct 1996 A
5568136 Hochstein et al. Oct 1996 A
5569868 Leung Oct 1996 A
5570291 Dudle et al. Oct 1996 A
5578797 Hewitt et al. Nov 1996 A
5590173 Beasley Dec 1996 A
5594226 Stegar Jan 1997 A
5594809 Kopec et al. Jan 1997 A
5604640 Zipf et al. Feb 1997 A
5612868 Off et al. Mar 1997 A
5615123 Davidson et al. Mar 1997 A
5623552 Lane Apr 1997 A
5625770 Nomura Apr 1997 A
5635694 Tuhro Jun 1997 A
5640002 Ruppert et al. Jun 1997 A
5640193 Wellner Jun 1997 A
5640726 Fichner-Rathus Jun 1997 A
5644408 Li et al. Jul 1997 A
5663748 Huffman et al. Sep 1997 A
5671282 Wolff et al. Sep 1997 A
5680528 Korszun Oct 1997 A
5682540 Klotz, Jr. et al. Oct 1997 A
5721788 Powell et al. Feb 1998 A
5734823 Saigh et al. Mar 1998 A
5739512 Tognazzini Apr 1998 A
5751590 Cannon et al. May 1998 A
5767496 Swartz et al. Jun 1998 A
5774664 Hidary et al. Jun 1998 A
5774874 Veeneman et al. Jun 1998 A
5778181 Hidary et al. Jul 1998 A
5794217 Allen Aug 1998 A
5797330 Li Aug 1998 A
5801944 Kara Sep 1998 A
5804803 Cragun et al. Sep 1998 A
5821728 Schwind Oct 1998 A
5828730 Zebryk et al. Oct 1998 A
5841978 Rhoads Nov 1998 A
5848413 Wolff Dec 1998 A
5857156 Anderson Jan 1999 A
5862260 Rhoads Jan 1999 A
5866888 Bravman et al. Feb 1999 A
5869819 Knowles et al. Feb 1999 A
5870030 DeLuca et al. Feb 1999 A
5870546 Kirsch Feb 1999 A
5872589 Morales Feb 1999 A
5902353 Reber et al. May 1999 A
5903729 Reber et al. May 1999 A
5905248 Russel et al. May 1999 A
5905251 Knowles May 1999 A
5913210 Call Jun 1999 A
5918214 Perkowski Jun 1999 A
5932863 Rathus et al. Aug 1999 A
5933829 Durst et al. Aug 1999 A
5938727 Ikeda Aug 1999 A
5940595 Reber et al. Aug 1999 A
5948040 DeLorme et al. Sep 1999 A
5950173 Perkowski Sep 1999 A
5963915 Kirsch Oct 1999 A
5963916 Kaplan Oct 1999 A
5971277 Cragun et al. Oct 1999 A
5978013 Jones et al. Nov 1999 A
5978648 George et al. Nov 1999 A
5978773 Hudetz et al. Nov 1999 A
5979757 Tracy et al. Nov 1999 A
5984366 Priddy Nov 1999 A
5991601 Anderson Nov 1999 A
5992752 Wilz, Sr. et al. Nov 1999 A
5995105 Reber et al. Nov 1999 A
6002491 Li et al. Dec 1999 A
6003774 Bard et al. Dec 1999 A
6012102 Shachar Jan 2000 A
6014090 Rosen et al. Jan 2000 A
6027024 Knowles Feb 2000 A
6032195 Reber et al. Feb 2000 A
6045048 Wilz, Sr. et al. Apr 2000 A
6064832 Sato et al. May 2000 A
6064979 Perkowski May 2000 A
6072186 Yokonuma et al. Jun 2000 A
6076733 Wilz, Sr. et al. Jun 2000 A
6085124 Choi Jul 2000 A
6085976 Sehr Jul 2000 A
6095418 Swartz et al. Aug 2000 A
6098065 Skillen et al. Aug 2000 A
6108656 Durst et al. Aug 2000 A
6119935 Jelen et al. Sep 2000 A
6122403 Rhoads Sep 2000 A
6148331 Parry Nov 2000 A
6149441 Pellegrino et al. Nov 2000 A
6152369 Wilz, Sr. et al. Nov 2000 A
6154738 Call Nov 2000 A
6157314 Loftus Dec 2000 A
6164534 Rathus et al. Dec 2000 A
6199048 Hudetz et al. Mar 2001 B1
6213401 Brown Apr 2001 B1
6226412 Schwab May 2001 B1
6246337 Rosenberg et al. Jun 2001 B1
6247047 Wolff Jun 2001 B1
6269336 Ladd et al. Jul 2001 B1
6269342 Brick et al. Jul 2001 B1
6270351 Roper Aug 2001 B1
6286036 Rhoads Sep 2001 B1
6311214 Rhoads Oct 2001 B1
6313732 DeLuca et al. Nov 2001 B1
6314457 Schena et al. Nov 2001 B1
6321991 Knowles Nov 2001 B1
6321992 Knowles et al. Nov 2001 B1
6330244 Swartz et al. Dec 2001 B1
6331864 Coco et al. Dec 2001 B1
6338434 Wilz, Sr. et al. Jan 2002 B1
6340931 Harrison et al. Jan 2002 B1
6345764 Knowles Feb 2002 B1
6353772 Silverbrook Mar 2002 B1
6373551 Manico et al. Apr 2002 B2
6375078 Russell et al. Apr 2002 B1
6377986 Philyaw et al. Apr 2002 B1
6381582 Walker et al. Apr 2002 B1
6384744 Philyaw et al. May 2002 B1
6385583 Ladd et al. May 2002 B1
6386453 Russell et al. May 2002 B1
6421675 Ryan et al. Jul 2002 B1
6446871 Buckley et al. Sep 2002 B1
6448979 Schena et al. Sep 2002 B1
6453301 Niwa Sep 2002 B1
6547726 Pratt et al. Apr 2003 B2
6560741 Gerety et al. May 2003 B1
6568595 Russell et al. May 2003 B1
6641037 Williams Nov 2003 B2
D483806 Fermgard Dec 2003 S
6719470 Berhin Apr 2004 B2
6753883 Schena et al. Jun 2004 B2
6783070 Faria et al. Aug 2004 B2
6796492 Gatto Sep 2004 B1
6824057 Rathus et al. Nov 2004 B2
6827256 Stobbe Dec 2004 B2
6827259 Rathus et al. Dec 2004 B2
6827267 Rathus et al. Dec 2004 B2
6827273 Wilz et al. Dec 2004 B2
6830187 Rathus et al. Dec 2004 B2
6830188 Rathus et al. Dec 2004 B2
6834804 Rathus et al. Dec 2004 B2
D500795 Andersson et al. Jan 2005 S
6843411 Rathus et al. Jan 2005 B2
6843419 Rathus et al. Jan 2005 B2
6866196 Rathus et al. Mar 2005 B1
6868433 Philyaw Mar 2005 B1
6873974 Schutzer Mar 2005 B1
6886750 Rathus et al. May 2005 B2
6951305 Overhultz et al. Oct 2005 B2
6988662 Russell et al. Jan 2006 B2
7016516 Rhoads Mar 2006 B2
7054461 Zeller et al. May 2006 B2
7055737 Tobin et al. Jun 2006 B1
7143946 Rathus et al. Dec 2006 B2
7143947 Rathus et al. Dec 2006 B2
7147160 Rathus et al. Dec 2006 B2
7154056 Bergqvist et al. Dec 2006 B2
7155202 Helal Dec 2006 B2
7162087 Bryborn Jan 2007 B2
7167164 Ericson et al. Jan 2007 B2
7180509 Fermgard et al. Feb 2007 B2
7185824 Hepworth et al. Mar 2007 B2
7213763 Rathus et al. May 2007 B2
7232057 Rathus et al. Jun 2007 B2
7269737 Robinson Sep 2007 B2
7299971 Marggraff et al. Nov 2007 B2
7314178 Rines et al. Jan 2008 B2
7333963 Widrow et al. Feb 2008 B2
7334728 Williams Feb 2008 B2
7337968 Wilz, Sr. et al. Mar 2008 B2
7341191 Russell et al. Mar 2008 B2
7341456 McAdams et al. Mar 2008 B2
7367049 Robinson et al. Apr 2008 B1
7374093 Rathus et al. May 2008 B2
7383984 Silverbrook et al. Jun 2008 B2
7392945 Philyaw Jul 2008 B1
7395969 Rathus et al. Jul 2008 B2
7398133 Wannier et al. Jul 2008 B2
7403968 Lee et al. Jul 2008 B2
7407099 Bhatti et al. Aug 2008 B1
7451932 Rathus et al. Nov 2008 B2
7455233 Rathus et al. Nov 2008 B2
7500596 Rathus et al. Mar 2009 B2
7523868 Rathus et al. Apr 2009 B2
20010011276 Durst, Jr. et al. Aug 2001 A1
20050006466 Overhultz et al. Jan 2005 A1
20050258961 Kimball et al. Nov 2005 A1
20070255652 Tumminaro et al. Nov 2007 A1
20080052168 Peters et al. Feb 2008 A1
20100023426 Wannier et al. Jan 2010 A1
Foreign Referenced Citations (35)
Number Date Country
2452202 Nov 1975 DE
357899 Mar 1990 EP
2494873 May 1982 FR
2109600 Jun 1981 GB
2 232 119 Dec 1990 GB
2300498 Nov 1995 GB
60-46684 Mar 1985 JP
62-98334 May 1987 JP
62264752 Nov 1987 JP
63203045 Aug 1988 JP
63296560 Dec 1988 JP
01226088 Sep 1989 JP
02103025 Apr 1990 JP
02127792 May 1990 JP
02279393 Nov 1990 JP
3060558 Mar 1991 JP
03151263 Jun 1991 JP
03174693 Jul 1991 JP
04-023183 Jan 1992 JP
4289048 Sep 1992 JP
06110913 Apr 1994 JP
6188962 Jul 1994 JP
6-233159 Aug 1994 JP
07056941 Mar 1995 JP
7064169 Mar 1995 JP
10269326 Oct 1995 JP
10269326 Oct 1998 JP
11355699 Dec 1998 JP
02000267966 Sep 2000 JP
02000293455 Oct 2000 JP
2001142971 May 2001 JP
02001265800 Sep 2001 JP
WO-9803923 Jan 1998 WO
WO-9806055 Feb 1998 WO
WO-0055738 Sep 2000 WO
Related Publications (1)
Number Date Country
20130008954 A1 Jan 2013 US
Divisions (1)
Number Date Country
Parent 08250799 May 1994 US
Child 08628246 US
Continuations (1)
Number Date Country
Parent 12590557 Nov 2009 US
Child 13605433 US
Continuation in Parts (6)
Number Date Country
Parent 11786540 Apr 2007 US
Child 12590557 US
Parent 11786704 Apr 2007 US
Child 11786540 US
Parent 10943798 Sep 2004 US
Child 11786704 US
Parent 09365961 Aug 1999 US
Child 10943798 US
Parent 08628246 Apr 1996 US
Child 09365961 US
Parent 10943798 Sep 2004 US
Child 11786540 Apr 2007 US