A most typical method of digitizing and utilizing a paper document is to scan the paper document with a scanner which is an external device connected to a user device (e.g., computer, laptop, tablet) with a cable and to utilize (edit, manipulate, search, and the like) the scanned image rendered on a monitor.
Examples are described in the following detailed description and in reference to the drawings, in which:
Various aspects of the present disclosure are directed to a mobile device with a transparent display and a scanner. More specifically, and as described in greater detail below, various aspects of the present disclosure are directed to a manner by which a mobile device can be used in a transparent display mode to scan documents or identified sections within documents.
Certain terms are used throughout the following description and claims to refer to particular system components. As one skilled in the art will appreciate, computer companies may refer to a component by different names. This document does not intend to distinguish between components that differ in name but not function. In the following discussion and in the claims, the terms “including” and “comprising” are used in an open-ended fashion, and thus should be interpreted to mean “including, but not limited to . . . .” Also, the term “couple” or “couples” is intended to mean either an indirect or direct connection. Thus, if a first device couples to a second device, that connection may be through a direct electrical or mechanical connection, through an indirect electrical or mechanical connection via other devices and connections, through an optical electrical connection, or through a wireless electrical connection. As used herein the term “approximately” means plus or minus 10%. In addition, as used herein, the phrase “user input device” refers to any suitable device for providing an input, by a user, into an electrical system such as, for example, a mouse, keyboard, a hand (or any finger thereof), a stylus, a pointing device, etc.
The following discussion is directed to various examples of the disclosure. Although one or more of these examples may be preferred, the examples disclosed should not be interpreted, or otherwise used, as limiting the scope of the disclosure, including the claims. In addition, one skilled in the art will understand that the following description has broad application, and the discussion of any example is meant only to be descriptive of that example, and not intended to intimate that the scope of the disclosure, including the claims, is limited to that example.
Aspects of the present disclosure described herein disclose allowing the user to scan content by placing the device over text (e.g., in a book, document, website). Among other things, this approach allows interaction with content through one screen by scanning the text and manipulating the scanned text as needed (e.g., search, share). The user can use touch gestures to control the device on the same display. Moreover, aspects of the present disclosure described herein allow utilizing a mobile device with a transparent OLED display which can become opaque as needed and an integrated scanner. The device may be stationed on a base, which can be a stand and a charger for the device in a public setting such as a library, school or office. Among other things, this approach may provide the flexibility of using a single device by a plurality of users.
In one example in accordance with the present disclosure, a method for managing a plurality of instances on a display screen of a display unit is provided. The method comprises receiving, by a processor, a selection of an area of a surface, the surface comprising text, instructing, by the processor, a scanner to scan the text in the selected area of the surface, and overlaying, by the processor, data on the text over the selected area.
In another example in accordance with the present disclosure, a mobile system is provided. The system comprises a transparent display screen, and a scanner head to scan an area seen through the transparent display. Position of the scanner is adjustable with respect to the transparent display. The scanner comprises at least one camera and at least one light source. The mobile system further comprises an outer frame, a cover to house the scanner head along with electronic components. The scanner head glides within the outer frame when scanning the area.
In a further example in accordance with the present disclosure, a non-transitory computer readable medium is provided. The non-transitory computer-readable medium comprises instructions which, when executed, cause a mobile device with a transparent display and a scanner to (i) receive a selection of an area of a surface, the surface comprising text, (ii) instruct the scanner to scan the text in the selected area of the surface, and (iii) overlay data over the selected area on the transparent display.
The device 100 includes a processor 110 (e.g., a central processing unit, a microprocessor, a microcontroller, or another suitable programmable device), a display screen 120, a memory unit 130, input interfaces 140, a communication interface 150, and a scanner 160. Each of these components or any additional components of the display unit 100 is operatively coupled to a bus 105. The bus 105 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures. In other examples, the display unit 100 includes additional, fewer, or different components for carrying out similar functionality described herein.
The processor 110 includes a control unit 115 and may be implemented using any suitable type of processing system where at least one processor executes computer-readable instructions stored in the memory 130. The processor 110 may be, for example, a central processing unit (CPU), a semiconductor-based microprocessor, an application specific integrated circuit (ASIC), a field-programmable gate array (FPGA) configured to retrieve and execute instructions, other electronic circuitry suitable for the retrieval and execution instructions stored on a computer readable storage medium (e.g., the memory 130), or a combination thereof. The machine readable medium 130 may be a non-transitory computer-readable medium that stores machine readable instructions, codes, data, and/or other information. The instructions, when executed by processor 110 (e.g., via one processing element or multiple processing elements of the processor) can cause processor 110 to perform processes described herein.
Further, the computer readable medium 130 may participate in providing instructions to the processor 110 for execution. The machine readable medium 130 may be one or more of a non-volatile memory, a volatile memory, and/or one or more storage devices. Examples of non-volatile memory include, but are not limited to, electronically erasable programmable read only memory (EEPROM) and read only memory (ROM). Examples of volatile memory include, but are not limited to, static random access memory (SRAM) and dynamic random access memory (DRAM). Examples of storage devices include, but are not limited to, hard disk drives, compact disc drives, digital versatile disc drives, optical devices, and flash memory devices. As discussed in more detail above, the processor 110 may be in data communication with the machine readable medium 130, which may include a combination of temporary and/or permanent storage. The machine readable medium 130 may include program memory that includes all programs and software such as an operating system, user detection software component, and any other application software programs. The machine readable medium 130 may also include data memory that may include multicast group information, various table settings, and any other data required by any element of the ASIC.
The communication interface 150 enables the device 100 to communicate with a plurality of networks and communication links. In some examples, the communication interface of the device 100 may include a Wi-Fi® interface, a Bluetooth interface, a 3G interface, a 4G interface, a near filed communication (NFC) Interface, and/or any other suitable interface that allows the computing device to communicate via one or more networks. The networks may include any suitable type or configuration of network to allow the device 100 to communicate with any external systems or devices.
The display screen 120 may be a transparent an organic light emitting diode (OLED) display, or any other suitable display. More specifically, the display screen 120 is transparent, and thus makes the rear background of the display screen be seen through. The display screen 120 may be a flexible display that can be wrapped and unwrapped from around a bar. An attachment section of the display screen 120 facilitates a coupling of flexible display to the bar in any conventional manner. In one implementation, the flexible display may have a magnetic disclosure, and the display wrapped around the bar may be held in place with the magnetic disclosure. Alternatively, a band may be used to hold the wrapped display around the bar. In various implementations, the flexible display screen 120 may have a variety of structural configuration and material composition. The display screen 120 is to display content from one or more applications communicated to the device 100. In one implementation, the display screen 120 comprises various display properties such as resolution, display pixel density, display orientation and/or display aspect ratio. The display screen 120 may be of different sizes and may support various types of display resolution, where display resolution is the number of distinct pixels in each dimension that can be displayed on the display screen 120. For example, the display screen 120 may support high display resolutions of 1920×1080, or any other suitable display resolutions. When the display screen supports a 1920×1080 display resolution, 1920 is the total number of pixels across the height of the display 120 and 1080 is the total number of pixels across the height of the display 120.
The scanner 160 comprises a plurality of color cameras 170 which is arranged to take an image of an object and/or document. The scanner 160 may include a support of which one end includes the camera 170 and another end is connected to a hinge provided on one point of an outer area of the device 100 and which has a predetermined length, and a hinge to be provided at the other end of the support to enable the support to move. The scanner 160 may further include at least one light source (e.g., LED) 180 and a lighting unit such as various kinds of lenses which provides an exterior light to a transparent display. The scanner 160 may be such that the camera 170 can be positioned to be in the front side or in the rear side of the transparent display 120, i.e., in the front side or in the rear side with respect to the device 100. In one implementation, the scanner 160 moves along a surface and scans the content present on the surface. For example, the scanner 160 may move along a document and scan text and images present on the document. In another implementation, a user of the device 100 may highlight a section of a document, and the scanner 160 may scan only that section. Moreover, the LEDs 180 may be on or blinking when the scanner 160 is actively scanning a surface. Further, the LEDs 180 may be off when the scanner 160 is not active. In another implementation, the scanner 160 may be a 3D image camera. As shown in
In one implementation, the camera 160 may communicate the identification of the document to the processor 110 to initiate deriving computer readable text from the images of text via optical character recognition (OCR). The images are displayed on the display screen 120. The text may comprise an e-mail, web-site, book, magazine, newspaper, advertisement, another display screen, or other. Moreover, the device 100 uses augmented reality technology. For example, a layer of computer readable text may be displayed on top of, or overlaid, the original image on the display screen 120. As the device 100 or the text on the document or object in view of the camera 160 moves, the display 120 is automatically updated to show the text currently being viewed by the camera 160. Accordingly, the computer readable text is also updated to correspond to the same currently imaged text. It should be noted while a camera is discussed in this specific implementation, other types of scanners may be incorporated in the device 100.
In one implementation, the device 100 may comprise an audio unit. In one implementation, the audio unit comprises an ambisonic sound system, providing three-dimensional (3D) sound in the environment. More specifically, the audio unit sends a sound signal with spatial information that enables the user to perceive the sound as originating from distinct spatial locations and different directions. In one example, the audio unit may target one user. That is, the audio unit may provide an effect of stereo sound when a single user is positioned within the direction of the speaker. In another example, the audio unit may provide a 3D sound for multiple users regardless of the users' positions.
As discussed above, the device 100 may be connected to other devices via VGA, HDMI, USB, Wi-Fi, Bluetooth, NFC over the local network or over the internet cloud. The other devices may be computing device, which includes one of various computing devices that have a keyboard/battery portion and a display screen portion. The computing devices may include, but not limited, to any one of various desktops, laptops, tablets, smart phones, watches and other similar devices. These devices may operate as a stationary computing device (e.g., personal computers (i.e., desktops), server computers, laptop computers (with permanently attached display screens), all in one devices, and other similar devices that possess comparable characteristics). In other implementations, these devices can be handheld devices, such as tablets and smart phones.
In other implementation, there may be additional components that are not shown in
Moreover, in another implementation, the device 100 may comprise a connection engine, which includes various components for establishing and maintaining device connections, such as computer-readable instructions for implementing communication protocols including TCP/IP. HTTP, Ethernet®, USB®, and FireWire®. The application engine may manage the operation of the instances that are displayed on the display unit. For example, the operation engine receives a command from the user to perform an action on one of the instances being displayed on the display unit. The user communicates the command by touching the instance window on the display screen of the display unit. For example, the display unit may display an image of the text that the camera of the device captures, and the user selects (e.g., moving tip of a finger) a section of the image containing a portion of the text that the camera of the device captures. The operation engine communicates the command (e.g., selection of a section of the image) received from the user to the device, and the device may visually indicate the selection of that section. For example, in one example, the device may highlight the selected section. In another implementation, the device may circle or magnify the selected section. An updated instance may be shown on the display screen, available for the user to further operate or manipulate.
In one implementation, a base may be a stand-alone component. As shown in
In one implementation, a content exploration method may be utilized. More specifically, the device is used to scan a text by a method similar to discussed with references to
In another implementation, a citation mechanism may be utilized. First, the user of the device may select a citation style (e.g., APA, MLA Chicago/Turubian). It should be noted that the user is allowed to make changes in the citation style if desired. In one implementation, the selected citation style may be saved in a database under a user profile. Accordingly, when the user is logged into the device, the device knows what the preferred citation style for that user is. Then, the user selects a section of a surface for scanning. Once the section is selected, the device identifies text in that section. In one example, the user may select the section of the surface by scrolling over the text by a finger. Once the user removes the finger from the text, the device starts a document allowing the user to save the text and initiate the citation mechanism. The mechanism is linked to some electronic packages (e.g., EasyBib, KnightCite, Endnote, and Google) and plagiarism software that are activated when the user saves the text. The packages and software provide an updated version of the text, wherein the text includes the citation style selected by the user. Further, the user is allowed to share the citation through various platforms such as social media and e-mail.
Turning now to the operation of the device 100,
The illustrated process 500 begins at block 505, where a device comprising a transparent display and a scanner receives a selection of an area on a surface which is seen through the transparent display by adjusting a position of at least one camera on the scanner. Further, the device may visually indicate the area on the transparent display. For example, the area may be highlighted or captured in a box. At block 510, the device controls the scanner to scan the surface by adjusting position of the scanner based on the selection of the area. More specifically, the device scans only the area that is identified in the selection. Such selection can be entered by a user of the system. The user may hoover or roll his finger on the area that is desired to be selected. At block 515, the device provides information corresponding to the scanned area and displays such information on the transparent display. For example, the device may allow the user to further manipulate the scanned area. More specifically, the user may choose to share content of the scanned area via a suitable platform (e.g., email, social media) or print or translate such content. In addition, augmented reality technology may be utilized to overlay data on the scanned area.
While the above disclosure has been shown and described with reference to the foregoing examples, it should be understood that other forms, details, and implementations may be made without departing from the spirit and scope of the disclosure that is defined in the following claims.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2016/054824 | 9/30/2016 | WO | 00 |