The present invention relates to a method for a camera based electronic device and in particular for a camera based mobile communication device comprising an optical detection, a determination of possible keywords or text-parts and a further selection of a keyword or the text-part as text input for a current software application instead of typing said text in to said device.
In particular the method replaces a manual typing of text into a mobile communication device by capturing an image of a printed page containing said text followed by a semi-automatic detection and selection of said text, which can be a keyword or a text-part, and using said text as text input for a current application running on the mobile communication device. As used in the context of the present description a page is intended as a printed page.
Mobile communication devices with integrated camera module are further approaching the market and show up already a good market penetration and use in daily life, taking over the market of simpler mobile phones. Far beyond phoning and email reading they get used for taking pictures, for Internet access and for other services including WAP services and camera based services. With an increasing interaction between the user and the mobile communication device and with increasing complexity of input data, the interface between the user and the mobile communication device gets more important. Better mobile phones and mobile communication devices from Research in Motion (Blackberry) and from Nokia (Communicator) for instance comprise a more complete keypad as they were mainly intended to be used for email reception and email sending besides a normal phoning. Following and latest mobile communication devices for instance from Apple (iPhone), Nokia, Motorola and others comprise an even bigger screen as touch-screen which allows a more convenient Internet use, looking at pictures as well as such touch-screens give a much better data input possibility. The term “smart-phone”, for sake of clarity, is also found in the market and is represented herein by the term mobile communication device. Speech recognition via built in microphone and speech recognition algorithms can be used for selecting phonebook names and for a small amount of control commands. However in mobile communication devices the touch-screen or the built-in keypad is necessarily small and it is still cumbersome to input data as steadily new applications on the market require even more interaction with an increasing amount of data input.
As an integration of camera modules in mobile phones and in mobile communication devices is already state of the art, there easier text input possibilities would be appreciated, as for instance using the build in camera module for text recognition and for an optical input of text.
From EP-08 169 713 by the same inventors as of the current invention there is disclosed a method and a device for using a mobile communication device as a reading device for blind or visually impaired people. Therefore an image of the text page, which shall be read, is captured semi-automatically, the image data are then transformed via optical character recognition (OCR) to text data and the text data are then converted via text-to-speech software to audio data which are then output via a loudspeaker of the mobile communication device. The text data within a book page, a newspaper, a menu or the like get instantly recognized and read as an image of said text is captured, wherein a filtering inclusive shadow compensation, rotation of the text, binarization, unwarping and the like for a following OCR and text-to-speech conversion is processed. In essence EP-08169713 describes a method for image filtering of captured images containing text, wherein the image is captured with camera modules of a limited quality, as they are built in mobile phones or the like, adapted for optimized OCR results.
From EP-09 161 549 of the same inventors as of the current invention there is disclosed a method for capturing an image, processing the image data, extracting text and objects thereof and sending this data via the same camera based mobile communication device to a server system for translations, for best price searches, for getting information about the local position and the like. There are described data communication channels with possible applications, which require input of respective keywords or text-parts as search targets. But keyword or text-part as search target has still to be typed in manually via the keypad of the mobile communication device.
WO 2008/063822 A1 discloses and describes a method for a mobile communication device with integrated camera, wherein a video stream is analyzed in real time to detect a word within a central region of the image which is indicated by a cross-hair in the display. The so detected word is indicated then by a background color or by highlighting the word, which can be selected then as text input. This method is very useful as alternative text input possibility to a keypad. But the mobile communication device with its cross-hair has to be targeted on the desired keyword to be detected precisely which might be difficult sometimes. In fact, while holding a camera module, little angle variations of the camera module can effect big position variations of the captured text within the camera image. A keypress for capturing the image can often result in enough movement of the mobile communication device, such that the desired keyword is out of the center point of the captured image and so is not detected anymore.
Reading a phone number on a text page which shall be used then for a phone call requires sometimes typing a thirteen digit long number (inclusive country code) in, which might be cumbersome, in particular for some senior people having no good short term memory anymore. So an easy optical method for a number recognition out of a text page and the conversion of said number as usable input for a mobile phone or the like would be an advantage.
WO 2005/101 193 discloses a system comprising a scanner and at least one display and/or a speaker to provide the user of the scanner an indication of actions available for a portion of a document from which scanned information is obtained. Therein the scanned information of the document is used either to identify the document among documents in a database or to identify via markups in the document next possible user actions, indicating the next possible actions on the display or via loudspeakers. Said disclosure comprises an intelligent identification process for the scanned information of the document. But not foreseen therein is a method applicable on mobile communication devices for a quick determination and indication of keywords within a scanned text and highlighting the keywords for a further selection.
It would also be desirable to have a solution, wherein a user could hold a camera based mobile communication device over a text page whereupon an image containing the desired text-part is captured, the text-part gets detected and selected, whereof possible keywords get determined for a further selection of the desired keyword, and whereupon the further selected keyword gets taken as text input for a search application or for the like.
The invention is set forth and characterized in the main independent claims, while dependent claims describe other advantageous characteristics of the invention.
The objective of the present invention is to overcome the shortcomings explained above and to provide a camera based method for text input and detection of a keyword displayed on any page or screen by simply directing a camera module on a section of the page or the screen containing the keyword.
The above objectives and purposes as well as further objectives and purposes which will become apparent from the following description are achieved by the features described in the independent claim and by additional features and characteristics described in the independent claims.
As apparent from the foregoing, the present invention provides a camera based method fox text input and keyword detection out of any page, screen or display containing said keyword, wherein the camera module gets directed on the page, the image of the camera module containing said keyword or a text-part. The method runs on electronic devices with a connected or integrated camera module, preferably on mobile communication devices.
A solution of a preferred embodiment according the present invention is disclosed in the following drawings with detailed description but it shall not be limiting the invention.
In order to achieve a more robust recognition of word blocks 3 (to be described herein below with respect to
The artifact reduced image 2b gets analyzed for image sections which contain preferably one single word, further called word blocks 3 (
Margin words 23 which are touching a margin of the captured image 2a or respectively of the artifact reduced image 2b get discarded. Image objects as for instance lines 20, 21 and 22 in the artifact reduced image 2b get analyzed as far as they are near a word. In fact, such image objects can be an underline as shown with reference numeral 24 in
Therefore an OCR analysis of each word block 3 is performed to get its text content, whereas each text content as word is inserted into a word table 40 (
The keyword probability determination rule takes into account preferably a database which contains filler words which have a very low probability parameter and words with a very high probability parameter. According to an application, for a translation dictionary for instance, the database could contain filler words which would have a low-medium probability parameter, for an online search application the filler words would in this case have a probability parameter of zero. Filler words contained in the database are for instance articles, adverbs, attributes and the like. It shall be mentioned that the database preferably can be selected among a set of databases according to the current application or that it can change itself within the current application as a database with learning characteristics. It is also imaginable that the probability parameters of word groups in the database get modified by the keyword probability determination rule according to the current application. Furthermore, it is imaginable that the database with its words gets downloaded or updated from a server system according to the current application or a certain field of interest, wherein for instance words with high occurrence get a respective high probability parameter.
The keyword probability determination rule takes into account preferably for each word also the distance of the respective word/word block 3 in regards to the center point 6.
The probability calculation for each word according the keyword probability determination rule results in values above or under a threshold, whereas preferably all words above the threshold are determined to be possibly the keyword 7, whereas the respective word block 3 is then determined to be a preferred A-block 4 with a higher probability containing the keyword 7, which gets indicated for a further selection. The threshold gets preferably calculated taking into account all probabilities of the detected words/word blocks 3.
The word table with the sorted words and rows according to their calculated probability, wherein the highest probability is on top, is shown in
For the further selection of the keyword 7 among the A-blocks 4 by the user, the A-blocks get assigned by attributes 5 (only some are indicated in
The word “Mongolei”, to which the attribute “1” is assigned and which is contained in A-block 10, 11 and 12, has the calculated highest probability to be the keyword 7 within the text of the captured image 2a, but despite of that it is not the desired keyword. The word “Volksrepublik”, to which the attribute “2” is assigned and which is contained in the A-blocks 13 occurs three times within the text and has the next highest probability to be the keyword 7. The words of A-blocks 14, 15 and 7 are close to the center point 6 of the captured image 2a, respectively to the displayed image 2c and get so a relatively high probability according to the preferred keyword probability determination rule in this example.
For instance for an application, wherein a phone number shall be detected and selected for making a phone call with that number, it is imaginable, that the keyword probability determination rule detects only numbers with more than 6 digits and less than 15 digits or the like, preferably prioritizing numbers which have a word “phone” in front or above, wherein the corresponding blocks 3 are then selected as A-blocks 4 for the further selection.
Another preferred method for a keyword detection is shown in
In case of an application in which a text-part comprising multiple words shall be detected as text input, the text-part gets selected by detection and selection of a first keyword 7.1, which is a starting word of the text-part and then by detection and selection of a second keyword 7.2, which is the last word of the text-part. The first 7.1 and the second keyword 7.2 get detected and selected as described before. An example is shown in
Another preferred method for a text-part detection and selection is shown in
Preferably the method described above can also comprise a zoom function for the camera module before or after the image 2a is captured, whereas it is imaginable that it can be zoomed into the captured image 2a resulting in that a section of the captured image 2a is further processed and filtered, whereupon within the section of the captured image 2a the word blocks 3 get detected and further processed as described above.
Preferably the keyword probability determination rule determines words which are hand underlined with the highest priority within the artifact reduced image 2b.
The indication of A-blocks 4 is preferably performed as overlaid blocks in a certain color under the written possible keywords but other indications are also imaginable as frames around the A-blocks 4, that the displayed word therein is written in a different color and/or as bold letters or as big capital letters or the like. As attributes 5 numbers are preferred but characters or numbers followed by characters after the first nine numerals are also imaginable, as illustrated for instance in
The further selection of the keyword 7, which is designated by its attribute 5 or by an indication of the respective A-block 4, is preferably performed by a keypress of the respective key of the keypad, whereas the keypad can also be a touch-screen. It is imaginable that the displayed and desired A-block 4 (respectively the keyword 7) can be also selected by arrow keys, a scroll-wheel or the like, wherein the selection jumps from A-block 4 to A-block 4 to the left, to the right, up or down in respect to the pressed key or scroll-wheel. A further selection of the respective attribute 5 via speech recognition is also imaginable. As a further alternative for the further selection of the keyword 7 or the corresponding A-block one use direct selection via touch screen if the mobile communication device includes the latter.
Upon the further selection of the keyword 7 or of the corresponding A-block a further correction of the displayed text is also imaginable which can be performed as usual by character keys or via T9-word corrective software which is available on most mobile communication devices 30.
The present invention shall not be limited to mobile communication devices 30 whereas it is also imaginable that the method gets applied on PCs, wherein a camera module is installed over a text page and the user has only to point out a certain keyword with his fingertip which gets detected and automatically selected for a translation for instance or for the like.
Moreover, the present invention can be applied also for automatic reading machines wherein a word which gets pointed out by a fingertip gets detected, converted into word characters and output by a loudspeaker device.
Furthermore it is imaginable to use this kind of rapid camera based keyword detection and selection for instance for quick search applications for word and text-block translations, for encyclopedia information, and for information related to the keyword, as for instance showing up in the display 32 manufacturers which are connected to the further selected keyword by a server database and the like.
Furthermore, although the present invention has been described in connection with certain specific embodiments of a mobile communication device 30 and flowcharts, the present invention is not limited thereto. Accordingly various modifications, adaptations and combinations of various features of the described embodiments can be practiced without departing from the scope of the present invention as set forth in the accompanying claims.
Where technical features mentioned in any claim are followed by reference signs, those reference signs have been included just for the sole purpose of increasing intelligibility of the claims and accordingly, such reference signs do not have any limiting effect on the scope of each element identified by way of example by such reference signs.
Number | Date | Country | Kind |
---|---|---|---|
10161624 | Apr 2010 | EP | regional |
Number | Name | Date | Kind |
---|---|---|---|
5384863 | Huttenlocher et al. | Jan 1995 | A |
5491760 | Withgott et al. | Feb 1996 | A |
5511148 | Wellner | Apr 1996 | A |
5642518 | Kiyama et al. | Jun 1997 | A |
5649222 | Mogilevsky | Jul 1997 | A |
7069506 | Rosenholtz et al. | Jun 2006 | B2 |
7330608 | Berkner et al. | Feb 2008 | B2 |
7450960 | Chen | Nov 2008 | B2 |
7787693 | Siegemund | Aug 2010 | B2 |
8098934 | Vincent et al. | Jan 2012 | B2 |
8208737 | Ie | Jun 2012 | B1 |
8713433 | Ouyang et al. | Apr 2014 | B1 |
20030169924 | Nishiyama et al. | Sep 2003 | A1 |
20040189804 | Borden et al. | Sep 2004 | A1 |
20050052558 | Yamazaki et al. | Mar 2005 | A1 |
20050169527 | Longe et al. | Aug 2005 | A1 |
20060017752 | Kurzweil et al. | Jan 2006 | A1 |
20060098874 | Lev | May 2006 | A1 |
20070047813 | Simske et al. | Mar 2007 | A1 |
20070106468 | Eichenbaum et al. | May 2007 | A1 |
20080118162 | Siegemund | May 2008 | A1 |
20090177653 | Hirabayashi | Jul 2009 | A1 |
20090227283 | Pylvanainen | Sep 2009 | A1 |
20100172590 | Foehr et al. | Jul 2010 | A1 |
20100293460 | Budelli | Nov 2010 | A1 |
20110081948 | Shirai et al. | Apr 2011 | A1 |
20110249900 | Thorn et al. | Oct 2011 | A1 |
20110267490 | Goktekin et al. | Nov 2011 | A1 |
20120131520 | Tang et al. | May 2012 | A1 |
20120134590 | Petrou et al. | May 2012 | A1 |
20130113943 | Wormald et al. | May 2013 | A1 |
20130234945 | Goktekin | Sep 2013 | A1 |
Number | Date | Country |
---|---|---|
2 189 926 | May 2010 | EP |
2 410 465 | Jan 2012 | EP |
2 383 970 | Jul 2013 | EP |
2 637 128 | Sep 2013 | EP |
WO 0225575 | Mar 2002 | WO |
WO 2005101193 | Oct 2005 | WO |
WO 2006002706 | Jan 2006 | WO |
WO 2007082536 | Jul 2007 | WO |
WO 2008063822 | May 2008 | WO |
Entry |
---|
Ho, Jason, “CapnTrans Demo,” Retrieved from the Internet: http://www.youtube.com/watch?v=QRZe3qV9gBO, Retrieved on: Jun. 18, 2012, XP007920719, whole document. |
“iPhone User Guide—for iPhone OS 3.1 Software,” Apple, Inc., Dec. 31, 2009, pp. 31-36. |
Iketani, A, et al., “Video Mosaicing Based on Structure from Motion for Distortion-Free Document Digitization,” Computer Vision ACCV, Part II, LNCS 4844, pp. 73-84 (2007). |
Nakajima, N, et al., “Video Mosaicing for Document Imaging,” Proc. CBDAR, Demo Session, pp. 171-178 (2007). |
Number | Date | Country | |
---|---|---|---|
20110267490 A1 | Nov 2011 | US |