The disclosure relates generally to a system and method for making a piece of print interactive.
Machine readable codes such as QR Codes, NFC, Watermarks and the like require that the code is designed into the piece before printing or added as a sticker. This restricts its use to new print that must be produced, distributed, placed, hung, installed etc.
Image Recognition systems require a digital file such as a PDF to be uploaded, indexed etc. via a computer before the image can be made interactive. In most instances, if an image is to link to a video, a still image from the video must be printed out to be the ‘marker’ file to access the video through Image Recognition.
There is a need for a more dynamic and user-friendly way to create user-generated interactive print.
The disclosure is particularly applicable to a cloud based system used for images with computing device having applications that interact with the cloud based system and it is in this context that the disclosure will be described. It will be appreciated, however, that the system and method has greater utility since it can be implemented in ways no specifically described below that are within the scope of the disclosure.
The system and method broadly allows print to made digital. Specifically, a digitised version of the printed piece of media is held in a backend of the system along with relevant metadata. When a piece of print media is scanned, such as using a enabled application running on a computing device in one implementation, a print recognition process in the backend searches for a match in a store, such as a database in one implementation, and performs a programmed action to direct the user to the required online destination for the digital media associated with the piece of printed media.
In one implementation, the system may have an application, such as a VideeCards app, on each computing device, such as a Smartphone, Tablet or Computer and then users can manipulate and place a photo, graphic or logo into a range of design templates. The application may also allow the user to edit text and graphics and/or record a video (or link to any existing video already stored on the computing device or other online content). The system also allows the user to place an order for a quantity of printed media that will have been made interactive by the system so that anyone scanning the piece of printed media with the computing device, such as by the VideeCards App being executed on the computing device, may be taken to the video or other predetermined content. The system may also further video content to be recorded and registered within the app such that when the card is subsequently scanned, the user will be taken to the new video. Additionally, the system allows customised and personalised content to be created and given a unique code. The code can be given to the targeted individual recipient such that when the generic card is scanned, the user is now prompted to enter the unique code and is then directed to the video (or other content) made especially for them. Furthermore, assuming explicit opt-in, the captured geo-location, time and handset details can, in real-time, influence the content that is delivered to individual recipients via programmatic conditional logic.
The piece of print media that may be made interactive by the system may include, but is not limited to, posters, paintings, contact cards, business cards, greetings cards, dating cards, resume cards, save the date cards, trading cards, postcards, reply cards, event cards, game cards, gift cards, charity cards etc. The interactive user-generated content concept also applies in the same way to ‘photo-merchandise’ products such as photo-books, canvas prints, hats, mugs, t-shirts, pillows, cushions, bags, fridge magnets, bumper stickers, labels, paintings, coasters, photo-jigsaws. More generally, the system may be used to make any existing piece of printed media interactive.
The communication path 206 may couple the computing devices to the backend and some of the backend systems to each other. The communication path 206 may be a wired communications network or a wired communication network. For example, the communication path 206 may be Ethernet, the web, a wireless data network, a computer network, a wired data network and the like.
The backend 204 may be implemented as one or more computing resources, such as cloud computing resources and server computers and each computing resource may have the typical components of the computing resource including one or more processors, memory, persistent storage and the like. The backend system 204 may further comprise an image recognition component 204a, a cloud component 204b, a video hosting component 204c and an image hosting component 204d. Each of the components of the system may be implemented using one or more computing resources and may be internal to the backend or external to the backend. In one embodiment, the image recognition component 204a, various image recognition providers can be used depending on the content or preference of the customer (for white labeled solutions). In one embodiment, the cloud component 204b may be hosted within Microsoft Azure utilizing both web and database functionality and may be geo-located in terms of redundancy for maximum availability. In one embodiment, the video and image components 204c, 204d may use various Video and Image hosting providers depending on the content and preference of the customer (for white labeled solutions). The images and videos could be stored at a later date within the VideeCards cloud server infrastructure.
The image recognition component 204a may host each file of the users that has a mark in it for the interactive content and may also store/host one or more image recognition processes that are used, based on a scan of a piece of printed media, to detect the mark on the piece of printed media and then provide one or more pieces of interactive content to a recipient of the piece of printed media. The cloud component 204b may perform the user registration and authentication processes, authenticate each computing device and provide a web server that allow a browser application to interact with the backend system. The cloud component 204b may also have a manager component that manages the overall operation of the system. The video and image components 204c, 204d store images and videos, respectively and allow the cloud component 204b to retrieve the images and videos as needed.
The system to create interactive printed media shown in
Generating a Piece of Interactive Printed Media
The user may then record a video (314) which activates and interacts with (including but not limited to) any on-board SD, HD or 3D video camera, external wirelessly connected and controlled remote video camera or other such device. Once recorded or captured, the video is given a unique reference by the system, such as the application running on the computing device, including but not limited to UDID—Unique Device Identifier—username, password, randomly generated and data based identifier consisting of 4/5/6/7/8 letters and numbers of upper and lower case etc. (316.) In one embodiment, the method may assign the unique identifier based on personal data of the user and the content of the video data. This video, upon accepting the legal terms and conditions of use and upon payment, is then ready to be uploaded (318) to the cloud component 204b or an external video-hosting site 204c including (but not limited to) YouTube, Vimeo and Video Streaming services.
The user may then add one or more links to videos, web sites, images or other content (320). The link may be assigned by entering a URL, or pURL (personalised URL generator). This can be content that can be generated on the fly, and based on parameters captured in real-time from the scanning system. The parameters may include, but is not limited to, GPS data, handset type, Operating System, time, repeat or unique visitor, API Key etc. Using conditional logic, unique data can be delivered to the handset—‘if this then do that’. Should the user scanning elect to ‘opt in’ and give personal data (including but not limited to) name, e-mail, physical address, membership number or other such identifier, then the UDID of the handset can be linked to the individual user in a database such that the data delivered upon scanning can be personalised to the known individual based upon API Key and/unique index that is scanned. In this way, unlimited copies of the same printed piece can lead those that scan the print to completely customized and personalized content (322), such as (including but not limited to) targeted coupons, customized rewards, prize draws, personal messages. The content for the user may be stored and indexed using the unique identifier based on personal information of the user, content information and device information (324.)
Once the Print file and video have been created, Ts & Cs accepted and payment is made (326) via (including but not limited to) In-App purchase, PayPal, Visa, MasterCard, Purchasing Card, Points, or given free as a promotion, via product code, sponsorship package etc., the Print file may then be uploaded to the cloud component 204B (328) where the Print file is made interactive (330). The piece of printed media file is interactive and a high resolution Print file is stored in a Print queue and a lower resolution, such as for example a jpeg file, is created and ‘indexed’ using a mark. The mark may be based on an Image Recognition process or an invisible digital watermark such that the visual attributes of the (eventually) printed piece can be subsequently recognized by the ‘scanning’ system(s) within the system together with (including but not limited to) the associated nomenclature, programmed URL destination and metadata. These ‘indexed’ files are associated with an ‘API Key’, (including but not limited to) an alphanumeric database ‘partition’ such that the indexed files on one Key may be isolated from those on another so that the ‘Scanning’ App can be made to be associated only with that Key (or such Keys that are required). In this way, the VideeCards App may be ‘White Labelled’ such that other brands may have their own version under their own brand. Indexed files under one brand may then be private, such that another version of the ‘Scanning’ system could not read another's files, only its own (unless a brand wished for its files to appear in multiple Keys). An example of this would be a branded version of the VideeCards App for Greetings Cards companies where they use their own copyright designs made suitable for use within the App. Similarly, a Football Club (including but not limited to) could have its own branded version where its own licensed merchandise and imagery is uniquely interactive on a private API Key for use only with the Club's version of the App.
When the design and video are finished and paid for, the digital file will be printed, boxed and shipped (332). Upon receipt, the Cards are interactive (334) such that, when scanned with the VideeCards App or relevant branded version, the user will be taken to the specified and/or dynamic online content, based upon the programmed parameters (such as location, time etc).
When presenting a Card to another individual, the Card owner will instruct the recipient how to download and scan the Card, and give a reason and motivation for them to do so. In this way, the recipients would have the capability to create their own Cards. Should a recipient then subsequently order, their UDID could be linked to the first person whose card was scanned, thereby creating an opportunity for a single layer of incentive commission. This will take the form of ‘credits’ against future orders.
For example, John takes a photo of his Auntie Joan at her 60th birthday party, using a WiFi enabled camera, linked to the app on his smartphone. This photo is made interactive, in the same way as above, and linked to a video of Joan at the party, also captured on the camera, through the app. John then orders, and pays for, prints of the photograph, to be sent to family members all over the world. These prints are produced remotely since the app sends the information wirelessly to a professional printer, where the photographs are printed and dispatched according to the delivery information entered on the app. These prints are already interactive, as the image is stored and indexed in the cloud component, so any user who scans the photo with the app is taken to the video of Joan at the party.
Interactivity of Existing Printed Media
In the method for making existing printed media interactive, there may be the following processes:
Image Capture—
Using the app on the computing device, a user creates an interactive image using image recognition technology utilised by the app. This is achieved either by taking a photo of an existing physical image, using any on-board SD, HD, 3D camera on the device (smartphone, tablet, camera) currently in use, or an external wirelessly connected and controlled remote camera or video camera, or other such device, to capture a new image which is to be made interactive, using the interface and functionality of the app. The subject matter can be any relatively planar object, including, but not limited to, photographs, paintings of all varieties, textile items of all varieties (assuming they are of a planar nature) such as clothing and accessories, carpets, wallpapers, tapestries etc., cases, carriers and covers, posters, signs, any printed material etc.
Make Interactive—
Once captured, a low resolution version of this image is uploaded to the components of the backend system in the cloud and an image recognition process is applied, such that the visual attributes of the image can subsequently be recognised by the scanning system of the app.
Assign Destination and Further Content—
The user then assigns the content to be displayed upon scanning of the image. The destination could be any accessible content on the mobile internet, online store, or any online function that is desired. This could take the form of (including, but not limited to) public video or photo sharing platforms such as YouTube, Flickr, etc., private (branded or unbranded) online media storage and sharing platforms etc. The app also has the functionality to create further user-generated content on the fly, that is to say using any on-board SD, HD, 3D camera on the device currently in use, or an external wirelessly connected and controlled remote camera or video camera, or other such device, to capture new content to which the image is to be linked. The new content will be uploaded and stored online, on a public video or photo sharing platform such as YouTube, Flickr, etc., or a private (branded or unbranded) online media storage and sharing platform. The destination, in whatever form, will be accessed from within the app, maintaining the app's appearance, skin, branding, advertising, and any other features.
Read—
When the image is subsequently scanned by any device using the app, the image recognition technology follows the directions given by the image in our database, resulting in the pre-assigned media, content or destination being accessed by the app.
For example, a grandson uses the app on his smartphone to take a photo of an existing, framed photograph of his late grandfather. The app uploads a low resolution version of the image of photograph to our database in the cloud, and the image recognition algorithm is applied. The grandson then uses his wireless-enabled digital video camera, which is wirelessly linked to his phone, on which he is using the app, to record a video of his grandmother's memories of his grandfather. The video is transferred to an online media storage platform through the app and the link created between the image of his grandfather, and the video of his grandmother. The grandson's cousin then scans the copy of the photograph of his grandfather that hangs on his wall, in Australia, using the app on his tablet. The cousin is presented with the video of his grandmother' memories of his late grandfather.
While the foregoing has been with reference to a particular embodiment of the invention, it will be appreciated by those skilled in the art that changes in this embodiment may be made without departing from the principles and spirit of the disclosure, the scope of which is defined by the appended claims.
This application is a continuation of and claims priority and benefit under 35 USC 120 to U.S. patent application Ser. No. 14/518,796, filed Oct. 20, 2014 that is turn is a continuation of and claims priority and benefit under 35 USC 120 to U.S. patent application Ser. No. 13/797,942 filed on Mar. 12, 2013 (issued on Oct. 21, 2014 as U.S. Pat. No. 8,867,106) that in turn claims the benefit and priority under 35 USC 119(e) and 120 to U.S. Provisional Patent Application Ser. No. 61/609,699 filed on Mar. 12, 2012 and entitled “Intelligent Print Recognition”, the entirety of both of which are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
5188624 | Young, Sr. | Feb 1993 | A |
5661517 | Budow | Aug 1997 | A |
5854894 | Lancaster | Dec 1998 | A |
6175663 | Huang | Jan 2001 | B1 |
6587599 | Huang | Jul 2003 | B1 |
6826727 | Mohr | Nov 2004 | B1 |
6914593 | Lapstun et al. | Jul 2005 | B2 |
7406603 | MacKay | Jul 2008 | B1 |
7466444 | Silverbrook | Dec 2008 | B2 |
7495795 | Graham et al. | Feb 2009 | B2 |
7739583 | Barrus et al. | Jun 2010 | B2 |
7746498 | Lapstun et al. | Jun 2010 | B2 |
7812986 | Graham et al. | Oct 2010 | B2 |
7930642 | Gerde | Apr 2011 | B1 |
7962691 | Nishihara | Jun 2011 | B2 |
7982898 | Lapstun et al. | Jul 2011 | B2 |
7991778 | Hull et al. | Aug 2011 | B2 |
8090628 | Shinohara et al. | Jan 2012 | B2 |
8156427 | Graham et al. | Apr 2012 | B2 |
8332401 | Hull et al. | Dec 2012 | B2 |
8335789 | Hull et al. | Dec 2012 | B2 |
8385660 | Moraleda et al. | Feb 2013 | B2 |
8390648 | Ptucha et al. | Mar 2013 | B2 |
8601526 | Nishimuyra | Dec 2013 | B2 |
8635531 | Graham et al. | Jan 2014 | B2 |
8681243 | Nozaki | Mar 2014 | B2 |
8797604 | Hannaway | Aug 2014 | B2 |
8867106 | Lancaster | Oct 2014 | B1 |
9036043 | Seg | May 2015 | B2 |
9092291 | Adib et al. | Jul 2015 | B1 |
9319640 | Ptucha et al. | Apr 2016 | B2 |
9332302 | Briggs | May 2016 | B2 |
9363448 | Seg | Jun 2016 | B2 |
9448625 | Kobayashi | Sep 2016 | B2 |
9656492 | Leynadier | May 2017 | B2 |
9754623 | Lancaster | Sep 2017 | B2 |
9979758 | Duncker | May 2018 | B2 |
20040047294 | Ain | Mar 2004 | A1 |
20050246590 | Lancaster | Nov 2005 | A1 |
20080262911 | Altberg | Oct 2008 | A1 |
20090210790 | Thomas | Aug 2009 | A1 |
20090313658 | Nishimura et al. | Dec 2009 | A1 |
20110157226 | Ptucha et al. | Jun 2011 | A1 |
20140198033 | Kobayashi | Jul 2014 | A1 |
20150270318 | Bhintade | Sep 2015 | A1 |
20160351062 | Mathews | Dec 2016 | A1 |
Number | Date | Country | |
---|---|---|---|
20170365293 A1 | Dec 2017 | US |
Number | Date | Country | |
---|---|---|---|
61609699 | Mar 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14518796 | Oct 2014 | US |
Child | 15694724 | US | |
Parent | 13797942 | Mar 2013 | US |
Child | 14518796 | US |