This patent for letters patent disclosure document describes inventive aspects that include various novel innovations (hereinafter “disclosure”) and contains material that is subject to copyright, mask work, and/or other intellectual property protection. The respective owners of such intellectual property have no objection to the facsimile reproduction of the disclosure by anyone as it appears in published Patent Office file/records, but otherwise reserve all rights.
The present innovations generally address apparatuses, methods, and systems for retail commerce, and more particularly, include TRANSACTION VISUAL CAPTURING APPARATUSES, METHODS AND SYSTEMS (“TVC”).
Consumer transactions typically require a customer to select a product from a store shelf or website, and then to check it out at a checkout counter or webpage. Product information is typically selected from a webpage catalog or entered into a point-of-sale terminal device, or the information is automatically entered by scanning an item barcode with an integrated barcode scanner, and the customer is usually provided with a number of payment options, such as cash, check, credit card or debit card. Once payment is made and approved, the point-of-sale terminal memorializes the transaction in the merchant's computer system, and a receipt is generated indicating the satisfactory consummation of the transaction.
The accompanying appendices and/or drawings illustrate various non-limiting, example, inventive aspects in accordance with the present disclosure:
FIGS. 5A-5F(1) provide exemplary UI diagrams illustrating TVC virtual shopping within embodiments of the TVC;
a-b show data flow diagrams illustrating processing gesture and vocal commands in some embodiments of the TVC;
a-c show logic flow diagrams illustrating processing gesture and vocal commands in some embodiments of the TVC;
a shows a data flow diagrams illustrating checking into a store in some embodiments of the TVC;
b-c show data flow diagrams illustrating accessing a virtual store in some embodiments of the TVC;
a shows a logic flow diagram illustrating checking into a store in some embodiments of the TVC;
b shows a logic flow diagram illustrating accessing a virtual store in some embodiments of the TVC;
a-c show schematic diagrams illustrating initiating transactions in some embodiments of the TVC;
The leading number of each reference number within the drawings indicates the figure in which that reference number is introduced and/or detailed. As such, a detailed discussion of reference number 101 would be found and/or introduced in
The TRANSACTION VISUAL CAPTURING APPARATUSES, METHODS AND SYSTEMS (hereinafter “TVC”) transform mobile device location coordinate information transmissions, real-time reality visual capturing, and mixed gesture capturing, via TVC components, into real-time behavior-sensitive product purchase related information, shopping purchase transaction notifications, and electronic receipts.
Within embodiments, the TVC may provide a merchant shopping assistance platform to facilitate consumers to engage their virtual mobile wallet to obtain shopping assistance at a merchant store, e.g., via a merchant mobile device user interface (UI). For example, a consumer may operate a mobile device (e.g., an Apple® iPhone, iPad, Google® Android, Microsoft® Surface, and/or the like) to “check-in” at a merchant store, e.g., by snapping a quick response (QR) code at a point of sale (PoS) terminal of the merchant store, by submitting GPS location information via the mobile device, etc. Upon being notified that a consumer is present in-store, the merchant may provide a mobile user interface (UI) to the consumer to assist the consumer's shopping experience, e.g., shopping item catalogue browsing, consumer offer recommendations, checkout assistance, and/or the like.
In one implementation, merchants may utilize the TVC mechanisms to create new TVC shopping experiences for their customers. For example, TVC may integrate with alert mechanisms (e.g., V.me wallet push systems, vNotify, etc.) for fraud preventions, and/or the like. As another example, TVC may provide/integrate with merchant-specific loyalty programs (e.g., levels, points, notes, etc.), facilitate merchants to provide personal shopping assistance to VIP customers. In further implementations, via the TVC merchant UI platform, merchants may integrate and/or synchronize a consumer's wish list, shopping cart, referrals, loyalty, merchandise delivery options, and other shopping preference settings between online and in-store purchase.
Within implementations, TVC may employ a virtual wallet alert mechanisms (e.g., vNotify) to allow merchants to communicate with their customers without sharing customer's personal information (e.g., e-mail, mobile phone number, residential addresses, etc.). In one implementation, the consumer may engage a virtual wallet applications (e.g., Visa® V.me wallet) to complete purchases at the merchant PoS without revealing the consumer's payment information (e.g., a PAN number) to the merchant.
Integration of an electronic wallet, a desktop application, a plug-in to existing applications, a standalone mobile application, a web based application, a smart prepaid card, and/or the like in capturing payment transaction related objects such as purchase labels, payment cards, barcodes, receipts, and/or the like reduces the number of network transactions and messages that fulfill a transaction payment initiation and procurement of payment information (e.g., a user and/or a merchant does not need to generate paper bills or obtain and send digital images of paper bills, hand in a physical payment card to a cashier, etc., to initiate a payment transaction, fund transfer, and/or the like). In this way, with the reduction of network communications, the number of transactions that may be processed per day is increased, i.e., processing efficiency is improved, and bandwidth and network latency is reduced.
It should be noted that although a mobile wallet platform is depicted (e.g., see
For example, the CSR app may include features such as described below in the discussion with reference to
Some embodiments of the TVC may feature a more streamlined login option for the consumer. For example, using a mobile device such as iPhone, the consumer may initially enter a device ID such as an Apple ID to get into the device. In one implementation, the device ID may be the ID used to gain access to the TVC application. As such, the TVC may use the device ID to identify the consumer and the consumer need not enter another set of credentials. In another implementation, the TVC application may identify the consumer using the device ID via federation. Again, the consumer may not need to enter his credentials to launch the TVC application. In some implementations, the consumer may also use their wallet credentials (e.g., V.me credentials) to access the TVC application. In such situations, the wallet credentials may be synchronized with the device credentials.
Once in the TVC application, the consumer may see some graphics that provide the consumer various options such as checking in and for carrying items in the store. In one implementation, as shown in
Some embodiments of the TVC application may include a dynamic key lock configuration. For example, the TVC application may include a dynamic keyboard that displays numbers or other characters in different configuration every time. Such a dynamic keypad would generate a different key entry pattern every time such that the consumer would need to enter their PIN every time. Such dynamic keypad may be used, for example, for entry of device ID, wallet PIN, and/or the like, and may provide an extra layer of security. In some embodiments, the dial and scrambled keypad may be provided based on user preference and settings. In other embodiments, the more cumbersome and intricate authentication mechanisms can be supplied based on increased seasoning and security requirements discussed in greater detail in U.S. patent application Ser. No. 13/434,818 filed Mar. 29, 2012 and titled “Graduated Security Seasoning Apparatuses, Methods and Systems,” and PCT international application serial no. PCT/US12/66898, filed Nov. 28, 2012, entitled “Transaction Security Graduated Seasoning And Risk Shifting Apparatuses, Methods And Systems,” which are all herein expressly incorporated by reference. These dynamic seasoned PIN authentication mechanisms may be used to authorize a purchase, and also to gain access to a purchasing application (e.g., wallet), to gain access to the device, and/or the like. In one embodiment, the GPS location of the device and/or discerned merchant may be used to determine a risk assessment of any purchasing made at such location and/or merchant, and as such may ratchet up or down the type of mechanism to be used for authentication/authorization.
In some embodiments, the TVC may also facilitate an outsourced customer service model wherein the customer service provider (e.g., sales associate) is remote, and the consumer may request help from the remote customer service provider by opening a communication channel from their mobile device application. The remote customer service provider may then guide the requesting user through the store and/or purchase.
With reference to
In an alternative implementation, a merchant 220 may optionally provide
33 a store check-in information 206 so that the consumer may snap a picture of the provided store check-in information. The store check-in information 206 may include barcodes (e.g., UPC, 2D, QR code, etc.), a trademark logo, a street address plaque, and/or the like, displayed at the merchant store 220. The consumer mobile device may then generate a check-in request 208 including the snapped picture of store check-in information 206 to the TVC server 210. In further implementations, the store check-in information 206 may include a store floor plan transmitted to the consumer via MMS, wallet push messages, email, and/or the like.
For example, the store information 206 to the TVCconsumer, substantially in the form of XML-formatted data, is provided below:
As another example, the consumer mobile device 203 may generate a (Secure) Hypertext Transfer Protocol (“HTTP(S)”) POST message including the consumer check-in information for the TVC server 210 in the form of data formatted according to the XML. An example listing of a checkout request 208 to the TVC server, substantially in the form of a HTTP(S) POST message including XML-formatted data, is provided below:
JFIF H H ÿâ'ICC_PROFILE ¤appl┐
desc P bdscm ′ Scprt └@ $wtpt └d ¶rXYZ └x ¶gXYZ
aarg A vcgt ...
The above exemplary check-in request message includes a snapped image (e.g., QR code, trademark logo, storefront, etc.) for the TVC server 210 to process and extract merchant information 209. In another implementation, the mobile device 203 may snap and extract merchant information from the snapped QR code, and include such merchant information into the consumer check-in information 208.
In another implementation, the check-in message 208 may further include the consumer's GPS coordinates for the TVC server 210 to associate a merchant store with the consumer's location. In further implementations, the check-in message 208 may include additional information, such as, but not limited to biometrics (e.g., voice, fingerprint, facial, etc.), e.g., a consumer provides biometric information to a merchant PoS terminal, etc., mobile device identity (e.g., IMEI, ESN, SIMid, etc.), mobile component security identifying information, trusted execution environment (e.g., Intel TXT, TrustZone, etc.), and/or the like.
In one implementation, upon TVC server obtaining merchant information 209 from the consumer check-in request message 208, TVC server 210 may query for related consumer loyalty profile 218 from a database 219. In one implementation, the consumer profile query 218 may be performed at the TVC server 210, and/or at the merchant 220 based on merchant previously stored consumer loyalty profile database. For example, the TVC database 219 may be a relational database responsive to Structured Query Language (“SQL”) commands. The TVC server may execute a hypertext preprocessor (“PHP”) script including SQL commands to query a database table (such as
In one implementation, the TVC may obtain the query result including the consumer loyalty offers profile (e.g., loyalty points with the merchant, with related merchants, product items the consumer previously purchased, product items the consumer previously scanned, locations of such items, etc.) 220, and may optionally provide the consumer profile information 223 to the merchant. For example, in one implementation, the queried consumer loyalty profile 220 and/or the profile information provided to the merchant CSR 223, substantially in the form of XML-formatted data, is provided below:
In the above example, TVC may optionally provide information on the consumer's previously viewed or purchased items to the merchant. For example, the consumer has previously scanned the QR code of a product “Michael Kors Flat Pants” and such information including the inventory availability, SKU location, etc. may be provided to the merchant CSR, so that the merchant CSR may provide a recommendation to the consumer. In one implementation, the consumer loyalty message 223 may not include sensitive information such as consumer's wallet account information, contact information, purchasing history, and/or the like, so that the consumer's private financial information is not exposed to the merchant.
Alternatively, the merchant 220 may query its local database for consumer loyalty profile associated with the merchant, and retrieve consumer loyalty profile information similar to message 223. For example, in one implementation, at the merchant 220, upon receiving consumer check-in information, the merchant may determine a CSR for the consumer 212. For example, the merchant may query a local consumer loyalty profile database to determine the consumer's status, e.g., whether the consumer is a returning customer, or a new customer, whether the consumer has been treated with a particular CSR, etc., to assign a CSR to the consumer. In one implementation, the CSR 230 may receive a consumer assignment 224 notification at a CSR terminal 240 (e.g., a PoS terminal, a mobile device, etc.). In one implementation, the consumer assignment notification message 224 may include consumer loyalty profile with the merchant, consumer's previous viewed or purchased item information, and/or the like (e.g., similar to that in message 223), and may be sent via email, SMS, instant messenger, PoS transmission, and/or the like. For example, in one implementation, the consumer assignment notification 224, substantially in the form of XML-formatted data, is provided below:
In the above example, the consumer assignment notification 224 includes basic consumer information, and CSR profile information (e.g., CSR specialty, availability, language support skills, etc.). Additionally, the consumer assignment notification 224 may include consumer loyalty profile that may take a form similar to that in 223.
In one implementation, the consumer may optionally submit in-store scanning information 225a to the CSR (e.g., the consumer may interact with the CSR so that the CSR may assist the scanning of an item, etc.), which may provide consumer interest indications to the CSR, and update the consumer's in-store location with the CSR. For example, in one implementation, the consumer scanning item message 225a, substantially in the form of XML-formatted data, is provided below:
Additionally, the consumer scanning information 225a may be provided to the TVC server to update consumer interests and location information.
Upon receiving consumer loyalty information and updated location information, the CSR terminal 240 may retrieve a list of complementary items for recommendations 225b, e.g., items close to the consumer's in-store location, items related to the consumer's previous viewed items, etc. In one implementation, the CSR may submit a selection of the retrieved items to recommend to the consumer 226, wherein such selection may be based on the real-time communication between the consumer and the CSR, e.g., in-person communication, SMS, video chat, TVC push messages (e.g., see 416a-b in
In one implementation, upon receiving the consumer assignment notification, CSR may interact with the consumer 202 to assist shopping. For example, the CSR 230 may present recommended item/offer information 227 (e.g., see 434d-3 in
In the above example, the location information included in the message 227 may be used to provide a store map, and directions to find the product item in the store floor plan (e.g., see
Continuing on with
In one implementation, the consumer may continue to checkout with a virtual wallet instantiated on the mobile device 203, e.g., see 444b
In one implementation, the TVC server 210 may optionally send a transaction confirmation message 241 to the merchant 220, wherein the transaction confirmation message 241 may have a data structure similar to the purchase receipt 245. The merchant 220 may confirm the completion of the purchase 242. In another implementation, as shown in
In one implementation, merchant store clerks 230a may be notified to their iPad 240 with the customer's loyalty profile. For example, in one implementation, the TVC server 210 may communicate with the merchant payment system 220a (e.g., PoS terminal) via a wallet API 251b to load consumer profile. In one implementation, the TVC server 210 may keep private consumer information anonymous from the merchant, e.g., consumer payment account information, address, telephone number, email addresses, and/or the like. In one implementation, the merchant payment system 220a may retrieve product inventory information from the merchant inventory system 220b, and provide such information to the PoS application of the sales clerk 230a. For example, the sales clerk may assist customer in shopping and adding items to iPad shopping cart (e.g., see 439 in
With reference to
In one implementation, the consumer may snap the QR code and generate a check-in message to the TVC server 310, which may receive the consumer check-in message 309 (e.g., see 208 in
In one implementation, if the consumer visit a merchant shopping site at 303, the consumer may similarly check-in with the merchant by snapping a QR code presented at the merchant site in a similar manner in 308-312. Alternatively, the consumer may log into a consumer account, e.g., a consumer account with the merchant, a consumer wallet account (e.g., V.me wallet payment account, etc.), to check-in with the merchant.
In one implementation, the merchant may receive consumer information from the TVC server (e.g., see 223 in
In one implementation, if a desired CSR match is not locally available 319 (e.g., not available at the merchant store, etc.), the TVC may expand the query to look for a remote CSR 321 which may communicate with the consumer via SMS, video chat, TVC push messages, etc., and allocate the CSR to the consumer based 322.
Alternatively, a pool of remote CSRs may be used to serve consumers and reduce overhead costs. In an alternative embodiment, online consumers may experience a store virtually by receiving a store floor plan for a designated location; and moving a consumer shopper avatar through the store floor plan to experience product offerings virtually, and the remote CSR may assist the virtual consumer, e.g., see
In one implementation, the consumer 302 may receive a check-in confirmation 324 (e.g., see 407 in
In another implementation, if the consumer has a transaction payment request (e.g., see 434g in
In one implementation, TVC server may receive the payment request from the consumer and may request PIN verification 341. For example, the TVC server may provide a PIN security challenge UI for the consumer to enter a PIN number 342, e.g., see 464 in
Continuing on with
In one implementation, the consumer may instantiate the shop 402c option, and check-in with a merchant store. For example, the consumer may operate the wallet application 403 to scan the merchant check-in QR code 404. Continuing on with
For example, with reference to
With reference to
In a further implementation, the consumer wallet may allow a consumer to include an image in the message with CSRs. In one implementation, the consumer may tap a camera icon 423 to snap a picture of an in-store advertisement, a front window display, a poster, etc., and submit the picture to the CSR to indicate the consumer's shopping interests. For example, the consumer may express interests in “Jeans” 427a, and may snap a picture of an in-store commercial poster of “men's jeans” 427b, and ask the CSR about “where to find” the jeans in display 427c.
With reference to
With reference to
With reference to
With reference to
In one implementation, upon submitting a “Pay” request 444b, the TVC may provide a PIN security challenge prior to payment processing to verify the consumer's identity. For example, the TVC may request a user to enter a PIN number 454 via a dial lock panel 455. In alternative implementations, as shown in
With reference to
With reference to
With reference to
With reference to
In another implementation, with reference to
In an alternative implementation, every aisle and shelving stack may include a numerous, wide-angle cameras having a specified accelerometer gyroscopic, positional/directional orientation, periodically taking a photograph of the opposing aisle/area, which may be submitted to the TVC server, so that the virtual store map may be continually updated and be kept up to date. For example, as shown in
In an alternative embodiment, such cameras may provide a continuous live video feed and still photos may be obtained from the live video frame grabs, which may be used to generate virtual store maps. In one implementation, a motion detection component may be used as a trigger to take still photos out of a live videos when the motion detection component detects no motion in the video and thereby provides unobstructed views for virtual map composition. In addition, when a consumer focuses on a particular shelf, aisle, stack, and/or region, e.g., a consumer turns their avatars parallel to a camera directional view, the consumer's view may then become filled with the live video feed of the camera closest to the consumer avatar's location.
In another implementation, as shown in
In further implementations, the consumer may be navigating a merchant's shopping site, having a shopping cart filled with product items, and the remote CSR may join the consumer's shopping session and provide assistance, allowing the CSR to provide the consumer with links to product items that may be of interests to the consumer; this may be achieved by having a CSR help/request button that may generate a pop-up window for audio/video chat with the CSR, and a dialogue box into which the CSR may place a link to the products. The consumer may click on the link provided by the CSR to be directed to a product page to view product details.
In a different embodiment, the users 611a-b may launch a TVC component instantiated on a camera-enabled mobile device 613a-103b to capture a view of the table, e.g., including the received invoice/bill 615 having a quick response (QR) code or barcode printed thereon, and a plurality of payment cards 619a-109b that the users 611a-b are going to pay for the bill. The users 611a-b may view virtual overlaid labels on top of the captured scene, so that they can tap on the option labels to split a bill equally, proportionally, and/or the like.
Within implementations, users 611a-b may facilitate payment from their payment cards upon TVC augmented reality capturing at the same mobile device/wallet. For example, user 611a may operate her mobile device 613a to capture a scene of the two payment cards 619a-b, while card 619b belongs to user 611b. In one implementation, the TVC component instantiated on the mobile device 613a may send an authorization request to a processing server, or a wallet management server to authorize split payment transaction on the payment card 613b. In such scenarios, users 611a-b may conduct a transaction including payments from two wallets on the same mobile device, without user 611b independently initiates a transaction using his mobile device 613b. Further implementations of restaurant bill payment scenarios are illustrated in
Within embodiments, the different layers 715a-d may comprise interdependent information. For example, merchant layer 715a and/or retailer layer 715b may provide information of related products based on user reviews from the social payer 715d. A variety of commerce participants, such as, but not limited to manufacturers, merchants, retailers, distributors, transaction processing networks, issuers, acquirers, payment gateway servers, and/or the like, may bid for layer space in the augmented reality shopping experience.
In one implementation, as shown at 719 in
As another example, a consumer 811 may walk into a merchant store and the mobile device 813 may capture the consumer's GPS coordinates 826. The TVC may then determine the consumer is located at a retailer shop based on the GPS coordinates 827, and may provide a retailer layer of augmented reality overlay labels 829 to the mobile device captured in-store scenes, e.g., including retailer discounts, in-store map, related products inventories, and/or the like.
For example, a user may move a sliding bar 907a to enable or disable a smart finger tip component 903a, e.g., when the smart finger tip component is enabled, the TVC may capture a human finger point within a captured reality scene (e.g., see also 912, etc.), etc. In one implementation, the smart finger tip component 903a may engage fingertip motion detection component (e.g., see
In another example, a user may move the sliding bar 907b to enable or disable auto card detection 903b, e.g., when the auto card detection component is enabled, the TVC may automatically detect and identify whether any rectangular object in a captured reality scene comprise a payment card, etc. In another example, a user may move the sliding bar 907c to enable or disable facial recognition 903c, e.g., when the facial recognition component is enabled, the TVC may automatically recognize human faces (e.g., including a human, a printed facial image on a magazine, a friend's picture displayed on a digital screen, etc.) that are presented in the reality scene and identify whether the human face matches with any of previously stored contacts. In another example, a user may move the sliding bar 907d to enable or disable smart bill tender component 903d, e.g., when the smart bill tender component is enabled, the TVC may provide option labels based on a type of the bill. When the bill is a restaurant bill, the TVC may provide options to facilitate tip calculation, bill splitting per actual consumption, and/or the like. In another example, a user may move the sliding bar 907e to enable or barcode reading component 903e, e.g., the TVC may read a barcode, and/or a QR code printed on a purchase label, invoice or bill to provide payment information via overlaid labels on the captured reality scene.
In one implementation, the user may configure a maximum one-time payment amount 904 via the TVC initiated transaction, e.g., by sliding the bar 905 to select a maximum amount of $500.00. In another implementation, a user may select to include social connections 906 into the TVC capturing component, e.g., the TVC may obtain social data such as user reviews, ratings with regard to a capture purchase item in the reality scene (see 1435 in
Within implementations, when a user places a camera-enabled mobile device (e.g., 913) to capture a reality scene, a user may view a plurality of virtual labels overlaid on top of the captured reality scene. For example, the user may view a sliding bar 910 to control whether to enable the smart finger tip component. As shown in
Continuing on with
With reference to
In one implementation, upon enrolling the card, the TVC may switch back to the visual capturing scene, with an overlaid notification showing the card is ready to use 926, and provide a plurality of overlaid option labels beneath the card 911, such as, but not limited to view balance 927a (e.g., a user may tap and see the current balance of the card), view history 927b (e.g., the user may tap and view recent transaction history associated with the card), transfer money from 927c (e.g., the user may select to transfer money from the card to another account), transfer money to 927d (e.g., the user may transfer money to the card from another account, etc.), pay shopping cart 927e (e.g., the user may engage the card to pay the current shopping cart 908a), and/or the like. Various other option labels related to the card may be contemplated.
In one implementation, if the user selects to tap on the “transfer $$ to” button 927d, with reference to
In one implementation, the user may move his finger to point to another card in the real scene so that the smart finger tip component may capture the payee card. In another implementation, as shown in
With reference to 9E, upon user confirming fund transfer, the TVC may provide a message notifying completion of the transaction 937, and the user may select to view the transaction receipt 938. In one implementation, the TVC may provide a virtual receipt 939 including a barcode 940 summarizing the transaction. In one implementation, the user may email 941 the virtual receipt (e.g., for reimbursement, etc.), or to earn points 942 from the transaction.
In one implementation, upon the facial recognition, the TVC may provide a plurality of option labels overlaid on top of the reality scene, so that the user may select to call the contact 1008a, send a SMS 1008b, email the contact 1008c, transfer funds to the contact 1008d, connect to the contact on social media 1008e, view the contact's published purchasing history 1008f, and/or the like. In one implementation, if the user selects to transfer money to the contact, the TVC may retrieve a previously stored account associated with the contact, or prompt the user to enter account information to facilitate the transfer.
With reference to
In another implementation, when the TVC determines the user tapped portion of the screen comprises a user's DMV license, 1113, the TVC may provide a plurality of option labels, such as view DMV profile 1114a, view pending tickets 1114b, pay ticket 1114c, file a dispute request 1114d, and/or the like.
With reference to
In another implementation, when the TVC determines the user tapped portion comprises a store membership card 1220, e.g., a PF Chang's card, the TVC may provide a plurality of labels including viewpoints 1221a, pay with the card 1221b, buy points 1221d-e, call to order 1221e, and/or the like.
With reference to
In another implementation, when the TVC determines the user tapped portion comprises a bill including a barcode 1326, e.g., a purchase invoice, a restaurant bill, a utility bill, a medical bill, etc., the TVC may provide a plurality of labels including view bill details 1327a, pay the bill 1327b, request extension 1327c, dispute bill 1327d, insurance reimbursement 1327e (e.g., for medical bills, etc.), and/or the like.
With reference to
In one implementation, if the user selects view social rating 1433f of the product, the TVC may retrieve social data from various social media platforms (e.g., Facebook, Twitter, Tumblr, etc.) related to the featured product, so that the user may review other users' comments related to the product.
In one implementation, when the user selects action consumption 1505c, the PVTC may provide tags of the consumed items 1507a-b, e.g., by reading the bill barcode 1502, or by performing OCR on the bill image, etc. In one implementation, a user may drag the item 1507a, e.g., a “bloody Mary” 1508 into the “I Pay” bowl 1510. The user may tap on the plus sign 1509 to increase quantity of the consumed item. In one implementation, the user may tap on a card 1511 to indicate pay with this card for the item in the “I Pay” bowl 1510 as summarized in label 1512. In one implementation, the TVC may provide option labels for tips, including suggested tip percentage (e.g., 15% or 20%) 1513 or enter tip amount 1514.
Continuing on with
With reference to
With reference to
Continuing on with
In one implementation, if the consumer elects “add card” 1542, the consumer may proceed with card enrollment in a similar manner as 215 in
Continuing on with
In one implementation, a user may place two payment cards in the scene so that the TVC may capture the cards. For example, the TVC may capture the type of the card, e.g., Visa 1608a and MasterCard 1608b, and provide labels to show rebate/rewards policy associated with each card for such a transaction 1609a-b. As such, the user may select to pay with a card to gain the provided rebate/rewards.
In an alternative embodiment, as shown in
With reference to
In another implementation, a consumer may slide the information layer 1611a to obtain another layer, e.g., retail information 1611b, social information 1611c, item information 1611d, and/or the like. For example, PVTC may capture a receipt and/or certificate in the scene, and provide information including other Cartier products 1618, purchase item description and price information 1615, retail store inventory information (e.g., stores where the purchase item is available) including physical stores 1623 and online shopping sites 1625, and/or the like.
In further embodiments, a consumer may tap on the provided virtual label of a “Cartier” store, e.g., 1613, 1623, etc., and be directed to a store map including inventory information, e.g., as shown in
With reference to
In another implementation, for additional captured objects 1630 in the scene (e.g., objects without textual contents, etc.), TVC may perform a pattern recognition to provide information of the recognized object 1630. For example, the pattern recognition may be correlated with other contexts within the scene to determine what the captured object is, e.g., the ring shaped object 1630 may be a piece of “Cartier” branded jewelry as the “Cartier” logo is captured in the same scene. In one implementation, the TVC may provide identified item information 1631 in a virtual label, and alternative item recognition information 1632, 1633, 1634. For example, for the ring-shaped product 1630, the TVC may recognize it as a “Cartier” branded bracelet 1631/1632, or ring shaped jewelry products of related brands 1633, 1634, and/or provide an option to the consumer to see more similar products 1635.
In one implementation, the TVC may automatically identify goods that are eligible for restricted-use accounts in a merchant store. For example, the TVC may allow a user to place a camera enabled device at a merchant store (e.g., scanning), and view a camera scene with augmented reality labels to indicate possible items eligible for a restricted-use account.
For example, in one implementation, when the user operate the camera enabled device to obtain a view inside the merchant store 1750, the user may also obtain augmented reality labels 1751 which identifies various products/items on the shelf, and show one or more possible eligible restricted-use accounts 1752. For example, over the counter drugs may be labeled as eligible for “FSA, HSA, HRA,” etc., 1752; grocery products may be eligible for food stamp usage; and infant food may be eligible for a children nutrition benefit account, and/or the like.
In further implementation, if the TVC does not automatically determine an item as eligible for any restricted-use accounts, e.g., an “Ester-C” supplement, a user may tap on the screen to select it, and may view a list of accounts 1863 to select a user desired reallocation account, e.g., any restricted-use account, loyalty account, and/or the like.
In further implementations, the TVC may identify a payment account that has been used to fulfill the transaction associated with the receipt, e.g., a Visa account 1866a, and/or obtain account information from the barcode printed on the receipt 1866b. In one implementation, the TVC may match the “*1234” Visa account with any of user's enrolled account in the wallet, and recommend the user to reimburse funds into an identified “Visa *1234” account if such account is identified from the wallet 1865. In another implementation, the TVC may prompt the user to select other accounts for depositing reimbursement funds 1865.
Continuing on with
In another implementation, if the user selects to tap on 1963 in
In one implementation, upon receiving user finger indication, the TVC may obtain an image of the scene (or the user finger pointed portion) 2006, e.g., grabbing a video frame, etc. In one implementation, the TVC may detect fingertip position within the video frame, and determine an object around the fingertip position for recognition 2007. The TVC may then perform OCR and/or pattern recognition on the obtained image (e.g., around the fingertip position) 2008 to determine a type of the object in the image 2010. For example, in one implementation, the TVC may start from the finger point and scan outwardly to perform edge detection so as to determine a contour of the object. The TVC may then perform OCR within the determined contour to determine a type of the object, e.g., whether there is card number presented 2011, whether there is a barcode or QR code presented 2012, whether there is a human face 2013, and/or the like.
In one implementation, if there is a payment card in the reality scene 2011, the TVC may determine a type of the card 2015 and the card number 2017. For example, the TVC may determine whether the card is a payment card (e.g., a credit card, a debit card, etc.), a membership card (e.g., a metro card, a store points card, a library card, etc.), a personal ID (e.g., a driver's license, etc.), an insurance card, and/or the like, based on the obtained textual content via OCR from the card. In one implementation, the TVC may query the user wallet for the card information 2018 to determine whether the card matches with any enrolled user account, and may generate and present overlay labels 2030 based on the type of the card (e.g., see overlay labels 927a-e for an identified Visa credit card 911 in
In another implementation, if there is a barcode and/or QR code detected within the reality scene 2012, the TVC may extract information from the barcode/QR code 2022, and determine a type of the object 2023, e.g., the barcode information may indicate whether the object comprises a purchase item, a bill, an invoice, and/or the like. In one implementation, the TVC may retrieve merchant information when the object comprises a purchase item, and/or biller information when the object comprises a bill 2028, and generate overlay labels accordingly, e.g., see overlay labels 1327a-e for an identified invoice 1326 in
In another implementation, if there is a human face detected from the reality scene 2013, the TVC may perform facial recognition to identify whether the presented human face matches with an existing contact 2024. In one implementation, the TVC may retrieve contact information if the contact is located from a contact list 2026, and/or add a new contact 2027 per user selection if the human face does not match with any existing contact record. The TVC may then generate and present overlay labels for the detected human face, e.g., see overlay labels 1008a-f for an identified face 1002 in
Upon user selection of the overlay labels, the TVC may proceed to transfer funds to an identified card, identified contact, and/or the like. The TVC may send financial transaction requests to an issuer network for processing, which may be performed in a similar manner as in
In one implementation, a consumer may engage in user interests indicative activities (e.g., web searches, wallet check-in, etc) 2031. For example, as shown in
In one implementation, when a consumer uses a mobile device to capture a reality scene (e.g., 2003/2004), TVC may determine a type of the object in the captured visual scene 2036, e.g., an item, card, barcode, receipt, etc. In one implementation, the TVC may retrieve stored user interest record 2038, and obtain information in the stored record. If the user interests record comprise a search term 2041, TVC may correlate the search term with product information 2044 (e.g., include price comparison information if the user is interested in finding the lowest price of a product, etc.), and generate an information layer for the virtual overlay 2049. In one implementation, the TVC may optionally capture mixed gestures within the captured reality scene 2029, e.g., consumer motion gestures, verbal gestures by articulating a command, etc. (see
In another implementation, if the user interests record comprise a real-time wallet check-in information 2042 of the consumer checking in at a retail store, the TVC may insert a retailer layer of virtual labels 2046 to the consumer device. In another implementation, the TVC may parse the user activity record for user interests indicators 2048 for other types of user activity data, e.g., browsing history, recent purchases, and/or the like, and determine an information layer of virtual overlay 2047. The consumer may obtain an automatically recommended injected layer of virtual label overlays 2050, and may switch to another layer of information labels by sliding on the layer, e.g., see 1611a-d in
As shown in
In one implementation, the TVC may determine whether the difference region has a “pointer” shape 2082, e.g., a fingertip, a pencil, etc. If not, e.g., the difference region may be noise caused by camera movement, etc., the TVC may determine whether the time lapse has exceeded a threshold. For example, if the TVC has been capturing the video scene for more than 10 seconds and detects no “pointer” shapes or “fingertip,” TVC may proceed to OCR/pattern recognition of the entire image 2087. Otherwise, the TVC may re-generate video frames at 2071.
In one implementation, if a “fingertip” or a “pointer” is detected at 2082, the TVC may determine a center point of the fingertip, e.g., by taking a middle point of the X and Y coordinates of the “fingertip.” The TVC may perform edge detection starting from the determined center point to determine the boundary of a consumer pointed object 2085. For example, the TVC may employ edge detection components such as, but not limited to Adobe Photoshop edge detection, Java edge detection package, and/or the like. Within implementations, upon TVC has defined boundaries of an object, the TVC may perform OCR and pattern recognition of the defined area 2088 to determine a type of the object.
Within implementations, TVC may receive information related to the determined object 2057 (e.g., 2018, 2027, 2028 in
For example, a data structure of a generated virtual label, substantially in the form of XML-formatted data, is provided below:
In the above example, the generated virtual label data structure includes fields such as size of the video frame, the captured object (e.g., the object is a barcode, etc.), information to be included in the virtual label, orientation of the label, format of the virtual label (e.g., template, font, background, transparency, etc.), injection position of the label , and/or the like. In one implementation, the virtual label may contain an informational link, e.g., for the product information in the above example, an Amazon link may be provided, etc. In one implementation, the injection position may be determined based on the position of the object (e.g., X, Y coordinates of the area on the image, determined by a barcode detector, etc.).
a-b show data flow diagrams illustrating processing gesture and vocal commands in some embodiments of the TVC. In some implementations, the user 2201 may initiate an action by providing both a physical gesture 2202 and a vocal command 2203 to an electronic device 2206. In some implementations, the user may use the electronic device itself in the gesture; in other implementations, the user may use another device (such as a payment device), and may capture the gesture via a camera on the electronic device 2207, or an external camera 2204 separate from the electronic device 2205. In some implementations, the camera may record a video of the device; in other implementations, the camera may take a burst of photos. In some implementations, the recording may begin when the user presses a button on the electronic device indicating that the user would like to initiate an action; in other implementations, the recording may begin as soon as the user enters a command application and begins to speak. The recording may end as soon as the user stops speaking, or as soon as the user presses a button to end the collection of video or image data. The electronic device may then send a command message 2208 to the TVC database, which may include the gesture and vocal command obtained from the user.
In some implementations, an exemplary XML-encoded command message 2208 may take a form similar to the following:
In some implementations, the electronic device may reduce the size of the vocal file by cropping the audio file to when the user begins and ends the vocal command. In some implementations, the TVC may process the gesture and audio data 2210 in order to determine the type of gesture performed, as well as the words spoken by the user. In some implementations, a composite gesture generated from the processing of the gesture and audio data may be embodied in an XML-encoded data structure similar to the following:
In some implementations, fields in the composite gesture data structure may be left blank depending on whether the particular gesture type (e.g., finger gesture, object gesture, and/or the like) has been made. The TVC may then match 2211 the gesture and the words to the various possible gesture types stored in the TVC database. In some implementations, the TVC may query the database for particular disparate gestures in a manner similar to the following:
In some implementations, the result of each query in the above example may be used to search for the composite gesture in the Multi-Disparate Gesture Action (MDGA) table of the database. For example, if $fingerresult is “tap check,” $objectresult is “swipe,” and $voiceresult is “pay total of check with this payment device,” TVC may search the MDGA table using these three results to narrow down the precise composite action that has been performed. If a match is found, the TVC may request confirmation that the right action was found, and then may perform the action 2212 using the user's account. In some implementations, the TVC may access the user's financial information and account 2213 in order to perform the action. In some implementations, TVC may update a gesture table 2214 in the TVC database 2215 to refine models for usable gestures based on the user's input, to add new gestures the user has invented, and/or the like. In some implementations, an update 2214 for a finger gesture may be performed via a PHP/MySQL command similar to the following:
After successfully updating the table 2216, the TVC may send the user to a confirmation page 2217 (or may provide an augmented reality (AR) overlay to the user) which may indicate that the action was successfully performed. In some implementations, the AR overlay may be provided to the user through use of smart glasses, contacts, and/or a like device (e.g. Google Glasses).
As shown in
The TVC may then perform the action specified 2221, accessing any information necessary to conduct the action 2222, and may send a confirmation page or AR overlay to the user 2223. In some implementations, the XML-encoded data structure for the AR overlay may take a form similar to the following:
a-23c show logic flow diagrams illustrating processing gesture and vocal commands in some embodiments of the TVC. In some implementations, the user 201 may perform a gesture and a vocal command 2301 equating to an action to be performed by TVC. The user's device 206 may capture the gesture 2302 via a set of images or a full video recorded by an on-board camera, or via an external camera-enabled device connected to the user's device, and may capture the vocal command via an on-board microphone, or via an external microphone connected to the user's device. The device may determine when both the gesture and the vocal command starts and ends 2303 based on when movement in the video or images starts and ends, based on when the user's voice starts and ends the vocal command, when the user presses a button in an action interface on the device, and/or the like. In some implementations, the user's device may then use the start and end points determined in order to package the gesture and voice data 2304, while keeping the packaged data a reasonable size. For example, in some implementations, the user's device may eliminate some accelerometer or gyroscope data, may eliminate images or crop the video of the gesture, based on the start and end points determined for the gesture. The user's device may also crop the audio file of the vocal command, based on the start and end points for the vocal command. This may be performed in order to reduce the size of the data and/or to better isolate the gesture or the vocal command. In some implementations, the user's device may package the data without reducing it based on start and end points.
In some implementations, TVC may receive 2305 the data from the user's device, which may include accelerometer and/or gyroscope data pertaining to the gesture, a video and/or images of the gesture, an audio file of the vocal command, and/or the like. In some implementations, TVC may determine what sort of data was sent by the user's device in order to determine how to process it. For example, if the user's device provides accelerometer and/or gyroscope data 2306, TVC may determine the gesture performed by matching the accelerometer and/or gyroscope data points with pre-determined mathematical gesture models 2309. For example, if a particular gesture would generate accelerometer and/or gyroscope data that would fit a linear gesture model, TVC will determine whether the received accelerometer and/or gyroscope data matches a linear model.
If the user's device provides a video and/or images of the gesture 2307, TVC may use an image processing component in order to process the video and/or images 2310 and determine what the gesture is. In some implementations, if a video is provided, the video may also be used to determine the vocal command provided by the user. As shown in
If the user's device provides an audio file 2308, then TVC may determine the vocal command given using an audio analytics component 2311. In some implementations, the audio analytics component may process the audio file and produce a text translation of the vocal command. As discussed above, in some implementations, the audio analytics component may also use a video, if provided, as input to produce a text translation of the user's vocal command.
As shown in
If the action is a single-party payment-related action 2318 (i.e., concerning one person and/or entity transferring funds to his/her/itself), TVC may retrieve the account information of the one user 2319, and may use it to access the relevant financial and/or other accounts associated in the transaction. For example, if one user is transferring funds from a bank account to a refillable gift card owned by the same user, then TVC would access the user's account in order to obtain information about both the bank account and the gift card, and would use the information to transfer funds from the bank account to the gift card 2320.
In either the multi-party or the single-party action, TVC may update 2321 the data of the affected accounts (including: saving a record of the transaction, which may include to whom the money was given to, the date and time of the transaction, the size of the transaction, and/or the like), and may send a confirmation of this update 2322 to the user.
If the action is related to obtaining information about a product and/or service 2323, TVC may send a request 2324 to the relevant merchant database(s) in order to get information about the product and/or service the user would like to know more about. TVC may provide any information obtained from the merchant to the user 2325. In some implementations, TVC may provide the information via an AR overlay, or via an information page or pop-up which displays all the retrieved information.
a shows a data flow diagram illustrating checking into a store or a venue in some embodiments of the TVC. In some implementations, the user 2401 may scan a QR code 2402 using their electronic device 2403 in order to check-in to a store. The electronic device may send check-in message 2404 to TVC server 2405, which may allow TVC to store information 2406 about the user based on their active e-wallet profile. In some implementations, an exemplary XML-encoded check-in message 2404 may take a form similar to the following:
In some implementations, the user, while shopping through the store, may also scan 2407 items with the user's electronic device, in order to obtain more information about them, in order to add them to the user's cart, and/or the like. In such implementations, the user's electronic device may send a scanned item message 2408 to the TVC server. In some implementations, an exemplary XML-encoded scanned item message 2408 may take a form similar to the following:
In some implementations, TVC may then determine the location 2409 of the user based on the location of the scanned item, and may send a notification 2410 to a sale's representative 2411 indicating that a user has checked into the store and is browsing items in the store. In some implementations, an exemplary XML-encoded notification message 2410 may comprise of the scanned item message of scanned item message 2408.
The sale's representative may use the information in the notification message to determine products and/or services to recommend 2412 to the user, based on the user's profile, location in the store, items scanned, and/or the like. Once the sale's representative has chosen at least one product and/or service to suggest, it may send the suggestion 2413 to the TVC server. In some implementations, an exemplary XML-encoded suggestion 2413 may take a form similar to the following:
Inc</item_manufacturer>
In some implementations, TVC may also use the user's profile information, location, scanned items, and/or the like to determine its own products and/or services to recommend 2414 to the user. In some implementations, TVC may determine where in the store any suggested product and/or service is 2415, based on aisle information in the item data structure, and may generate a map from the user's location to the location of the suggested product and/or service. In some implementations, the map overlays a colored path on a store map from the user's location to the suggested product and/or service. TVC may send 2416 this map, along with the suggested product and/or item, to the user, who may use it to find the suggested item, and add the suggested item to its shopping cart 2440 if the user would like to purchase it.
b-c show data flow diagrams illustrating accessing a virtual store in some embodiments of the TVC. In some implementations, a user 2417 may have a camera (either within an electronic device 2420 or an external camera 2419, such as an Xbox Kinect device) take a picture 2418 of the user. The user may also choose to provide various user attributes, such as the user's clothing size, the item(s) the user wishes to search for, and/or like information. The electronic device 2420 may also obtain 2421 stored attributes (such as a previously-submitted clothing size, color preference, and/or the like) from the TVC database, including whenever the user chooses not to provide attribute information. The electronic device may send a request 2422 to the TVC database 2423, and may receive all the stored attributes 2424 in the database. The electronic device may then send an apparel preview request 2425 to the TVC server 2426, which may include the photo of the user, the attributes provided, and/or the like. In some implementations, an exemplary XML-encoded apparel preview request 2425 may take a form similar to the following:
In some implementations, TVC may conduct its own analysis of the user based on the photo 2427, including analyzing the image to determine the user's body size, body shape, complexion, and/or the like. In some implementations, TVC may use these attributes, along with any provided through the apparel preview request, to search the database 2428 for clothing that matches the user's attributes and search criteria. In some implementations, TVC may also update 2429 the user's attributes stored in the database, based on the attributes provided in the apparel preview request or based on TVC' analysis of the user's photo. After TVC receives confirmation that the update is successful 2430, TVC may send a virtual closet 2431 to the user, comprising a user interface for previewing clothing, accessories, and/or the like chosen for the user based on the user's attributes and search criteria. In some implementations, the virtual closet may be implemented via HTML and Javascript.
In some implementations, as shown in
a shows a logic flow diagram illustrating checking into a store in some embodiments of the TVC. In some implementations, the user may scan a check-in code 2501, which may allow TVC to receive a notification 2502 that the user has checked in, and may allow TVC to use the user profile identification information provided to create a store profile for the user. In some implementations, the user may scan a product 2503, which may cause TVC to receive notification of the user's item scan 2504, and may prompt TVC to determine where the user is based on the location of the scanned item 2505. In some implementations, TVC may then send a notification of the check-in and/or the item scan to a sale's representative 2506. TVC may then determine (or may receive from the sale's representative) at least one product and/or service to recommend to the user 2507, based on the user's profile, shopping cart, scanned item, and/or the like. TVC may then determine the location of the recommended product and/or service 2508, and may use the user's location and the location of the recommended product and/or service to generate a map from the user's location to the recommended product and/or service 2509. TVC may then send the recommmended product and/or service, along with the generated map, to the user 2510, so that the user may find its way to the recommended product and add it to a shopping cart if desired.
b shows a logic flow diagram illustrating accessing a virtual store in some embodiments of the TVC. In some implementations, the user's device may take a picture 2511 of the user, and may request from the user attribute data 2512, such as clothing size, clothing type, and/or like information. If the user chooses not to provide information 2513, the electronic device may access the user profile in the TVC database in order to see if any previously-entered user attribute data exists 2514. In some implementations, anything found is sent with the user image to TVC 2515. If little to no user attribute information is provided, TVC may use an image processing component to predict the user's clothing size, complexion, body type, and/or the like 2516, and may retrieve clothing from the database 2517. In some implementations, if the user chose to provide information 2513, then TVC automatically searches the database 2517 for clothing without attempting to predict the user's clothing size and/or the like. In some implementations, TVC may use the user attributes and search criteria to search the retrieved clothing 2518 for any clothing tagged with attributes matching that of the user (e.g. clothing tagged with a similar size as the user, and/or the like). TVC may send the matching clothing to the user 2519 as recommended items to preview via a virtual closet interface. Depending upon further search parameters provided by the user (e.g., new colors, higher or lower prices, and/or the like), TVC may update the clothing loaded into the virtual closet 2520 based on the further search parameters (e.g., may only load red clothing if the user chooses to only see the red clothing in the virtual closet, and/or the like).
In some implementations, the user may provide a selection of at least one article of clothing to try on 2521, prompting TVC to determine body and/or joint locations and markers in the user photo 2522, and to scale the image of the article of clothing to match the user image 2523, based on those body and/or joint locations and markers. In some implementations, TVC may also format the clothing image 2524, including altering shadows in the image, blurring the image, and/or the like, in order to match the look of the clothing image to the look of the user image. TVC may superimpose 2525 the clothing image on the user image to allow the user to virtually preview the article of clothing on the user, and may allow the user to change options such as the clothing color, size, and/or the like while the article of clothing is being previewed on the user. In some implementations, TVC may receive a request to purchase at least one article of clothing 2526, and may retrieve user information 2527, including the user's ID, shipping address, and/or the like. TVC may further retrieve the user's payment information 2528, including the user's preferred payment device or account, and/or the like, and may contact the user's issuer (and that of the merchant) 2529 in order to process the transaction. TVC may send a confirmation to the user when the transaction is completed 2530.
a-d show schematic diagrams illustrating initiating transactions in some embodiments of the TVC. In some implementations, as shown in
As shown in
As shown in
In one embodiment, for example, a user may select the option current items 3215, as shown in the left most user interface of
With reference to
With reference to
In one implementation, a user may select Joe P. for payment. Joe P., as shown in the user interface, has an email icon 3217g next to his name indicating that Joe P. accepts payment via email. When his name is selected, the user interface may display his contact information such as email, phone, etc. If a user wishes to make a payment to Joe P. by a method other than email, the user may add another transfer mode 3217j to his contact information and make a payment transfer. With reference to
With reference to
With reference to
With reference to
In one implementation, the user may combine funds from multiple sources to pay for the transaction. The amount 3315 displayed on the user interface may provide an indication of the amount of total funds covered so far by the selected forms of payment (e.g., Discover card and rewards points). The user may choose another form of payment or adjust the amount to be debited from one or more forms of payment until the amount 3315 matches the amount payable 3314. Once the amounts to be debited from one or more forms of payment are finalized by the user, payment authorization may begin.
In one implementation, the user may select a secure authorization of the transaction by selecting the cloak button 3322 to effectively cloak or anonymize some (e.g., pre-configured) or all identifying information such that when the user selects pay button 3321, the transaction authorization is conducted in a secure and anonymous manner. In another implementation, the user may select the pay button 3321 which may use standard authorization techniques for transaction processing. In yet another implementation, when the user selects the social button 3323, a message regarding the transaction may be communicated to one of more social networks (set up by the user) which may post or announce the purchase transaction in a social forum such as a wall post or a tweet. In one implementation, the user may select a social payment processing option 3323. The indicator 3324 may show the authorizing and sending social share data in progress.
In another implementation, a restricted payment mode 3325 may be activated for certain purchase activities such as prescription purchases. The mode may be activated in accordance with rules defined by issuers, insurers, merchants, payment processor and/or other entities to facilitate processing of specialized goods and services. In this mode, the user may scroll down the list of forms of payments 3326 under the funds tab to select specialized accounts such as a flexible spending account (FSA) 3327, health savings account (HAS), and/or the like and amounts to be debited to the selected accounts. In one implementation, such restricted payment mode 1925 processing may disable social sharing of purchase information.
In one embodiment, the wallet mobile application may facilitate importing of funds via the import funds user interface 3328. For example, a user who is unemployed may obtain unemployment benefit fund 3329 via the wallet mobile application. In one implementation, the entity providing the funds may also configure rules for using the fund as shown by the processing indicator message 3330. The wallet may read and apply the rules prior, and may reject any purchases with the unemployment funds that fail to meet the criteria set by the rules. Example criteria may include, for example, merchant category code (MCC), time of transaction, location of transaction, and/or the like. As an example, a transaction with a grocery merchant having MCC 5411 may be approved, while a transaction with a bar merchant having an MCC 5813 may be refused.
With reference to
Similarly, when a German user operates a wallet in Germany, the mobile wallet application user interface may be dynamically updated to reflect the country of operation 3332 and the currency 3334. In a further implementation, the wallet application may rearrange the order in which different forms of payment 3336 are listed based on their acceptance level in that country. Of course, the order of these forms of payments may be modified by the user to suit his or her own preferences.
With reference to
With reference to
With reference to
With reference to
In one implementation, the user may select a transaction, for example transaction 3415, to view the details of the transaction. For example, the user may view the details of the items associated with the transaction and the amounts 3416 of each item. In a further implementation, the user may select the show option 3417 to view actions 3418 that the user may take in regards to the transaction or the items in the transaction. For example, the user may add a photo to the transaction (e.g., a picture of the user and the iPad the user bought). In a further implementation, if the user previously shared the purchase via social channels, a post including the photo may be generated and sent to the social channels for publishing. In one implementation, any sharing may be optional, and the user, who did not share the purchase via social channels, may still share the photo through one or more social channels of his or her choice directly from the history mode of the wallet application. In another implementation, the user may add the transaction to a group such as company expense, home expense, travel expense or other categories set up by the user. Such grouping may facilitate year-end accounting of expenses, submission of work expense reports, submission for value added tax (VAT) refunds, personal expenses, and/or the like. In yet another implementation, the user may buy one or more items purchased in the transaction. The user may then execute a transaction without going to the merchant catalog or site to find the items. In a further implementation, the user may also cart one or more items in the transaction for later purchase.
The history mode, in another embodiment, may offer facilities for obtaining and displaying ratings 3419 of the items in the transaction. The source of the ratings may be the user, the user's friends (e.g., from social channels, contacts, etc.), reviews aggregated from the web, and/or the like. The user interface in some implementations may also allow the user to post messages to other users of social channels (e.g., TWITTER or FACEBOOK). For example, the display area 3420 shows FACEBOOK message exchanges between two users. In one implementation, a user may share a link via a message 3421. Selection of such a message having embedded link to a product may allow the user to view a description of the product and/or purchase the product directly from the history mode.
In one embodiment, the history mode may also include facilities for exporting receipts. The export receipts pop up 3422 may provide a number of options for exporting the receipts of transactions in the history. For example, a user may use one or more of the options 3425, which include save (to local mobile memory, to server, to a cloud account, and/or the like), print to a printer, fax, email, and/or the like. The user may utilize his or her address book 3423 to look up email or fax number for exporting. The user may also specify format options 3424 for exporting receipts. Example format options may include, without limitation, text files (.doc, .txt, .rtf, iif, etc.), spreadsheet (.csv, .xls, etc.), image files (.jpg, .tff, .png, etc.), portable document format (.pdf), postscript (.ps), and/or the like. The user may then click or tap the export button 3427 to initiate export of receipts.
With reference to
As shown, the user may enter a search term (e.g., bills) in the search bar 2121. The user may then identify in the tab 3522 the receipt 3523 the user wants to reallocate. Alternatively, the user may directly snap a picture of a barcode on a receipt, and the snap mode may generate and display a receipt 3523 using information from the barcode. The user may now reallocate 3525. In some implementations, the user may also dispute the transaction 3524 or archive the receipt 3526.
In one implementation, when the reallocate button 3525 is selected, the wallet application may perform optical character recognition (OCR) of the receipt. Each of the items in the receipt may then be examined to identify one or more items which could be charged to which payment device or account for tax or other benefits such as cash back, reward points, etc. In this example, there is a tax benefit if the prescription medication charged to the user's Visa card is charged to the user's FSA. The wallet application may then perform the reallocation as the back end. The reallocation process may include the wallet contacting the payment processor to credit the amount of the prescription medication to the Visa card and debit the same amount to the user's FSA account. In an alternate implementation, the payment processor (e.g., Visa or MasterCard) may obtain and OCR the receipt, identify items and payment accounts for reallocation and perform the reallocation. In one implementation, the wallet application may request the user to confirm reallocation of charges for the selected items to another payment account. The receipt 3527 may be generated after the completion of the reallocation process. As discussed, the receipt shows that some charges have been moved from the Visa account to the FSA.
With reference to
In one implementation, the user may decide to pay with default 3534. The wallet application may then use the user's default method of payment, in this example the wallet, to complete the purchase transaction. Upon completion of the transaction, a receipt may be automatically generated for proof of purchase. The user interface may also be updated to provide other options for handling a completed transaction. Example options include social 3537 to share purchase information with others, reallocate 3538 as discussed with regard to
With reference to
In one implementation, after the offer or coupon 3546 is applied, the user may have the option to find qualifying merchants and/or products using find, the user may go to the wallet using 3548, and the user may also save the offer or coupon 3546 for later use.
With reference to
For example, a user may go to doctor's office and desire to pay the co-pay for doctor's appointment. In addition to basic transactional information such as account number and name, the app may provide the user the ability to select to transfer medical records, health information, which may be provided to the medical provider, insurance company, as well as the transaction processor to reconcile payments between the parties. In some implementations, the records may be sent in a Health Insurance Portability and Accountability Act (HIPAA)-compliant data format and encrypted, and only the recipients who are authorized to view such records may have appropriate decryption keys to decrypt and view the private user information.
With reference to
In some implementations, the TVC may utilize a text challenge procedure to verify the authenticity of the user, e.g., 3725. For example, the TVC may communicate with the user via text chat, SMS messages, electronic mail, Facebook® messages, Twitter™ tweets, and/or the like. The TVC may pose a challenge question, e.g., 3726, for the user. The app may provide a user input interface element(s) (e.g., virtual keyboard 3728) to answer the challenge question posed by the TVC. In some implementations, the challenge question may be randomly selected by the TVC automatically; in some implementations, a customer service representative may manually communicate with the user. In some implementations, the user may not have initiated the transaction, e.g., the transaction is fraudulent. In such implementations, the user may cancel the text challenge. The TVC may cancel the transaction, and/or initiate fraud investigation on behalf of the user.
In some embodiments, the merchant server may obtain the checkout request from the client, and extract the checkout detail (e.g., XML data) from the checkout request. For example, the merchant server may utilize a parser such as the example parsers described below in the discussion with reference to
In some embodiments, in response to obtaining the product data, the merchant server may generate, e.g., 3816, checkout data to provide for the PoS client. In some embodiments, such checkout data, e.g., 3817, may be embodied, in part, in a HyperText Markup Language (“HTML”) page including data for display, such as product detail, product pricing, total pricing, tax information, shipping information, offers, discounts, rewards, value-added service information, etc., and input fields to provide payment information to process the purchase transaction, such as account holder name, account number, billing address, shipping address, tip amount, etc. In some embodiments, the checkout data may be embodied, in part, in a Quick Response (“QR”) code image that the PoS client can display, so that the user may capture the QR code using a user's device to obtain merchant and/or product data for generating a purchase transaction processing request. In some embodiments, a user alert mechanism may be built into the checkout data. For example, the merchant server may embed a URL specific to the transaction into the checkout data. In some embodiments, the alerts URL may further be embedded into optional level 3 data in card authorization requests, such as those discussed further below with reference to
Upon obtaining the checkout data, e.g., 3817, the PoS client may render and display, e.g., 3818, the checkout data for the user.
In some embodiments, upon authenticating the user for access to virtual wallet features, the user wallet device may provide a transaction authorization input, e.g., 4014, to a point-of-sale (“PoS”) client, e.g., 4002. For example, the user wallet device may communicate with the PoS client via Bluetooth, Wi-Fi, cellular communication, one- or two-way near-field communication (“NFC”), and/or the like. In embodiments where the user utilizes a plastic card instead of the user wallet device, the user may swipe the plastic card at the PoS client to transfer information from the plastic card into the PoS client. For example, the PoS client may obtain, as transaction authorization input 4014, track 1 data from the user's plastic card (e.g., credit card, debit card, prepaid card, charge card, etc.), such as the example track 1 data provided below:
In embodiments where the user utilizes a user wallet device, the user wallet device may provide payment information to the PoS client, formatted according to a data formatting protocol appropriate to the communication mechanism employed in the communication between the user wallet device and the PoS client. An example listing of transaction authorization input 4014, substantially in the form of XML-formatted data, is provided below:
In some embodiments, the PoS client may generate a card authorization request, e.g., 4015, using the obtained transaction authorization input from the user wallet device, and/or product/checkout data (see, e.g.,
In some embodiments, the card authorization request generated by the user device may include a minimum of information required to process the purchase transaction. For example, this may improve the efficiency of communicating the purchase transaction request, and may also advantageously improve the privacy protections provided to the user and/or merchant. For example, in some embodiments, the card authorization request may include at least a session ID for the user's shopping session with the merchant. The session ID may be utilized by any component and/or entity having the appropriate access authority to access a secure site on the merchant server to obtain alerts, reminders, and/or other data about the transaction(s) within that shopping session between the user and the merchant. In some embodiments, the PoS client may provide the generated card authorization request to the merchant server, e.g., 4016. The merchant server may forward the card authorization request to a pay gateway server, e.g., 4004a, for routing the card authorization request to the appropriate payment network for payment processing. For example, the pay gateway server may be able to select from payment networks, such as Visa, Mastercard, American Express, Paypal, etc., to process various types of transactions including, but not limited to: credit card, debit card, prepaid card, B2B and/or like transactions. In some embodiments, the merchant server may query a database, e.g., merchant/acquirer database 4003b, for a network address of the payment gateway server, for example by using a portion of a user payment card number, or a user ID (such as an email address) as a keyword for the database query. For example, the merchant server may issue PHP/SQL commands to query a database table (such as
In response, the merchant/acquirer database may provide the requested payment gateway address, e.g., 4018. The merchant server may forward the card authorization request to the pay gateway server using the provided address, e.g., 4019. In some embodiments, upon receiving the card authorization request from the merchant server, the pay gateway server may invoke a component to provide one or more services associated with purchase transaction authorization. For example, the pay gateway server may invoke components for fraud prevention, loyalty and/or rewards, and/or other services for which the user-merchant combination is authorized. The pay gateway server may forward the card authorization request to a pay network server, e.g., 4005a, for payment processing. For example, the pay gateway server may be able to select from payment networks, such as Visa, Mastercard, American Express, Paypal, etc., to process various types of transactions including, but not limited to: credit card, debit card, prepaid card, B2B and/or like transactions. In some embodiments, the pay gateway server may query a database, e.g., pay gateway database 4004b, for a network address of the payment network server, for example by using a portion of a user payment card number, or a user ID (such as an email address) as a keyword for the database query. For example, the pay gateway server may issue PHP/SQL commands to query a database table (such as
In response, the payment gateway database may provide the requested payment network address, e.g., 4022. The pay gateway server may forward the card authorization request to the pay network server using the provided address, e.g., 4023.
With reference to
In some embodiments, the pay network server may generate a query, e.g., 4024, for issuer server(s) corresponding to the user-selected payment options. For example, the user's account may be linked to one or more issuer financial institutions (“issuers”), such as banking institutions, which issued the account(s) for the user. For example, such accounts may include, but not be limited to: credit card, debit card, prepaid card, checking, savings, money market, certificates of deposit, stored (cash) value accounts and/or the like. Issuer server(s), e.g., 4006a, of the issuer(s) may maintain details of the user's account(s). In some embodiments, a database, e.g., pay network database 4005b, may store details of the issuer server(s) associated with the issuer(s). In some embodiments, the pay network server may query a database, e.g., pay network database 4005b, for a network address of the issuer(s) server(s), for example by using a portion of a user payment card number, or a user ID (such as an email address) as a keyword for the database query. For example, the merchant server may issue PHP/SQL commands to query a database table (such as
In response to obtaining the issuer server query, e.g., 4024, the pay network database may provide, e.g., 4025, the requested issuer server data to the pay network server. In some embodiments, the pay network server may utilize the issuer server data to generate funds authorization request(s), e.g., 4026, for each of the issuer server(s) selected based on the pre-defined payment settings associated with the user's virtual wallet, and/or the user's payment options input, and provide the funds authorization request(s) to the issuer server(s). In some embodiments, the funds authorization request(s) may include details such as, but not limited to: the costs to the user involved in the transaction, card account details of the user, user billing and/or shipping information, and/or the like. An example listing of a funds authorization request 4026, substantially in the form of a HTTP(S) POST message including XML-formatted data, is provided below:
In some embodiments, an issuer server may parse the authorization request(s), and based on the request details may query a database, e.g., user profile database 4006b, for data associated with an account linked to the user. For example, the merchant server may issue PHP/SQL commands to query a database table (such as
In some embodiments, on obtaining the user account(s) data, e.g., 4028, the issuer server may determine whether the user can pay for the transaction using funds available in the account, 4029. For example, the issuer server may determine whether the user has a sufficient balance remaining in the account, sufficient credit associated with the account, and/or the like. Based on the determination, the issuer server(s) may provide a funds authorization response, e.g., 4030, to the pay network server. For example, the issuer server(s) may provide a HTTP(S) POST message similar to the examples above. In some embodiments, if at least one issuer server determines that the user cannot pay for the transaction using the funds available in the account, the pay network server may request payment options again from the user (e.g., by providing an authorization fail message to the user device and requesting the user device to provide new payment options), and re-attempt authorization for the purchase transaction. In some embodiments, if the number of failed authorization attempts exceeds a threshold, the pay network server may abort the authorization process, and provide an “authorization fail” message to the merchant server, user device and/or client.
In some embodiments, the pay network server may obtain the funds authorization response including a notification of successful authorization, and parse the message to extract authorization details. Upon determining that the user possesses sufficient funds for the transaction, e.g., 4031, the pay network server may invoke a component to provide value-add services for the user.
In some embodiments, the pay network server may generate a transaction data record from the authorization request and/or authorization response, and store the details of the transaction and authorization relating to the transaction in a transactions database. For example, the pay network server may issue PHP/SQL commands to store the data to a database table (such as
In some embodiments, the pay network server may forward a transaction authorization response, e.g., 4032, to the user wallet device, PoS client, and/or merchant server. The merchant may obtain the transaction authorization response, and determine from it that the user possesses sufficient funds in the card account to conduct the transaction. The merchant server may add a record of the transaction for the user to a batch of transaction data relating to authorized transactions. For example, the merchant may append the XML data pertaining to the user transaction to an XML data file comprising XML data for transactions that have been authorized for various users, e.g., 4033, and store the XML data file, e.g., 4034, in a database, e.g., merchant database 404. For example, a batch XML data file may be structured similar to the example XML data structure template provided below:
In some embodiments, the server may also generate a purchase receipt, e.g., 4033, and provide the purchase receipt to the client, e.g., 4035. The client may render and display, e.g., 4036, the purchase receipt for the user. In some embodiments, the user's wallet device may also provide a notification of successful authorization to the user. For example, the PoS client/user device may render a webpage, electronic message, text/SMS message, buffer a voicemail, emit a ring tone, and/or play an audio message, etc., and provide output including, but not limited to: sounds, music, audio, video, images, tactile feedback, vibration alerts (e.g., on vibration-capable client devices such as a smartphone etc.), and/or the like.
In some embodiments, upon authenticating the user for access to virtual wallet features, the user wallet device may provide a transaction authorization input, e.g., 4104, to a point-of-sale (“PoS”) client. For example, the user wallet device may communicate with the PoS client via Bluetooth, Wi-Fi, cellular communication, one- or two-way near-field communication (“NFC”), and/or the like. In embodiments where the user utilizes a plastic card instead of the user wallet device, the user may swipe the plastic card at the PoS client to transfer information from the plastic card into the PoS client. In embodiments where the user utilizes a user wallet device, the user wallet device may provide payment information to the PoS client, formatted according to a data formatting protocol appropriate to the communication mechanism employed in the communication between the user wallet device and the PoS client.
In some embodiments, the PoS client may obtain the transaction authorization input, and parse the input to extract payment information from the transaction authorization input, e.g., 4105. For example, the PoS client may utilize a parser, such as the example parsers provided below in the discussion with reference to
In some embodiments, the PoS client may provide the generated card authorization request to the merchant server. The merchant server may forward the card authorization request to a pay gateway server, for routing the card authorization request to the appropriate payment network for payment processing. For example, the pay gateway server may be able to select from payment networks, such as Visa, Mastercard, American Express, Paypal, etc., to process various types of transactions including, but not limited to: credit card, debit card, prepaid card, B2B and/or like transactions. In some embodiments, the merchant server may query a database, e.g., 4108, for a network address of the payment gateway server, for example by using a portion of a user payment card number, or a user ID (such as an email address) as a keyword for the database query. In response, the merchant/acquirer database may provide the requested payment gateway address, e.g., 4110. The merchant server may forward the card authorization request to the pay gateway server using the provided address. In some embodiments, upon receiving the card authorization request from the merchant server, the pay gateway server may invoke a component to provide one or more service associated with purchase transaction authorization, e.g., 4111. For example, the pay gateway server may invoke components for fraud prevention (see e.g., VerifyChat,
The pay gateway server may forward the card authorization request to a pay network server for payment processing, e.g., 4114. For example, the pay gateway server may be able to select from payment networks, such as Visa, Mastercard, American Express, Paypal, etc., to process various types of transactions including, but not limited to: credit card, debit card, prepaid card, B2B and/or like transactions. In some embodiments, the pay gateway server may query a database, e.g., 4112, for a network address of the payment network server, for example by using a portion of a user payment card number, or a user ID (such as an email address) as a keyword for the database query. In response, the payment gateway database may provide the requested payment network address, e.g., 4113. The pay gateway server may forward the card authorization request to the pay network server using the provided address, e.g., 4114.
With reference to
In response to obtaining the issuer server query, the pay network database may provide, e.g., 4116, the requested issuer server data to the pay network server. In some embodiments, the pay network server may utilize the issuer server data to generate funds authorization request(s), e.g., 4117, for each of the issuer server(s) selected based on the pre-defined payment settings associated with the user's virtual wallet, and/or the user's payment options input, and provide the funds authorization request(s) to the issuer server(s). In some embodiments, the funds authorization request(s) may include details such as, but not limited to: the costs to the user involved in the transaction, card account details of the user, user billing and/or shipping information, and/or the like. In some embodiments, an issuer server may parse the authorization request(s), e.g., 4118, and based on the request details may query a database, e.g., 4119, for data associated with an account linked to the user.
In some embodiments, on obtaining the user account(s) data, e.g., 4120, the issuer server may determine whether the user can pay for the transaction using funds available in the account, e.g., 4121. For example, the issuer server may determine whether the user has a sufficient balance remaining in the account, sufficient credit associated with the account, and/or the like. Based on the determination, the issuer server(s) may provide a funds authorization response, e.g., 4122, to the pay network server. In some embodiments, if at least one issuer server determines that the user cannot pay for the transaction using the funds available in the account, the pay network server may request payment options again from the user (e.g., by providing an authorization fail message to the user device and requesting the user device to provide new payment options), and re-attempt authorization for the purchase transaction. In some embodiments, if the number of failed authorization attempts exceeds a threshold, the pay network server may abort the authorization process, and provide an “authorization fail” message to the merchant server, user device and/or client.
In some embodiments, the pay network server may obtain the funds authorization response including a notification of successful authorization, and parse the message to extract authorization details. Upon determining that the user possesses sufficient funds for the transaction, e.g., 4123, the pay network server may invoke a component to provide value-add services for the user, e.g., 4123.
In some embodiments, the pay network server may forward a transaction authorization response to the user wallet device, PoS client, and/or merchant server. The merchant may parse, e.g., 4124, the transaction authorization response, and determine from it that the user possesses sufficient funds in the card account to conduct the transaction, e.g., 4125, option“Yes.” The merchant server may add a record of the transaction for the user to a batch of transaction data relating to authorized transactions. For example, the merchant may append the XML data pertaining to the user transaction to an XML data file comprising XML data for transactions that have been authorized for various users, e.g., 4126, and store the XML data file, e.g., 4127, in a database. In some embodiments, the server may also generate a purchase receipt, e.g., 4128, and provide the purchase receipt to the client. The client may render and display, e.g., 4129, the purchase receipt for the user. In some embodiments, the user's wallet device may also provide a notification of successful authorization to the user. For example, the PoS client/user device may render a webpage, electronic message, text /SMS message, buffer a voicemail, emit a ring tone, and/or play an audio message, etc., and provide output including, but not limited to: sounds, music, audio, video, images, tactile feedback, vibration alerts (e.g., on vibration-capable client devices such as a smartphone etc.), and/or the like.
With reference to
In some embodiments, the issuer server may generate a payment command, e.g., 4227. For example, the issuer server may issue a command to deduct funds from the user's account (or add a charge to the user's credit card account). The issuer server may issue a payment command, e.g., 4227, to a database storing the user's account information, e.g., user profile database 4206b. The issuer server may provide an individual payment confirmation, e.g., 4228, to the pay network server, which may forward, e.g., 4229, the funds transfer message to the acquirer server. An example listing of an individual payment confirmation 4228, substantially in the form of a HTTP(S) POST message including XML-formatted data, is provided below:
In some embodiments, the acquirer server may parse the individual payment confirmation, and correlate the transaction (e.g., using the request ID field in the example above) to the merchant. The acquirer server may then transfer the funds specified in the funds transfer message to an account of the merchant. For example, the acquirer server may query, e.g. 4230, an acquirer database 4207b for payment ledger and/or merchant account data, e.g., 4231. The acquirer server may utilize payment ledger and/or merchant account data from the acquirer database, along with the individual payment confirmation, to generate updated payment ledger and/or merchant account data, e.g., 4232. The acquirer server may then store, e.g., 4233, the updated payment ledger and/or merchant account data to the acquire database.
The pay network server may parse the batch payment request obtained from the acquirer server, and extract the transaction data for each transaction stored in the batch payment request, e.g., 4308. The pay network server may store the transaction data, e.g., 4309, for each transaction in a pay network database. In some embodiments, the pay network server may invoke a component, e.g., 4310, to provide analytics based on the transactions of the merchant for whom purchase transaction are being cleared.
With reference to
In some embodiments, the acquirer server may parse the individual payment confirmation, and correlate the transaction (e.g., using the request ID field in the example above) to the merchant. The acquirer server may then transfer the funds specified in the funds transfer message to an account of the merchant. For example, the acquirer server may query, e.g. 4319, an acquirer database for payment ledger and/or merchant account data, e.g., 4320. The acquirer server may utilize payment ledger and/or merchant account data from the acquirer database, along with the individual payment confirmation, to generate updated payment ledger and/or merchant account data, e.g., 4321. The acquirer server may then store, e.g., 4322, the updated payment ledger and/or merchant account data to the acquire database.
Typically, users, e.g., 4433a, which may be people and/or other systems, may engage information technology systems (e.g., computers) to facilitate information processing. In turn, computers employ processors to process information; such processors 4403 may be referred to as central processing units (CPU). One form of processor is referred to as a microprocessor. CPUs use communicative circuits to pass binary encoded signals acting as instructions to enable various operations. These instructions may be operational and/or data instructions containing and/or referencing other instructions and data in various processor accessible and operable areas of memory 4429 (e.g., registers, cache memory, random access memory, etc.). Such communicative instructions may be stored and/or transmitted in batches (e.g., batches of instructions) as programs and/or data components to facilitate desired operations. These stored instruction codes, e.g., programs, may engage the CPU circuit components and other motherboard and/or system components to perform desired operations. One type of program is a computer operating system, which, may be executed by CPU on a computer; the operating system enables and facilitates users to access and operate computer information technology and resources. Some resources that may be employed in information technology systems include: input and output mechanisms through which data may pass into and out of a computer; memory storage into which data may be saved; and processors by which information may be processed. These information technology systems may be used to collect data for later retrieval, analysis, and manipulation, which may be facilitated through a database program. These information technology systems provide interfaces that allow users to access and operate various system components.
In one embodiment, the TVC controller 4401 may be connected to and/or communicate with entities such as, but not limited to: one or more users from user input devices 4411; peripheral devices 4412; an optional cryptographic processor device 4428; and/or a communications network 4413. For example, the TVC controller 4401 may be connected to and/or communicate with users, e.g., 4433a, operating client device(s), e.g., 4433b, including, but not limited to, personal computer(s), server(s) and/or various mobile device(s) including, but not limited to, cellular telephone(s), smartphone(s) (e.g., iPhone®, Blackberry®, Android OS-based phones etc.), tablet computer(s) (e.g., Apple iPad™, HP Slate™, Motorola Xoom™, etc.), eBook reader(s) (e.g., Amazon Kindle™, Barnes and Noble's Nook™ eReader, etc.), laptop computer(s), notebook(s), netbook(s), gaming console(s) (e.g., XBOX Live™, Nintendo® DS, Sony PlayStation® Portable, etc.), portable scanner(s), and/or the like.
Networks are commonly thought to comprise the interconnection and interoperation of clients, servers, and intermediary nodes in a graph topology. It should be noted that the term “server” as used throughout this application refers generally to a computer, other device, program, or combination thereof that processes and responds to the requests of remote users across a communications network. Servers serve their information to requesting “clients.” The term “client” as used herein refers generally to a computer, program, other device, user and/or combination thereof that is capable of processing and making requests and obtaining and processing any responses from servers across a communications network. A computer, other device, program, or combination thereof that facilitates, processes information and requests, and/or furthers the passage of information from a source user to a destination user is commonly referred to as a “node.” Networks are generally thought to facilitate the transfer of information from source points to destinations. A node specifically tasked with furthering the passage of information from a source to a destination is commonly called a “router.” There are many forms of networks such as Local Area Networks (LANs), Pico networks, Wide Area Networks (WANs), Wireless Networks (WLANs), etc. For example, the Internet is generally accepted as being an interconnection of a multitude of networks whereby remote clients and servers may access and interoperate with one another.
The TVC controller 4401 may be based on computer systems that may comprise, but are not limited to, components such as: a computer systemization 4402 connected to memory 4429.
A computer systemization 4402 may comprise a clock 4430, central processing unit (“CPU(s)” and/or “processor(s)” (these terms are used interchangeable throughout the disclosure unless noted to the contrary)) 4403, a memory 4429 (e.g., a read only memory (ROM) 4406, a random access memory (RAM) 4405, etc.), and/or an interface bus 4407, and most frequently, although not necessarily, are all interconnected and/or communicating through a system bus 4404 on one or more (mother)board(s) 4402 having conductive and/or otherwise transportive circuit pathways through which instructions (e.g., binary encoded signals) may travel to effectuate communications, operations, storage, etc. The computer systemization may be connected to a power source 4486; e.g., optionally the power source may be internal. Optionally, a cryptographic processor 4426 and/or transceivers (e.g., ICs) 4474 may be connected to the system bus. In another embodiment, the cryptographic processor and/or transceivers may be connected as either internal and/or external peripheral devices 4412 via the interface bus I/O. In turn, the transceivers may be connected to antenna(s) 4475, thereby effectuating wireless transmission and reception of various communication and/or sensor protocols; for example the antenna(s) may connect to: a Texas Instruments WiLink WL1283 transceiver chip (e.g., providing 802.1 in, Bluetooth 3.0, FM, global positioning system (GPS) (thereby allowing TVC controller to determine its location)); Broadcom BCM4329FKUBG transceiver chip (e.g., providing 802.11n, Bluetooth 2.1+EDR, FM, etc.); a Broadcom BCM4750IUB8 receiver chip (e.g., GPS); an Infineon Technologies X-Gold 618-PMB9800 (e.g., providing 2G/3G HSDPA/HSUPA communications); and/or the like. The system clock typically has a crystal oscillator and generates a base signal through the computer systemization's circuit pathways. The clock is typically coupled to the system bus and various clock multipliers that will increase or decrease the base operating frequency for other components interconnected in the computer systemization. The clock and various components in a computer systemization drive signals embodying information throughout the system. Such transmission and reception of instructions embodying information throughout a computer systemization may be commonly referred to as communications. These communicative instructions may further be transmitted, received, and the cause of return and/or reply communications beyond the instant computer systemization to: communications networks, input devices, other computer systemizations, peripheral devices, and/or the like. It should be understood that in alternative embodiments, any of the above components may be connected directly to one another, connected to the CPU, and/or organized in numerous variations employed as exemplified by various computer systems.
The CPU comprises at least one high-speed data processor adequate to execute program components for executing user and/or system-generated requests. Often, the processors themselves will incorporate various specialized processing units, such as, but not limited to: integrated system (bus) controllers, memory management control units, floating point units, and even specialized processing sub-units like graphics processing units, digital signal processing units, and/or the like. Additionally, processors may include internal fast access addressable memory, and be capable of mapping and addressing memory 4429 beyond the processor itself; internal memory may include, but is not limited to: fast registers, various levels of cache memory (e.g., level 1, 2, 3, etc.), RAM, etc. The processor may access this memory through the use of a memory address space that is accessible via instruction address, which the processor can construct and decode allowing it to access a circuit path to a specific memory address space having a memory state. The CPU may be a microprocessor such as: AMD's Athlon, Duron and/or Opteron; ARM's application, embedded and secure processors; IBM and/or Motorola's DragonBall and PowerPC; IBM's and Sony's Cell processor; Intel's Celeron, Core (2) Duo, Itanium, Pentium, Xeon, and/or XScale; and/or the like processor(s). The CPU interacts with memory through instruction passing through conductive and/or transportive conduits (e.g., (printed) electronic and/or optic circuits) to execute stored instructions (i.e., program code) according to conventional data processing techniques. Such instruction passing facilitates communication within the TVC controller and beyond through various interfaces. Should processing requirements dictate a greater amount speed and/or capacity, distributed processors (e.g., Distributed TVC), mainframe, multi-core, parallel, and/or super-computer architectures may similarly be employed. Alternatively, should deployment requirements dictate greater portability, smaller Personal Digital Assistants (PDAs) may be employed.
Depending on the particular implementation, features of the TVC may be achieved by implementing a microcontroller such as CAST's R8051XC2 microcontroller; Intel's MCS 51 (i.e., 8051 microcontroller); and/or the like. Also, to implement certain features of the TVC, some feature implementations may rely on embedded components, such as: Application-Specific Integrated Circuit (“ASIC”), Digital Signal Processing (“DSP”), Field Programmable Gate Array (“FPGA”), and/or the like embedded technology. For example, any of the TVC component collection (distributed or otherwise) and/or features may be implemented via the microprocessor and/or via embedded components; e.g., via ASIC, coprocessor, DSP, FPGA, and/or the like. Alternately, some implementations of the TVC may be implemented with embedded components that are configured and used to achieve a variety of features or signal processing.
Depending on the particular implementation, the embedded components may include software solutions, hardware solutions, and/or some combination of both hardware/software solutions. For example, TVC features discussed herein may be achieved through implementing FPGAs, which are a semiconductor devices containing programmable logic components called “logic blocks”, and programmable interconnects, such as the high performance FPGA Virtex series and/or the low cost Spartan series manufactured by Xilinx. Logic blocks and interconnects can be programmed by the customer or designer, after the FPGA is manufactured, to implement any of the TVC features. A hierarchy of programmable interconnects allow logic blocks to be interconnected as needed by the TVC system designer/administrator, somewhat like a one-chip programmable breadboard. An FPGA's logic blocks can be programmed to perform the operation of basic logic gates such as AND, and XOR, or more complex combinational operators such as decoders or simple mathematical operations. In most FPGAs, the logic blocks also include memory elements, which may be circuit flip-flops or more complete blocks of memory. In some circumstances, the TVC may be developed on regular FPGAs and then migrated into a fixed version that more resembles ASIC implementations. Alternate or coordinating implementations may migrate TVC controller features to a final ASIC instead of or in addition to FPGAs. Depending on the implementation all of the aforementioned embedded components and microprocessors may be considered the “CPU” and/or “processor” for the TVC.
The power source 4486 may be of any standard form for powering small electronic circuit board devices such as the following power cells: alkaline, lithium hydride, lithium ion, lithium polymer, nickel cadmium, solar cells, and/or the like. Other types of AC or DC power sources may be used as well. In the case of solar cells, in one embodiment, the case provides an aperture through which the solar cell may capture photonic energy. The power cell 4486 is connected to at least one of the interconnected subsequent components of the TVC thereby providing an electric current to all subsequent components. In one example, the power source 4486 is connected to the system bus component 4404. In an alternative embodiment, an outside power source 4486 is provided through a connection across the I/O 4408 interface. For example, a USB and/or IEEE 1394 connection carries both data and power across the connection and is therefore a suitable source of power.
Interface bus(ses) 4407 may accept, connect, and/or communicate to a number of interface adapters, conventionally although not necessarily in the form of adapter cards, such as but not limited to: input output interfaces (I/O) 4408, storage interfaces 4409, network interfaces 4410, and/or the like. Optionally, cryptographic processor interfaces 4427 similarly may be connected to the interface bus. The interface bus provides for the communications of interface adapters with one another as well as with other components of the computer systemization. Interface adapters are adapted for a compatible interface bus. Interface adapters conventionally connect to the interface bus via a slot architecture. Conventional slot architectures may be employed, such as, but not limited to: Accelerated Graphics Port (AGP), Card Bus, (Extended) Industry Standard Architecture ((E)ISA), Micro Channel Architecture (MCA), NuBus, Peripheral Component Interconnect (Extended) (PCI(X)), PCI Express, Personal Computer Memory Card International Association (PCMCIA), and/or the like.
Storage interfaces 4409 may accept, communicate, and/or connect to a number of storage devices such as, but not limited to: storage devices 4414, removable disc devices, and/or the like. Storage interfaces may employ connection protocols such as, but not limited to: (Ultra) (Serial) Advanced Technology Attachment (Packet Interface) ((Ultra) (Serial) ATA(PI)), (Enhanced) Integrated Drive Electronics ((E)IDE), Institute of Electrical and Electronics Engineers (IEEE) 1394, fiber channel, Small Computer Systems Interface (SCSI), Universal Serial Bus (USB), and/or the like.
Network interfaces 4410 may accept, communicate, and/or connect to a communications network 4413. Through a communications network 4413, the TVC controller is accessible through remote clients 4433b (e.g., computers with web browsers) by users 4433a. Network interfaces may employ connection protocols such as, but not limited to: direct connect, Ethernet (thick, thin, twisted pair 10/100/1000 Base T, and/or the like), Token Ring, wireless connection such as IEEE 802.11a-x, and/or the like. Should processing requirements dictate a greater amount speed and/or capacity, distributed network controllers (e.g., Distributed TVC), architectures may similarly be employed to pool, load balance, and/or otherwise increase the communicative bandwidth required by the TVC controller. A communications network may be any one and/or the combination of the following: a direct interconnection; the Internet; a Local Area Network (LAN); a Metropolitan Area Network (MAN); an Operating Missions as Nodes on the Internet (OMNI); a secured custom connection; a Wide Area Network (WAN); a wireless network (e.g., employing protocols such as, but not limited to a Wireless Application Protocol (WAP), I-mode, and/or the like); and/or the like. A network interface may be regarded as a specialized form of an input output interface. Further, multiple network interfaces 4410 may be used to engage with various communications network types 4413. For example, multiple network interfaces may be employed to allow for the communication over broadcast, multicast, and/or unicast networks.
Input Output interfaces (I/O) 4408 may accept, communicate, and/or connect to user input devices 4411, peripheral devices 4412, cryptographic processor devices 4428, and/or the like. I/O may employ connection protocols such as, but not limited to: audio: analog, digital, monaural, RCA, stereo, and/or the like; data: Apple Desktop Bus (ADB), IEEE 1394a-b, serial, universal serial bus (USB); infrared; joystick; keyboard; midi; optical; PC AT; PS/2; parallel; radio; video interface: Apple Desktop Connector (ADC), BNC, coaxial, component, composite, digital, Digital Visual Interface (DVI), high-definition multimedia interface (HDMI), RCA, RF antennae, S-Video, VGA, and/or the like; wireless transceivers: 802.11a/b/g/n/x; Bluetooth; cellular (e.g., code division multiple access (CDMA), high speed packet access (HSPA(+)), high-speed downlink packet access (HSDPA), global system for mobile communications (GSM), long term evolution (LTE), WiMax, etc.); and/or the like. One typical output device may include a video display, which typically comprises a Cathode Ray Tube (CRT) or Liquid Crystal Display (LCD) based monitor with an interface (e.g., DVI circuitry and cable) that accepts signals from a video interface, may be used. The video interface composites information generated by a computer systemization and generates video signals based on the composited information in a video memory frame. Another output device is a television set, which accepts signals from a video interface. Typically, the video interface provides the composited video information through a video connection interface that accepts a video display interface (e.g., an RCA composite video connector accepting an RCA composite video cable; a DVI connector accepting a DVI display cable, etc.).
User input devices 4411 often are a type of peripheral device 4412 (see below) and may include: card readers, dongles, finger print readers, gloves, graphics tablets, joysticks, keyboards, microphones, mouse (mice), remote controls, retina readers, touch screens (e.g., capacitive, resistive, etc.), trackballs, trackpads, sensors (e.g., accelerometers, ambient light, GPS, gyroscopes, proximity, etc.), styluses, and/or the like.
Peripheral devices 4412 may be connected and/or communicate to I/O and/or other facilities of the like such as network interfaces, storage interfaces, directly to the interface bus, system bus, the CPU, and/or the like. Peripheral devices may be external, internal and/or part of the TVC controller. Peripheral devices may include: antenna, audio devices (e.g., line-in, line-out, microphone input, speakers, etc.), cameras (e.g., still, video, webcam, etc.), dongles (e.g., for copy protection, ensuring secure transactions with a digital signature, and/or the like), external processors (for added capabilities; e.g., crypto devices 4428), force-feedback devices (e.g., vibrating motors), network interfaces, printers, scanners, storage devices, transceivers (e.g., cellular, GPS, etc.), video devices (e.g., goggles, monitors, etc.), video sources, visors, and/or the like. Peripheral devices often include types of input devices (e.g., cameras).
It should be noted that although user input devices and peripheral devices may be employed, the TVC controller may be embodied as an embedded, dedicated, and/or monitor-less (i.e., headless) device, wherein access would be provided over a network interface connection.
Cryptographic units such as, but not limited to, microcontrollers, processors 4426, interfaces 4427, and/or devices 4428 may be attached, and/or communicate with the TVC controller. A MC68HC16 microcontroller, manufactured by Motorola Inc., may be used for and/or within cryptographic units. The MC68HC16 microcontroller utilizes a 16-bit multiply-and-accumulate instruction in the 16 MHz configuration and requires less than one second to perform a 512-bit RSA private key operation. Cryptographic units support the authentication of communications from interacting agents, as well as allowing for anonymous transactions. Cryptographic units may also be configured as part of the CPU. Equivalent microcontrollers and/or processors may also be used. Other commercially available specialized cryptographic processors include: the Broadcom's CryptoNetX and other Security Processors; nCipher's nShield, SafeNet's Luna PCI (e.g., 7100) series; Semaphore Communications' 40 MHz Roadrunner 184; Sun's Cryptographic Accelerators (e.g., Accelerator 6000 PCIe Board, Accelerator 500 Daughtercard); Via Nano Processor (e.g., L2100, L2200, U2400) line, which is capable of performing 500+MB/s of cryptographic instructions; VLSI Technology's 33 MHz 6868; and/or the like.
Generally, any mechanization and/or embodiment allowing a processor to affect the storage and/or retrieval of information is regarded as memory 4429. However, memory is a fungible technology and resource, thus, any number of memory embodiments may be employed in lieu of or in concert with one another. It is to be understood that the TVC controller and/or a computer systemization may employ various forms of memory 4429. For example, a computer systemization may be configured wherein the operation of on-chip CPU memory (e.g., registers), RAM, ROM, and any other storage devices are provided by a paper punch tape or paper punch card mechanism; however, such an embodiment would result in an extremely slow rate of operation. In a typical configuration, memory 4429 will include ROM 4406, RAM 4405, and a storage device 4414. A storage device 4414 may be any conventional computer system storage. Storage devices may include a drum; a (fixed and/or removable) magnetic disk drive; a magneto-optical drive; an optical drive (i.e., Blueray, CD ROM/RAM/Recordable (R)/ReWritable (RW), DVD R/RW, HD DVD R/RW etc.); an array of devices (e.g., Redundant Array of Independent Disks (RAID)); solid state memory devices (USB memory, solid state drives (SSD), etc.); other processor-readable storage mediums; and/or other devices of the like. Thus, a computer systemization generally requires and makes use of memory.
The memory 4429 may contain a collection of program and/or database components and/or data such as, but not limited to: operating system component(s) 4415 (operating system); information server component(s) 4416 (information server); user interface component(s) 4417 (user interface); Web browser component(s) 4418 (Web browser); database(s) 4419; mail server component(s) 4421; mail client component(s) 4422; cryptographic server component(s) 4420 (cryptographic server); the TVC component(s) 4435; and/or the like (i.e., collectively a component collection). These components may be stored and accessed from the storage devices and/or from storage devices accessible through an interface bus. Although non-conventional program components such as those in the component collection, typically, are stored in a local storage device 4414, they may also be loaded and/or stored in memory such as: peripheral devices, RAM, remote storage facilities through a communications network, ROM, various forms of memory, and/or the like.
The operating system component 4415 is an executable program component facilitating the operation of the TVC controller. Typically, the operating system facilitates access of I/O, network interfaces, peripheral devices, storage devices, and/or the like. The operating system may be a highly fault tolerant, scalable, and secure system such as: Apple Macintosh OS X (Server); AT&T Nan 9; Be OS; Unix and Unix-like system distributions (such as AT&T's UNIX; Berkley Software Distribution (BSD) variations such as FreeBSD, NetBSD, OpenBSD, and/or the like; Linux distributions such as Red Hat, Ubuntu, and/or the like); and/or the like operating systems. However, more limited and/or less secure operating systems also may be employed such as Apple Macintosh OS, IBM OS/2, Microsoft DOS, Microsoft Windows 2000/2003/3.1/95/98/CE/Millenium/NT/Vista/XP (Server), Palm OS, and/or the like. An operating system may communicate to and/or with other components in a component collection, including itself, and/or the like. Most frequently, the operating system communicates with other program components, user interfaces, and/or the like. For example, the operating system may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses. The operating system, once executed by the CPU, may enable the interaction with communications networks, data, I/O, peripheral devices, program components, memory, user input devices, and/or the like. The operating system may provide communications protocols that allow the TVC controller to communicate with other entities through a communications network 4413. Various communication protocols may be used by the TVC controller as a subcarrier transport mechanism for interaction, such as, but not limited to: multicast, TCP/IP, UDP, unicast, and/or the like.
An information server component 4416 is a stored program component that is executed by a CPU. The information server may be a conventional Internet information server such as, but not limited to Apache Software Foundation's Apache, Microsoft's Internet Information Server, and/or the like. The information server may allow for the execution of program components through facilities such as Active Server Page (ASP), ActiveX, (ANSI) (Objective-) C (++), C# and/or .NET, Common Gateway Interface (CGI) scripts, dynamic (D) hypertext markup language (HTML), FLASH, Java, JavaScript, Practical Extraction Report Language (PERL), Hypertext Pre-Processor (PHP), pipes, Python, wireless application protocol (WAP), WebObjects, and/or the like. The information server may support secure communications protocols such as, but not limited to, File Transfer Protocol (FTP); HyperText Transfer Protocol (HTTP); Secure Hypertext Transfer Protocol (HTTPS), Secure Socket Layer (SSL), messaging protocols (e.g., America Online (AOL) Instant Messenger (AIM), Application Exchange (APEX), ICQ, Internet Relay Chat (IRC), Microsoft Network (MSN) Messenger Service, Presence and Instant Messaging Protocol (PRIM), Internet Engineering Task Force's (IETF's) Session Initiation Protocol (SIP), SIP for Instant Messaging and Presence Leveraging Extensions (SIMPLE), open XML-based Extensible Messaging and Presence Protocol (XMPP) (i.e., Jabber or Open Mobile Alliance's (OMA's) Instant Messaging and Presence Service (IMPS)), Yahoo! Instant Messenger Service, and/or the like. The information server provides results in the form of Web pages to Web browsers, and allows for the manipulated generation of the Web pages through interaction with other program components. After a Domain Name System (DNS) resolution portion of an HTTP request is resolved to a particular information server, the information server resolves requests for information at specified locations on the TVC controller based on the remainder of the HTTP request. For example, a request such as http://123.124.125.126/myInformation.html might have the IP portion of the request “123.124.125.126” resolved by a DNS server to an information server at that IP address; that information server might in turn further parse the http request for the “/myInformation.html” portion of the request and resolve it to a location in memory containing the information “myInformation.html.” Additionally, other information serving protocols may be employed across various ports, e.g., FTP communications across port 21, and/or the like. An information server may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the information server communicates with the TVC database 4419, operating systems, other program components, user interfaces, Web browsers, and/or the like.
Access to the TVC database may be achieved through a number of database bridge mechanisms such as through scripting languages as enumerated below (e.g., CGI) and through inter-application communication channels as enumerated below (e.g., CORBA, WebObjects, etc.). Any data requests through a Web browser are parsed through the bridge mechanism into appropriate grammars as required by the TVC. In one embodiment, the information server would provide a Web form accessible by a Web browser. Entries made into supplied fields in the Web form are tagged as having been entered into the particular fields, and parsed as such. The entered terms are then passed along with the field tags, which act to instruct the parser to generate queries directed to appropriate tables and/or fields. In one embodiment, the parser may generate queries in standard SQL by instantiating a search string with the proper join/select commands based on the tagged text entries, wherein the resulting command is provided over the bridge mechanism to the TVC as a query. Upon generating query results from the query, the results are passed over the bridge mechanism, and may be parsed for formatting and generation of a new results Web page by the bridge mechanism. Such a new results Web page is then provided to the information server, which may supply it to the requesting Web browser.
Also, an information server may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
Computer interfaces in some respects are similar to automobile operation interfaces. Automobile operation interface elements such as steering wheels, gearshifts, and speedometers facilitate the access, operation, and display of automobile resources, and status. Computer interaction interface elements such as check boxes, cursors, menus, scrollers, and windows (collectively and commonly referred to as widgets) similarly facilitate the access, capabilities, operation, and display of data and computer hardware and operating system resources, and status. Operation interfaces are commonly called user interfaces. Graphical user interfaces (GUIs) such as the Apple Macintosh Operating System's Aqua, IBM's OS/2, Microsoft's Windows 2000/2003/3.1/95/98/CE/Millenium/NT/XP/Vista/7 (i.e., Aero), Unix's X-Windows (e.g., which may include additional Unix graphic interface libraries and layers such as K Desktop Environment (KDE), mythTV and GNU Network Object Model Environment (GNOME)), web interface libraries (e.g., ActiveX, AJAX, (D)HTML, FLASH, Java, JavaScript, etc. interface libraries such as, but not limited to, Dojo, jQuery(UI), MooTools, Prototype, script.aculo.us, SWFObject, Yahoo! User Interface, any of which may be used and) provide a baseline and means of accessing and displaying information graphically to users.
A user interface component 4417 is a stored program component that is executed by a CPU. The user interface may be a conventional graphic user interface as provided by, with, and/or atop operating systems and/or operating environments such as already discussed. The user interface may allow for the display, execution, interaction, manipulation, and/or operation of program components and/or system facilities through textual and/or graphical facilities. The user interface provides a facility through which users may affect, interact, and/or operate a computer system. A user interface may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the user interface communicates with operating systems, other program components, and/or the like. The user interface may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
A Web browser component 4418 is a stored program component that is executed by a CPU. The Web browser may be a conventional hypertext viewing application such as Microsoft Internet Explorer or Netscape Navigator. Secure Web browsing may be supplied with 128 bit (or greater) encryption by way of HTTPS, SSL, and/or the like. Web browsers allowing for the execution of program components through facilities such as ActiveX, AJAX, (D)HTML, FLASH, Java, JavaScript, web browser plug-in APIs (e.g., FireFox, Safari Plug-in, and/or the like APIs), and/or the like. Web browsers and like information access tools may be integrated into PDAs, cellular telephones, and/or other mobile devices. A Web browser may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the Web browser communicates with information servers, operating systems, integrated program components (e.g., plug-ins), and/or the like; e.g., it may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses. Also, in place of a Web browser and information server, a combined application may be developed to perform similar operations of both. The combined application would similarly affect the obtaining and the provision of information to users, user agents, and/or the like from the TVC enabled nodes. The combined application may be nugatory on systems employing standard Web browsers.
A mail server component 4421 is a stored program component that is executed by a CPU 4403. The mail server may be a conventional Internet mail server such as, but not limited to sendmail, Microsoft Exchange, and/or the like. The mail server may allow for the execution of program components through facilities such as TVC, ActiveX, (ANSI) (Objective-) C (++), C# and/or .NET, CGI scripts, Java, JavaScript, PERL, PHP, pipes, Python, WebObjects, and/or the like. The mail server may support communications protocols such as, but not limited to: Internet message access protocol (IMAP), Messaging Application Programming Interface (MAPI)/Microsoft Exchange, post office protocol (POP3), simple mail transfer protocol (SMTP), and/or the like. The mail server can route, forward, and process incoming and outgoing mail messages that have been sent, relayed and/or otherwise traversing through and/or to the TVC.
Access to the TVC mail may be achieved through a number of APIs offered by the individual Web server components and/or the operating system.
Also, a mail server may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, information, and/or responses.
A mail client component 4422 is a stored program component that is executed by a CPU 4403. The mail client may be a conventional mail viewing application such as Apple Mail, Microsoft Entourage, Microsoft Outlook, Microsoft Outlook Express, Mozilla, Thunderbird, and/or the like. Mail clients may support a number of transfer protocols, such as: IMAP, Microsoft Exchange, POP3, SMTP, and/or the like. A mail client may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the mail client communicates with mail servers, operating systems, other mail clients, and/or the like; e.g., it may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, information, and/or responses. Generally, the mail client provides a facility to compose and transmit electronic mail messages.
A cryptographic server component 4420 is a stored program component that is executed by a CPU 4403, cryptographic processor 4426, cryptographic processor interface 4427, cryptographic processor device 4428, and/or the like. Cryptographic processor interfaces will allow for expedition of encryption and/or decryption requests by the cryptographic component; however, the cryptographic component, alternatively, may run on a conventional CPU. The cryptographic component allows for the encryption and/or decryption of provided data. The cryptographic component allows for both symmetric and asymmetric (e.g., Pretty Good Protection (PGP)) encryption and/or decryption. The cryptographic component may employ cryptographic techniques such as, but not limited to: digital certificates (e.g., X.509 authentication framework), digital signatures, dual signatures, enveloping, password access protection, public key management, and/or the like. The cryptographic component will facilitate numerous (encryption and/or decryption) security protocols such as, but not limited to: checksum, Data Encryption Standard (DES), Elliptical Curve Encryption (ECC), International Data Encryption Algorithm (IDEA), Message Digest 5 (MD5, which is a one way hash operation), passwords, Rivest Cipher (RC5), Rijndael, RSA (which is an Internet encryption and authentication system that uses an algorithm developed in 1977 by Ron Rivest, Adi Shamir, and Leonard Adleman), Secure Hash Algorithm (SHA), Secure Socket Layer (SSL), Secure Hypertext Transfer Protocol (HTTPS), and/or the like. Employing such encryption security protocols, the TVC may encrypt all incoming and/or outgoing communications and may serve as node within a virtual private network (VPN) with a wider communications network. The cryptographic component facilitates the process of “security authorization” whereby access to a resource is inhibited by a security protocol wherein the cryptographic component effects authorized access to the secured resource. In addition, the cryptographic component may provide unique identifiers of content, e.g., employing and MD5 hash to obtain a unique signature for an digital audio file. A cryptographic component may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. The cryptographic component supports encryption schemes allowing for the secure transmission of information across a communications network to enable the TVC component to engage in secure transactions if so desired. The cryptographic component facilitates the secure accessing of resources on the TVC and facilitates the access of secured resources on remote systems; i.e., it may act as a client and/or server of secured resources. Most frequently, the cryptographic component communicates with information servers, operating systems, other program components, and/or the like. The cryptographic component may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
The TVC database component 4419 may be embodied in a database and its stored data. The database is a stored program component, which is executed by the CPU; the stored program component portion configuring the CPU to process the stored data. The database may be a conventional, fault tolerant, relational, scalable, secure database such as Oracle or Sybase. Relational databases are an extension of a flat file. Relational databases consist of a series of related tables. The tables are interconnected via a key field. Use of the key field allows the combination of the tables by indexing against the key field; i.e., the key fields act as dimensional pivot points for combining information from various tables. Relationships generally identify links maintained between tables by matching primary keys. Primary keys represent fields that uniquely identify the rows of a table in a relational database. More precisely, they uniquely identify rows of a table on the “one” side of a one-to-many relationship.
Alternatively, the TVC database may be implemented using various standard data-structures, such as an array, hash, (linked) list, struct, structured text file (e.g., XML), table, and/or the like. Such data-structures may be stored in memory and/or in (structured) files. In another alternative, an object-oriented database may be used, such as Frontier, ObjectStore, Poet, Zope, and/or the like. Object databases can include a number of object collections that are grouped and/or linked together by common attributes; they may be related to other object collections by some common attributes. Object-oriented databases perform similarly to relational databases with the exception that objects are not just pieces of data but may have other types of capabilities encapsulated within a given object. If the TVC database is implemented as a data-structure, the use of the TVC database 4419 may be integrated into another component such as the TVC component 4435. Also, the database may be implemented as a mix of data structures, objects, and relational structures. Databases may be consolidated and/or distributed in countless variations through standard data processing techniques. Portions of databases, e.g., tables, may be exported and/or imported and thus decentralized and/or integrated.
In one embodiment, the database component 4419 includes several tables 4419a-q. A Users table 4419a may include fields such as, but not limited to: user_id, ssn, dob, first_name, last_name, age, state, address_firstline, address_secondline, zipcode, devices_list, contact_info, contact_type, alt_contact_info, alt_contact_type, user_gender, user_clothing_size, user_body_type, user_eye_color, user_hair_color, user_complexion, user_personalized_gesture_models, user_recommended_items, user_image, user_image date, user_body_joint_location, and/or the like. The Users table may support and/or track multiple entity accounts on a TVC. A Devices table 4419b may include fields such as, but not limited to: device_ID, device_name, device_IP, device_GPS, device_MAC, device_serial, device_ECID, device_UDID, device_browser, device_type, device_model, device_version, device_OS, device_apps_list, device_securekey, wallet_app_installed_flag, and/or the like. An Apps table 4419c may include fields such as, but not limited to: app_ID, app_name, app_type, app_dependencies, app_access_code, user_pin, and/or the like. An Accounts table 4419d may include fields such as, but not limited to: account_number, account_security_code, account_name, issuer_acquirer_flag, issuer_name, acquirer_name, account_address, routing_number, access_API_call, linked_wallets_list, and/or the like. A Merchants table 4419e may include fields such as, but not limited to: merchant_id, merchant_name, merchant address, store_id, ip_address, mac_address, auth_key, port_num, security_settings_list, and/or the like. An Issuers table 4419f may include fields such as, but not limited to: issuer_id, issuer_name, issuer_address, ip_address, mac_address, auth_key, port_num, security_settings_list, and/or the like. An Acquirers table 4419g may include fields such as, but not limited to: account_firstname, account_lastname, account_type, account_num, account_balance_list, billingaddress_line1, billingaddress_line2, billing_zipcode, billing_state, shipping_preferences, shippingaddress_line1, shippingaddress_line2, shipping_zipcode, shipping_state, and/or the like. A Pay Gateways table 4419h may include fields such as, but not limited to: gateway_ID, gateway_IP, gateway_MAC, gateway_secure_key, gateway_access_list, gateway_API_call_list, gateway_services_list, and/or the like. A Shop Sessions table 4419i may include fields such as, but not limited to: user_id, session_id, alerts_URL, timestamp, expiry_lapse, merchant_id, store_id, device_type, device_ID, device_IP, device_MAC, device_browser, device_serial, device_ECID, device_model, device_OS, wallet_app_installed, total_cost, cart_ID_list, product_params_list, social_flag, social_message, social_networks_list, coupon_lists, accounts_list, CVV2_lists, charge_ratio_list, charge_priority_list, value_exchange_symbols_list, bill_address, ship_address, cloak_flag, pay_mode, alerts_rules_list, and/or the like. A Transactions table 4419j may include fields such as, but not limited to: order_id, user_id, timestamp, transaction_cost, purchase_details_list, num_products, products_list, product_type, product_params_list, product_title, product_summary, quantity, user_id, client_id, client_ip, client_type, client_model, operating_system, os_version, app_installed_flag, user_id, account_firstname, account_lastname, account_type, account_num, account_priority_account_ratio, billingaddress_line1, billingaddress_line2, billing_zipcode, billing_state, shipping_preferences, shippingaddress_line1, shippingaddress_line2, shipping_zipcode, shipping_state, merchant_id, merchant_name, merchant_auth_key, and/or the like. A Batches table 4419k may include fields such as, but not limited to: batch_id, transaction_id_list, timestamp_list, cleared_flag_list, clearance_trigger_settings, and/or the like. A Ledgers table 44191 may include fields such as, but not limited to: request_id, timestamp, deposit_amount, batch_id, transaction_id, clear_flag, deposit_account, transaction_summary, payor_name, payor_account, and/or the like. A Products table 4419m may include fields such as, but not limited to: product_ID, product_title, product_attributes_list, product_price, tax_info_list, related_products_list, offers_list, discounts_list, rewards_list, merchants_list, merchant_availability_list, product_date_added, product_image, product_qr, product_manufacturer, product_model, product_aisle, product_stack, product_shelf, product_type, and/or the like. An Offers table 4419n may include fields such as, but not limited to: offer_ID, offer_title, offer_attributes_list, offer_price, offer_expiry, related_products_list, discounts_list, rewards_list, merchants_list, merchant_availability_list, and/or the like. A Behavior Data table 44190 may include fields such as, but not limited to: user_id, timestamp, activity_type, activity_location, activity_attribute_list, activity_attribute_values_list, and/or the like. A Label Analytics table 4419p may include fields such as, but not limited to: label_id, label_name, label_format, label_account_type, labe_session_id, label_session_type, label_product_id, label_product_type, Label_transaction_id, label_transaction_type, and/or the like. A Social table 4419q may include fields such as, but not limited to: social_id, social_name, social_server_id, social_server_ip, social_domain_id, social_source, social_feed_id, social_feed_source, social_comment, social_comment_time, social_comment_keyterms, social_comment_product_id, and/or the like. A MDGA table 4419r includes fields such as, but not limited to: MDGA_id, MDGA_name, MDGA_touch_gestures, MDGA_finger_gestures, MDGA_QR_gestures, MDGA_object_gestures, MDGA_vocal_commands, MDGA_merchant, and/or the like. The MDGA table may support and/or track multiple possible composite actions on a TVC. A payment device table 4419s includes fields such as, but not limited to: pd_id, pd_user, pd_type, pd_issuer, pd_issuer_id, pd_qr, pd_date_added, and/or the like. The payment device table may support and/or track multiple payment devices used on a TVC. An object gestures table 4419t includes fields such as, but not limited to: object_gesture_id, object_gesture_type, object_gesture_x, object_gesture_x, object_gesture_merchant, and/or the like. The object gesture table may support and/or track multiple object gestures performed on a TVC. A touch gesture table 4419u includes fields such as, but not limited to: touch_gesture_id, touch_gesture_type, touch_gesture_x, touch_gesture_x, touch_gesture_merchant, and/or the like. The touch gestures table may support and/or track multiple touch gestures performed on a TVC.A finger gesture table 4419v includes fields such as, but not limited to: finger_gesture_id, finger_gesture_type, finger_gesture_x, finger_gesture_x, finger_gesture_merchant, and/or the like. The finger gestures table may support and/or track multiple finger gestures performed on a TVC. A QR gesture table 4419w includes fields such as, but not limited to: QR_gesture_id, QR_gesture_type, QR_gesture_x, QR_gesture_x, QR_gesture_merchant, and/or the like. The QR gestures table may support and/or track multiple QR gestures performed on a TVC. A vocal command table 4419x includes fields such as, but not limited to: vc_id, vc_name, vc_command_list, and/or the like. The vocal command gestures table may support and/or track multiple vocal commands performed on a TVC.
In one embodiment, the TVC database may interact with other database systems. For example, employing a distributed database system, queries and data access by search TVC component may treat the combination of the TVC database, an integrated data security layer database as a single database entity.
In one embodiment, user programs may contain various user interface primitives, which may serve to update the TVC. Also, various accounts may require custom database tables depending upon the environments and the types of clients the TVC may need to serve. It should be noted that any unique fields may be designated as a key field throughout. In an alternative embodiment, these tables have been decentralized into their own databases and their respective database controllers (i.e., individual database controllers for each of the above tables). Employing standard data processing techniques, one may further distribute the databases over several computer systemizations and/or storage devices. Similarly, configurations of the decentralized database controllers may be varied by consolidating and/or distributing the various database components 4419a-x. The TVC may be configured to keep track of various settings, inputs, and parameters via database controllers.
The TVC database may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the TVC database communicates with the TVC component, other program components, and/or the like. The database may contain, retain, and provide information regarding other nodes and data.
The TVC component 4435 is a stored program component that is executed by a CPU. In one embodiment, the TVC component incorporates any and/or all combinations of the aspects of the TVC discussed in the previous figures. As such, the TVC affects accessing, obtaining and the provision of information, services, transactions, and/or the like across various communications networks.
The TVC component may transform reality scene visual captures (e.g., see 213 in
The TVC component enabling access of information between nodes may be developed by employing standard development tools and languages such as, but not limited to: Apache components, Assembly, ActiveX, binary executables, (ANSI) (Objective-) C (++), C# and/or .NET, database adapters, CGI scripts, Java, JavaScript, mapping tools, procedural and object oriented development tools, PERL, PHP, Python, shell scripts, SQL commands, web application server extensions, web development environments and libraries (e.g., Microsoft's ActiveX; Adobe AIR, FLEX & FLASH; AJAX; (D)HTML; Dojo, Java; JavaScript; jQuery(UI); MooTools; Prototype; script.aculo.us; Simple Object Access Protocol (SOAP); SWFObject; Yahoo! User Interface; and/or the like), WebObjects, and/or the like. In one embodiment, the TVC server employs a cryptographic server to encrypt and decrypt communications. The TVC component may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the TVC component communicates with the TVC database, operating systems, other program components, and/or the like. The TVC may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
The structure and/or operation of any of the TVC node controller components may be combined, consolidated, and/or distributed in any number of ways to facilitate development and/or deployment. Similarly, the component collection may be combined in any number of ways to facilitate deployment and/or development. To accomplish this, one may integrate the components into a common code base or in a facility that can dynamically load the components on demand in an integrated fashion.
The component collection may be consolidated and/or distributed in countless variations through standard data processing and/or development techniques. Multiple instances of any one of the program components in the program component collection may be instantiated on a single node, and/or across numerous nodes to improve performance through load-balancing and/or data-processing techniques. Furthermore, single instances may also be distributed across multiple controllers and/or storage devices; e.g., databases. All program component instances and controllers working in concert may do so through standard data processing communication techniques.
The configuration of the TVC controller will depend on the context of system deployment. Factors such as, but not limited to, the budget, capacity, location, and/or use of the underlying hardware resources may affect deployment requirements and configuration. Regardless of if the configuration results in more consolidated and/or integrated program components, results in a more distributed series of program components, and/or results in some combination between a consolidated and distributed configuration, data may be communicated, obtained, and/or provided. Instances of components consolidated into a common code base from the program component collection may communicate, obtain, and/or provide data. This may be accomplished through intra-application data processing communication techniques such as, but not limited to: data referencing (e.g., pointers), internal messaging, object instance variable communication, shared memory space, variable passing, and/or the like.
If component collection components are discrete, separate, and/or external to one another, then communicating, obtaining, and/or providing data with and/or to other components may be accomplished through inter-application data processing communication techniques such as, but not limited to: Application Program Interfaces (API) information passage; (distributed) Component Object Model ((D)COM), (Distributed) Object Linking and Embedding ((D)OLE), and/or the like), Common Object Request Broker Architecture (CORBA), Jini local and remote application program interfaces, JavaScript Object Notation (JSON), Remote Method Invocation (RMI), SOAP, process pipes, shared files, and/or the like. Messages sent between discrete component components for inter-application communication or within memory spaces of a singular component for intra-application communication may be facilitated through the creation and parsing of a grammar. A grammar may be developed by using development tools such as lex, yacc, XML, and/or the like, which allow for grammar generation and parsing capabilities, which in turn may form the basis of communication messages within and between components.
For example, a grammar may be arranged to recognize the tokens of an HTTP post command, e.g.:
where Value1 is discerned as being a parameter because “http://” is part of the grammar syntax, and what follows is considered part of the post value. Similarly, with such a grammar, a variable “Value1” may be inserted into an “http://” post command and then sent. The grammar syntax itself may be presented as structured data that is interpreted and/or otherwise used to generate the parsing mechanism (e.g., a syntax description text file as processed by lex, yacc, etc.). Also, once the parsing mechanism is generated and/or instantiated, it itself may process and/or parse structured data such as, but not limited to: character (e.g., tab) delineated text, HTML, structured text streams, XML, and/or the like structured data. In another embodiment, inter-application data processing protocols themselves may have integrated and/or readily available parsers (e.g., JSON, SOAP, and/or like parsers) that may be employed to parse (e.g., communications) data. Further, the parsing grammar may be used beyond message parsing, but may also be used to parse: databases, data collections, data stores, structured data, and/or the like. Again, the desired configuration will depend upon the context, environment, and requirements of system deployment.
For example, in some implementations, the TVC controller may be executing a PHP script implementing a Secure Sockets Layer (“SSL”) socket server via the information server, which listens to incoming communications on a server port to which a client may send data, e.g., data encoded in JSON format. Upon identifying an incoming communication, the PHP script may read the incoming message from the client device, parse the received JSON-encoded text data to extract information from the JSON-encoded text data into PHP script variables, and store the data (e.g., client identifying information, etc.) and/or extracted information in a relational database accessible using the Structured Query Language (“SQL”). An exemplary listing, written substantially in the form of PHP/SQL commands, to accept JSON-encoded input data from a client device via a SSL connection, parse the data to extract variables, and store the data to a database, is provided below:
Also, the following resources may be used to provide example embodiments regarding SOAP parser implementation:
and other parser implementations:
all of which are hereby expressly incorporated by reference herein.
In order to address various issues and advance the art, the entirety of this application for TRANSACTION VISUAL CAPTURING APPARATUSES, METHODS AND SYSTEMS (including the Cover Page, Title, Headings, Field, Background, Summary, Brief Description of the Drawings, Detailed Description, Claims, Abstract, Figures, Appendices and/or otherwise) shows by way of illustration various embodiments in which the claimed innovations may be practiced. The advantages and features of the application are of a representative sample of embodiments only, and are not exhaustive and/or exclusive. They are presented only to assist in understanding and teach the claimed principles. It should be understood that they are not representative of all claimed innovations. As such, certain aspects of the disclosure have not been discussed herein. That alternate embodiments may not have been presented for a specific portion of the innovations or that further undescribed alternate embodiments may be available for a portion is not to be considered a disclaimer of those alternate embodiments. It will be appreciated that many of those undescribed embodiments incorporate the same principles of the innovations and others are equivalent. Thus, it is to be understood that other embodiments may be utilized and functional, logical, operational, organizational, structural and/or topological modifications may be made without departing from the scope and/or spirit of the disclosure. As such, all examples and/or embodiments are deemed to be non-limiting throughout this disclosure. Also, no inference should be drawn regarding those embodiments discussed herein relative to those not discussed herein other than it is as such for purposes of reducing space and repetition. For instance, it is to be understood that the logical and/or topological structure of any combination of any program components (a component collection), other components and/or any present feature sets as described in the figures and/or throughout are not limited to a fixed operating order and/or arrangement, but rather, any disclosed order is exemplary and all equivalents, regardless of order, are contemplated by the disclosure. Furthermore, it is to be understood that such features are not limited to serial execution, but rather, any number of threads, processes, services, servers, and/or the like that may execute asynchronously, concurrently, in parallel, simultaneously, synchronously, and/or the like are contemplated by the disclosure. As such, some of these features may be mutually contradictory, in that they cannot be simultaneously present in a single embodiment. Similarly, some features are applicable to one aspect of the innovations, and inapplicable to others. In addition, the disclosure includes other innovations not presently claimed. Applicant reserves all rights in those presently unclaimed innovations, including the right to claim such innovations, file additional applications, continuations, continuations in part, divisions, and/or the like thereof. As such, it should be understood that advantages, embodiments, examples, functional, features, logical, operational, organizational, structural, topological, and/or other aspects of the disclosure are not to be considered limitations on the disclosure as defined by the claims or limitations on equivalents to the claims. It is to be understood that, depending on the particular needs and/or characteristics of a TVC individual and/or enterprise user, database configuration and/or relational model, data type, data transmission and/or network framework, syntax structure, and/or the like, various embodiments of the TVC may be implemented that enable a great deal of flexibility and customization. For example, aspects of the TVC may be adapted for (electronic/financial) trading systems, financial planning systems, and/or the like. While various embodiments and discussions of the TVC have been directed to retail commerce, however, it is to be understood that the embodiments described herein may be readily configured and/or customized for a wide variety of other applications and/or implementations.
For example, further embodiments may include:
1. An augmented retail shopping processor-implemented method, comprising:
obtaining a user shopping assistance request including user check-in information from a user mobile device upon user entry into a merchant store to engage in a shopping experience;
2. An augmented retail shopping processor-implemented method, comprising:
3. The method of embodiment 2, wherein the user check-in message is generated by a user snapping a merchant store provided quick response (QR) code.
4. The method of embodiment 2, wherein the user check-in message is sent to a remote server.
5. The method of embodiment 2, wherein the user check-in message includes geo-location information of the user.
6. The method of embodiment 2, wherein the merchant store assigns a sales clerk to the user upon user check-in at the merchant store.
7. The method of embodiment 6, wherein the sales clerk comprises any of a store employee and a virtual shopping assistant.
8. The method of embodiment 6, wherein the sales clerk assignment is determined based on user loyalty levels.
9. The method of embodiment 6, wherein the sales clerk comprises any of a local representative and a remote representative.
10. The method of embodiment 2, wherein the user profile comprises user loyalty information and past purchasing history with the merchant store.
11. The method of embodiment 2, wherein the user profile is previously stored at a local database at the merchant store.
12. The method of embodiment 2, wherein the user profile is stored at a remote server and transmitted to the merchant store.
13. The method of embodiment 2, wherein the real-time in-store behavior data comprises any of:
14. The method of embodiment 2, wherein the product purchase recommendation comprises any of:
15. The method of embodiment 14, further comprising:
16. The method of embodiment 2, further comprising:
receiving a user communication indicating shopping interests.
17. The method of embodiment 16, wherein the user communication is conducted via any of:
in-person communication between the user and a sales clerk;
video chat;
audio chat;
instant messages; and
text messages.
18. The method of embodiment 16, wherein the shopping interests further comprises:
19. The method of embodiment 16, wherein the shopping interests further comprises:
20. The method of embodiment 16, wherein the shopping interests further comprises:
21. The method of embodiment 16, wherein the shopping interests further comprises:
22. The method of embodiment 2, wherein check-out information page includes a QR code encoding product item information and a payment amount due.
23. The method of embodiment 22, wherein the purchase transaction is initiated upon the user snapping the QR code using the user mobile device, and submitting a wallet payment request to an electronic payment processing network.
24. The method of embodiment 22, wherein the purchase transaction is initiated at the merchant store.
25. The method of embodiment 22, wherein the electronic receipt is sent to the user mobile device via a third party notification system.
26. The method of embodiment 22, wherein the electronic receipt is provided by the merchant store.
27. The method of embodiment 2, further comprising:
maintaining a shopping cart for the user; and
adding the product item to the shopping cart.
28. The method of embodiment 2, further comprising:
receiving a shopping list from the user mobile device; and
obtaining product item information from the shopping list.
29. The method of embodiment 28, further comprising:
obtaining inventory information and stock keeping unit (SKU) information of the obtained product information; and
generating a store map with tags indicating locations of product items on the shopping list.
30. The method of embodiment 28, further comprising:
generating an augmenter reality in-store scan indicating locations of product items on the shopping list.
31. An augmented retail shopping system, comprising:
means for obtaining a user check-in message indicating user entry at a merchant store from a user mobile device;
means for retrieving a user profile associated with the merchant store;
means for obtaining user real-time in-store behavior data from the user mobile device;
means for generating a product purchase recommendation based on the user profile and the user real-time in-store behavior;
means for providing the product purchase recommendation to the user;
means for obtaining a user interests indication that the user wishes to make a purchase of a product;
means for initiating a purchase transaction for the product; and
means for providing an electronic receipt to the user mobile device for the purchase transaction upon completion of the purchase transaction.
32. An augmented retail shopping apparatus, comprising:
a processor; and
a memory disposed in communication with the processor and storing processor-executable instructions to:
33. An augmented retail shopping non-transitory computer-readable medium storing processor-executable instructions, said instructions executable by a processor to:
34. A payment transaction visual capturing processor-implemented method, comprising:
35. The method of embodiment 34, wherein the identified object comprises any of a payment card, an invoice and a purchase item.
36. The method of embodiment 34, wherein the user interactive option labels comprise any of the labels for fund transfer, view balance, and pay for a purchase.
37. A payment transaction visual capturing processor-implemented method, comprising:
38. The method of embodiment 37, wherein the user activity records include any of a web search key term, a GPS location check-in event, and a browsing history.
39. The method of embodiment 37, wherein two or more objects are identified from the captured reality scene, and each of the two or more objects is associated with augmented reality virtual labels.
40. The method of embodiment 37, further comprising:
determining a fingertip motion within the captured reality scene.
41. A transaction visual capturing processor-implemented method, comprising:
42. The method of embodiment 41, wherein the identified first object is a financial payment card having an account resolvable identifier.
43. The method of embodiment 41, wherein the identified second object is a financial payment card having an account resolvable identifier.
44. The method of embodiment 41, wherein the identified second object is a sales bill including a QR code.
45. The method of embodiment 41, wherein the identified second object is a metro card.
46. The method of embodiment 41, wherein the payment from the first financial account to the second financial account comprises a fund transfer from one financial payment card to another financial payment card.
47. The method of embodiment 41, wherein the payment from the first financial account to the second financial account comprises a bill payment from a financial payment card to a merchant for a product purchase.
48. The method of embodiment 41, wherein the payment from the first financial account to the second financial account comprises a fund refill from a financial payment card to a metro card.
49. The method of embodiment 41, wherein the image processing comprises obtaining screen grabs of the obtained live visual capture.
50. The method of embodiment 41, wherein the user transaction command comprises an audio command.
51. The method of embodiment 41, wherein the user transaction command comprises a fingertip motion of moving from the first object to the second object.
52. The method of embodiment 41, further comprising:
obtaining information pertaining to the identified first financial account and the identified second object based on the determined first account identifier.
53. The method of embodiment 41, further comprising:
54. The method of embodiment 41, wherein the first account identifier and the second account identifier are visibly determinable via any of:
barcode reading;
QR code decoding; and
optical character recognition (OCR).
55. The method of embodiment 41, further comprising:
obtaining authorization credentials for the payment from the first financial account to the second financial account.
56. The method of embodiment 55, further comprising:
requesting a user to input a passcode for user identify confirmation.
57. The method of embodiment 41, wherein the first account identifier comprises a 16 digit bank card number.
58. The method of embodiment 41, wherein the second account identifier comprises a merchant identifier.
59. The method of embodiment 41, wherein the second account identifier comprises a 16 digit bank card number.
60. The method of embodiment 41, further comprising:
generating a security alert request when the second object comprises a financial payment card with a cardholder; and
sending the security alert to the cardholder of the second object.
61. A visual capturing processor-implemented method, comprising:
obtaining a list of product items indicating user demands at a user mobile device;
determining a product category and a product identifier for each product item on the obtained list of product items;
obtaining a user indication of a merchant store;
obtaining product inventory and stock keeping data of the merchant store;
querying the obtained product inventory and stock keeping data based on the product identifier and the product category for each product item;
determining an in-store stock keeping location for each product item based on the query;
obtaining a visual layout of the merchant store;
tagging the visual layout of the merchant store with the determined in-store stock keeping location for each product item; and
presenting the tagged visual layout of the merchant store at the user mobile device.
62. The method of embodiment 61, wherein the list of product items comprises a shopping list entered by a user.
63. The method of embodiment 62, wherein the shopping list is generated via audio commands from the user.
64. The method of embodiment 62, wherein the shopping list is generated by extracting product item information from a previously stored sales receipt.
65. The method of embodiment 61, wherein the user indication of the merchant store comprises a user check-in message at a merchant store.
66. The method of embodiment 61, wherein the user indication of the merchant store comprises GPS coordinates of a user.
67. The method of embodiment 61, wherein the product inventory and stock keeping data comprises a table listing an aisle number and a stack number of an in-stock product at the merchant store.
68. The method of embodiment 61, wherein the in-store stock keeping location for each product item comprises any of a aisle number, a stack number, and a shelf number.
69. The method of embodiment 61, wherein the visual layout of the merchant store comprises a static store floor plan map.
70. The method of embodiment 69, further comprising highlighting the static store floor plan map with labels illustrating a location of each product item.
71. The method of embodiment 61, wherein the visual layout of the merchant store comprises a live visual capture of an in-store reality scene.
72. The method of embodiment 71, further comprising injecting user interactive augmented reality labels overlaying the live visual capture of the in-store reality scene, said augmented reality labels indicating a location of each product item within the in-store reality scene.
73. The method of embodiment 72, wherein said augmented reality labels may comprise a semi-transparent bound box covering a product item within the in-store reality scene.
74. The method of embodiment 61, wherein more than one merchant stores are processed for multi-merchant shopping.
75. An augmented retail shopping apparatus, comprising:
a processor; and
a memory in communication with the processor containing processor-readable instructions to:
76. An augmented retail shopping system, comprising:
77. An augmented retail shopping non-transitory computer-readable medium storing processor-executable instructions, said instructions executable by a processor to:
obtain a user shopping assistance request including user check-in information from a user mobile device upon user entry into a merchant store to engage in a shopping experience;
extract a user identifier based on the user check-in information;
access a database for a user profile based on the extracted user identifier;
determine a user prior behavior pattern from the accessed user profile;
obtain user real-time in-store behavior data from the user mobile device;
generate a product purchase recommendation using the user real-time in-store behavior and the user prior behavior pattern;
provide, via a network communication device over a merchant network, the product purchase recommendation to the user mobile device;
add a product for purchase by the user to a shopping cart over the merchant network, based on the provided recommendation;
obtain a transaction interests indication that the user wishes to purchase the product added to the cart;
provide a check-out information page to the user including product item information and payment information;
initiate a purchase transaction for the product added to the cart through an encrypted, non-merchant, bandwidth and network latency reducing, and out-of-band network communication via an electronic payment communication network; and
provide an electronic receipt to the user mobile device for the purchase transaction for the product added to the cart.
78. The apparatus of embodiment 31, wherein the user check-in message is generated by a user snapping a merchant store provided quick response (QR) code.
79. The system of embodiment 31, wherein the user check-in message is sent to a remote server.
80. The system of embodiment 31, wherein the user check-in message includes geo-location information of the user.
81. The system of embodiment 31, wherein the merchant store assigns a sales clerk to the user upon user check-in at the merchant store.
82. The system of embodiment 81, wherein the sales clerk comprises any of a store employee and a virtual shopping assistant.
83. The system of embodiment 81, wherein the sales clerk assignment is determined based on user loyalty levels.
84. The system of embodiment 81, wherein the sales clerk comprises any of a local representative and a remote representative.
85. The system of embodiment 31, wherein the user profile comprises user loyalty information and past purchasing history with the merchant store.
86. The system of embodiment 31, wherein the user profile is previously stored at a local database at the merchant store.
87. The system of embodiment 31, wherein the user profile is stored at a remote server and transmitted to the merchant store.
88. The system of embodiment 31, wherein the real-time in-store behavior data comprises any of:
89. The system of embodiment 31, wherein the product purchase recommendation comprises any of:
90. The system of embodiment 89, further comprising:
91. The system of embodiment 31, further comprising:
means for receiving a user communication indicating shopping interests.
92. The system of embodiment 91, wherein the user communication is conducted via any of:
in-person communication between the user and a sales clerk;
video chat;
audio chat;
instant messages; and
text messages.
93. The system of embodiment 91, wherein the shopping interests further comprises:
94. The system of embodiment 91, wherein the shopping interests further comprises:
95. The system of embodiment 91, wherein the shopping interests further comprises:
96. The system of embodiment 91, wherein the shopping interests further comprises:
97. The system of embodiment 31, wherein check-out information page includes a QR code encoding product item information and a payment amount due.
98. The system of embodiment 97, wherein the purchase transaction is initiated upon the user snapping the QR code using the user mobile device, and submitting a wallet payment request to an electronic payment processing network.
99. The system of embodiment 97, wherein the purchase transaction is initiated at the merchant store.
100. The system of embodiment 97, wherein the electronic receipt is sent to the user mobile device via a third party notification system.
101. The system of embodiment 97, wherein the electronic receipt is provided by the merchant store.
102. The system of embodiment 31, further comprising:
means for maintaining a shopping cart for the user; and
means for adding the product item to the shopping cart.
103. The system of embodiment 31, further comprising:
means for receiving a shopping list from the user mobile device; and
means for obtaining product item information from the shopping list.
104. The system of embodiment 31, further comprising:
means for obtaining inventory information and stock keeping unit (SKU) information of the obtained product information; and
means for generating a store map with tags indicating locations of product items on the shopping list.
105. The system of embodiment 31, further comprising:
means for generating an augmenter reality in-store scan indicating locations of product items on the shopping list.
106. The apparatus of embodiment 32, wherein the user check-in message is generated by a user snapping a merchant store provided quick response (QR) code.
107. The apparatus of embodiment 32, wherein the user check-in message is sent to a remote server.
108. The apparatus of embodiment 32, wherein the user check-in message includes geo-location information of the user.
109. The apparatus of embodiment 32, wherein the merchant store assigns a sales clerk to the user upon user check-in at the merchant store.
110. The apparatus of embodiment 109, wherein the sales clerk comprises any of a store employee and a virtual shopping assistant.
111. The apparatus of embodiment 109, wherein the sales clerk assignment is determined based on user loyalty levels.
112. The apparatus of embodiment 109, wherein the sales clerk comprises any of a local representative and a remote representative.
113. The apparatus of embodiment 32, wherein the user profile comprises user loyalty information and past purchasing history with the merchant store.
114. The apparatus of embodiment 32, wherein the user profile is previously stored at a local database at the merchant store.
115. The apparatus of embodiment 32, wherein the user profile is stored at a remote server and transmitted to the merchant store.
116. The apparatus of embodiment 32, wherein the real-time in-store behavior data comprises any of:
117. The apparatus of embodiment 32, wherein the product purchase recommendation comprises any of:
118. The apparatus of embodiment 117, further comprising instructions to:
119. The apparatus of embodiment 32, further comprising instructions to:
receive a user communication indicating shopping interests.
120. The apparatus of embodiment 119, wherein the user communication is conducted via any of:
in-person communication between the user and a sales clerk;
video chat;
audio chat;
instant messages; and
text messages.
121. The apparatus of embodiment 119, wherein the shopping interests further comprises:
122. The apparatus of embodiment 119, wherein the shopping interests further comprises:
123. The apparatus of embodiment 119, wherein the shopping interests further comprises:
124. The apparatus of embodiment 119, wherein the shopping interests further comprises:
125. The apparatus of embodiment 32, wherein check-out information page includes a QR code encoding product item information and a payment amount due.
126. The apparatus of embodiment 125, wherein the purchase transaction is initiated upon the user snapping the QR code using the user mobile device, and submitting a wallet payment request to an electronic payment processing network.
127. The apparatus of embodiment 125, wherein the purchase transaction is initiated at the merchant store.
128. The apparatus of embodiment 125, wherein the electronic receipt is sent to the user mobile device via a third party notification system.
129. The apparatus of embodiment 125, wherein the electronic receipt is provided by the merchant store.
130. The apparatus of embodiment 32, further comprising instructions to:
maintain a shopping cart for the user; and
add the product item to the shopping cart.
131. The apparatus of embodiment 32, further comprising instructions to:
receive a shopping list from the user mobile device; and
obtain product item information from the shopping list.
132. The apparatus of embodiment 32, further comprising instructions to:
obtain inventory information and stock keeping unit (SKU) information of the obtained product information; and
generate a store map with tags indicating locations of product items on the shopping list.
133. The apparatus of embodiment 32, further comprising instructions to:
generate an augmenter reality in-store scan indicating locations of product items on the shopping list.
134. The medium of embodiment 33, wherein the user check-in message is generated by a user snapping a merchant store provided quick response (QR) code.
135. The medium of embodiment 33, wherein the user check-in message is sent to a remote server.
136. The medium of embodiment 33, wherein the user check-in message includes geo-location information of the user.
137. The medium of embodiment 33, wherein the merchant store assigns a sales clerk to the user upon user check-in at the merchant store.
138. The medium of embodiment 137, wherein the sales clerk comprises any of a store employee and a virtual shopping assistant.
139. The medium of embodiment 137, wherein the sales clerk assignment is determined based on user loyalty levels.
140. The medium of embodiment 137, wherein the sales clerk comprises any of a local representative and a remote representative.
141. The medium of embodiment 33, wherein the user profile comprises user loyalty information and past purchasing history with the merchant store.
142. The medium of embodiment 33, wherein the user profile is previously stored at a local database at the merchant store.
143. The medium of embodiment 33, wherein the user profile is stored at a remote server and transmitted to the merchant store.
144. The medium of embodiment 33, wherein the real-time in-store behavior data comprises any of:
145. The medium of embodiment 33, wherein the product purchase recommendation comprises any of:
146. The medium of embodiment 145, further comprising instructions to:
147. The medium of embodiment 33, further comprising instructions to:
receive a user communication indicating shopping interests.
148. The medium of embodiment 147, wherein the user communication is conducted via any of:
in-person communication between the user and a sales clerk;
video chat;
audio chat;
instant messages; and
text messages.
149. The medium of embodiment 147, wherein the shopping interests further comprises:
150. The medium of embodiment 147, wherein the shopping interests further comprises:
151. The medium of embodiment 147, wherein the shopping interests further comprises:
152. The medium of embodiment 147, wherein the shopping interests further comprises:
153. The medium of embodiment 33, wherein check-out information page includes a QR code encoding product item information and a payment amount due.
154. The medium of embodiment 153, wherein the purchase transaction is initiated upon the user snapping the QR code using the user mobile device, and submitting a wallet payment request to an electronic payment processing network.
155. The medium of embodiment 153, wherein the purchase transaction is initiated at the merchant store.
156. The medium of embodiment 153, wherein the electronic receipt is sent to the user mobile device via a third party notification system.
157. The medium of embodiment 153, wherein the electronic receipt is provided by the merchant store.
158. The medium of embodiment 33, further comprising instructions to:
maintain a shopping cart for the user; and
add the product item to the shopping cart.
159. The medium of embodiment 33, further comprising instructions to:
receive a shopping list from the user mobile device; and
obtain product item information from the shopping list.
160. The medium of embodiment 33, further comprising instructions to:
obtain inventory information and stock keeping unit (SKU) information of the obtained product information; and
generate a store map with tags indicating locations of product items on the shopping list.
161. The medium of embodiment 33, further comprising instructions to:
generate an augmenter reality in-store scan indicating locations of product items on the shopping list.
162. A payment transaction visual capturing apparatus, comprising:
a processor; and
a memory disposed in communication with the processor and storing processor-executable instructions to:
163. A payment transaction visual capturing system, comprising:
164. A payment transaction visual capturing non-transitory computer-readable medium storing processor-executable instructions, said instructions executable by a processor to:
165. The apparatus of embodiment 162, wherein the identified object comprises any of a payment card, an invoice and a purchase item.
166. The apparatus of embodiment 162, wherein the user interactive option labels comprise any of the labels for fund transfer, view balance, and pay for a purchase.
167. The system of embodiment 163, wherein the identified object comprises any of a payment card, an invoice and a purchase item.
168. The system of embodiment 163, wherein the user interactive option labels comprise any of the labels for fund transfer, view balance, and pay for a purchase.
169. The medium of embodiment 164, wherein the identified object comprises any of a payment card, an invoice and a purchase item.
170. The medium of embodiment 164, wherein the user interactive option labels comprise any of the labels for fund transfer, view balance, and pay for a purchase.
171. A payment transaction visual capturing system, comprising:
172. A payment transaction visual capturing apparatus, comprising:
a processor; and
a memory disposed in communication with the processor and storing processor-executable instructions to:
173. A payment transaction visual capturing non-transitory computer-readable medium storing processor-executable instructions, said instructions executable by a processor to:
174. The system of embodiment 171, wherein the user activity records include any of a web search key term, a GPS location check-in event, and a browsing history.
175. The system of embodiment 171, wherein two or more objects are identified from the captured reality scene, and each of the two or more objects is associated with augmented reality virtual labels.
176. The system of embodiment 171, further comprising:
means for determining a fingertip motion within the captured reality scene.
177. The apparatus of embodiment 172, wherein the user activity records include any of a web search key term, a GPS location check-in event, and a browsing history.
178. The apparatus of embodiment 172, wherein two or more objects are identified from the captured reality scene, and each of the two or more objects is associated with augmented reality virtual labels.
179. The apparatus of embodiment 172, further comprising instructions to:
determine a fingertip motion within the captured reality scene.
180. The medium of embodiment 173, wherein the user activity records include any of a web search key term, a GPS location check-in event, and a browsing history.
181. The medium of embodiment 173, wherein two or more objects are identified from the captured reality scene, and each of the two or more objects is associated with augmented reality virtual labels.
182. The medium of embodiment 173, further comprising instructions to:
determine a fingertip motion within the captured reality scene.
183. A transaction visual capturing system, comprising:
184. A transaction visual capturing apparatus, comprising:
a processor; and
a memory disposed in communication with the processor and storing processor-executable instructions to:
185. A transaction visual capturing non-transitory computer-readable medium storing processor-executable instructions, said instructions executable by a processor to:
186. The system of embodiment 183, wherein the identified first object is a financial payment card having an account resolvable identifier.
187. The system of embodiment 183, wherein the identified second object is a financial payment card having an account resolvable identifier.
188. The system of embodiment 183, wherein the identified second object is a sales bill including a QR code.
189. The system of embodiment 183, wherein the identified second object is a metro card.
190. The system of embodiment 183, wherein the payment from the first financial account to the second financial account comprises a fund transfer from one financial payment card to another financial payment card.
191. The system of embodiment 183, wherein the payment from the first financial account to the second financial account comprises a bill payment from a financial payment card to a merchant for a product purchase.
192. The system of embodiment 183, wherein the payment from the first financial account to the second financial account comprises a fund refill from a financial payment card to a metro card.
193. The system of embodiment 183, wherein the image processing comprises obtaining screen grabs of the obtained live visual capture.
194. The system of embodiment 183, wherein the user transaction command comprises an audio command.
195. The system of embodiment 183, wherein the user transaction command comprises a fingertip motion of moving from the first object to the second object.
196. The system of embodiment 183, further comprising:
means for obtaining information pertaining to the identified first financial account and the identified second object based on the determined first account identifier.
197. The system of embodiment 183, further comprising:
198. The system of embodiment 183, wherein the first account identifier and the second account identifier are visibly determinable via any of:
barcode reading;
QR code decoding; and
optical character recognition (OCR).
199. The system of embodiment 183, further comprising:
means for obtaining authorization credentials for the payment from the first financial account to the second financial account.
200. The system of embodiment 199, further comprising:
means for requesting a user to input a passcode for user identify confirmation.
201. The system of embodiment 183, wherein the first account identifier comprises a 16 digit bank card number.
202. The system of embodiment 183, wherein the second account identifier comprises a merchant identifier.
203. The system of embodiment 183, wherein the second account identifier comprises a 16 digit bank card number.
204. The system of embodiment 183, further comprising:
means for generating a security alert request when the second object comprises a financial payment card with a cardholder; and
means for sending the security alert to the cardholder of the second object.
205. The apparatus of embodiment 184, wherein the identified first object is a financial payment card having an account resolvable identifier.
206. The apparatus of embodiment 184, wherein the identified second object is a financial payment card having an account resolvable identifier.
207. The apparatus of embodiment 184, wherein the identified second object is a sales bill including a QR code.
208. The apparatus of embodiment 184, wherein the identified second object is a metro card.
209. The apparatus of embodiment 184, wherein the payment from the first financial account to the second financial account comprises a fund transfer from one financial payment card to another financial payment card.
210. The apparatus of embodiment 184, wherein the payment from the first financial account to the second financial account comprises a bill payment from a financial payment card to a merchant for a product purchase.
211. The apparatus of embodiment 184, wherein the payment from the first financial account to the second financial account comprises a fund refill from a financial payment card to a metro card.
212. The apparatus of embodiment 184, wherein the image processing comprises obtaining screen grabs of the obtained live visual capture.
213. The apparatus of embodiment 184, wherein the user transaction command comprises an audio command.
214. The apparatus of embodiment 184, wherein the user transaction command comprises a fingertip motion of moving from the first object to the second object.
215. The apparatus of embodiment 184, further comprising instructions to: obtain information pertaining to the identified first financial account and the identified second object based on the determined first account identifier.
216. The apparatus of embodiment 184, further comprising instructions to:
217. The apparatus of embodiment 184, wherein the first account identifier and the second account identifier are visibly determinable via any of:
barcode reading;
QR code decoding; and
optical character recognition (OCR).
218. The apparatus of embodiment 184, further comprising instructions to:
obtain authorization credentials for the payment from the first financial account to the second financial account.
219. The apparatus of embodiment 218, further comprising instructions to:
request a user to input a passcode for user identify confirmation.
220. The apparatus of embodiment 184, wherein the first account identifier comprises a 16 digit bank card number.
221. The apparatus of embodiment 184, wherein the second account identifier comprises a merchant identifier.
222. The apparatus of embodiment 184, wherein the second account identifier comprises a 16 digit bank card number.
223. The apparatus of embodiment 184, further comprising instructions to:
generate a security alert request when the second object comprises a financial payment card with a cardholder; and
send the security alert to the cardholder of the second object.
224. The medium of embodiment 185, wherein the identified first object is a financial payment card having an account resolvable identifier.
225. The medium of embodiment 185, wherein the identified second object is a financial payment card having an account resolvable identifier.
226. The medium of embodiment 185, wherein the identified second object is a sales bill including a QR code.
227. The medium of embodiment 185, wherein the identified second object is a metro card.
228. The medium of embodiment 185, wherein the payment from the first financial account to the second financial account comprises a fund transfer from one financial payment card to another financial payment card.
229. The medium of embodiment 185, wherein the payment from the first financial account to the second financial account comprises a bill payment from a financial payment card to a merchant for a product purchase.
230. The medium of embodiment 185, wherein the payment from the first financial account to the second financial account comprises a fund refill from a financial payment card to a metro card.
231. The medium of embodiment 185, wherein the image processing comprises obtaining screen grabs of the obtained live visual capture.
232. The medium of embodiment 185, wherein the user transaction command comprises an audio command.
233. The medium of embodiment 185, wherein the user transaction command comprises a fingertip motion of moving from the first object to the second object.
234. The medium of embodiment 185, further comprising instructions to:
obtain information pertaining to the identified first financial account and the identified second object based on the determined first account identifier.
235. The medium of embodiment 185, further comprising instructions to:
236. The medium of embodiment 185, wherein the first account identifier and the second account identifier are visibly determinable via any of:
barcode reading;
QR code decoding; and
optical character recognition (OCR).
237. The medium of embodiment 185, further comprising:
obtain authorization credentials for the payment from the first financial account to the second financial account.
238. The medium of embodiment 237, further comprising instructions to:
request a user to input a passcode for user identify confirmation.
239. The medium of embodiment 185, wherein the first account identifier comprises a 16 digit bank card number.
240. The medium of embodiment 185, wherein the second account identifier comprises a merchant identifier.
241. The medium of embodiment 185, wherein the second account identifier comprises a 16 digit bank card number.
242. The medium of embodiment 185, further comprising instructions to:
generate a security alert request when the second object comprises a financial payment card with a cardholder; and
send the security alert to the cardholder of the second object.
243. A visual capturing system, comprising:
means for obtaining a list of product items indicating user demands at a user mobile device;
means for determining a product category and a product identifier for each product item on the obtained list of product items;
means for obtaining a user indication of a merchant store;
obtaining product inventory and stock keeping data of the merchant store;
means for querying the obtained product inventory and stock keeping data based on the product identifier and the product category for each product item;
means for determining an in-store stock keeping location for each product item based on the query;
means for obtaining a visual layout of the merchant store;
means for tagging the visual layout of the merchant store with the determined in-store stock keeping location for each product item; and
means for presenting the tagged visual layout of the merchant store at the user mobile device.
244. A visual capturing apparatus, comprising:
a processor; and
a memory disposed in communication with the processor and storing processor-executable instructions to:
245. A visual capturing non-transitory computer-readable medium storing processor-executable instructions, said instructions executable by a processor to:
obtain a list of product items indicating user demands at a user mobile device;
determine a product category and a product identifier for each product item on the obtained list of product items;
obtain a user indication of a merchant store;
obtain product inventory and stock keeping data of the merchant store;
query the obtained product inventory and stock keeping data based on the product identifier and the product category for each product item;
determine an in-store stock keeping location for each product item based on the query;
obtain a visual layout of the merchant store;
tag the visual layout of the merchant store with the determined in-store stock keeping location for each product item; and
present the tagged visual layout of the merchant store at the user mobile device.
246. The system of embodiment 243, wherein the list of product items comprises a shopping list entered by a user.
247. The system of embodiment 246, wherein the shopping list is generated via audio commands from the user.
248. The system of embodiment 246, wherein the shopping list is generated by extracting product item information from a previously stored sales receipt.
249. The system of embodiment 243, wherein the user indication of the merchant store comprises a user check-in message at a merchant store.
250. The system of embodiment 243, wherein the user indication of the merchant store comprises GPS coordinates of a user.
251. The system of embodiment 243, wherein the product inventory and stock keeping data comprises a table listing an aisle number and a stack number of an in-stock product at the merchant store.
252. The system of embodiment 243, wherein the in-store stock keeping location for each product item comprises any of a aisle number, a stack number, and a shelf number.
253. The system of embodiment 243, wherein the visual layout of the merchant store comprises a static store floor plan map.
254. The system of embodiment 253, further comprising highlighting the static store floor plan map with labels illustrating a location of each product item.
255. The system of embodiment 243, wherein the visual layout of the merchant store comprises a live visual capture of an in-store reality scene.
256. The system of embodiment 255, further comprising injecting user interactive augmented reality labels overlaying the live visual capture of the in-store reality scene, said augmented reality labels indicating a location of each product item within the in-store reality scene.
257. The system of embodiment 256, wherein said augmented reality labels may comprise a semi-transparent bound box covering a product item within the in-store reality scene.
258. The system of embodiment 243, wherein more than one merchant stores are processed for multi-merchant shopping.
259. The apparatus of embodiment 244, wherein the list of product items comprises a shopping list entered by a user.
260. The apparatus of embodiment 259, wherein the shopping list is generated via audio commands from the user.
261. The apparatus of embodiment 259, wherein the shopping list is generated by extracting product item information from a previously stored sales receipt.
262. The apparatus of embodiment 244, wherein the user indication of the merchant store comprises a user check-in message at a merchant store.
263. The apparatus of embodiment 244, wherein the user indication of the merchant store comprises GPS coordinates of a user.
264. The apparatus of embodiment 244, wherein the product inventory and stock keeping data comprises a table listing an aisle number and a stack number of an in-stock product at the merchant store.
265. The apparatus of embodiment 244, wherein the in-store stock keeping location for each product item comprises any of a aisle number, a stack number, and a shelf number.
266. The apparatus of embodiment 244, wherein the visual layout of the merchant store comprises a static store floor plan map.
267. The apparatus of embodiment 266, further comprising highlighting the static store floor plan map with labels illustrating a location of each product item.
268. The apparatus of embodiment 244, wherein the visual layout of the merchant store comprises a live visual capture of an in-store reality scene.
269. The apparatus of embodiment 268, further comprising injecting user interactive augmented reality labels overlaying the live visual capture of the in-store reality scene, said augmented reality labels indicating a location of each product item within the in-store reality scene.
270. The apparatus of embodiment 269, wherein said augmented reality labels may comprise a semi-transparent bound box covering a product item within the in-store reality scene.
271. The apparatus of embodiment 244, wherein more than one merchant stores are processed for multi-merchant shopping.
272. The medium of embodiment 245, wherein the list of product items comprises a shopping list entered by a user.
273. The medium of embodiment 272, wherein the shopping list is generated via audio commands from the user.
274. The medium of embodiment 272, wherein the shopping list is generated by extracting product item information from a previously stored sales receipt.
275. The medium of embodiment 245, wherein the user indication of the merchant store comprises a user check-in message at a merchant store.
276. The medium of embodiment 245, wherein the user indication of the merchant store comprises GPS coordinates of a user.
277. The medium of embodiment 245, wherein the product inventory and stock keeping data comprises a table listing an aisle number and a stack number of an in-stock product at the merchant store.
278. The medium of embodiment 245, wherein the in-store stock keeping location for each product item comprises any of a aisle number, a stack number, and a shelf number.
279. The medium of embodiment 245, wherein the visual layout of the merchant store comprises a static store floor plan map.
280. The medium of embodiment 279, further comprising highlighting the static store floor plan map with labels illustrating a location of each product item.
281. The medium of embodiment 245, wherein the visual layout of the merchant store comprises a live visual capture of an in-store reality scene.
282. The medium of embodiment 281, further comprising injecting user interactive augmented reality labels overlaying the live visual capture of the in-store reality scene, said augmented reality labels indicating a location of each product item within the in-store reality scene.
283. The medium of embodiment 282, wherein said augmented reality labels may comprise a semi-transparent bound box covering a product item within the in-store reality scene.
284. The medium of embodiment 245, wherein more than one merchant stores are processed for multi-merchant shopping.
285. A processor-implemented method comprising:
receiving from a wallet user multiple gesture actions within a specified temporal quantum;
determining composite constituent gestures, gesture manipulated objects, and user account information from the received multiple gesture actions;
determining via a processor a composite gesture action associated with the determined composite constituent gestures and gesture manipulated objects; and
executing via a processor the composite gesture action to perform a transaction with a user account specified by the user account information.
286. The method of embodiment 285, wherein the multiple gesture actions contain a video file.
287. The method of embodiment 285, wherein the multiple gesture actions contain at least one image file.
288. The method of embodiment 285, wherein the wherein the multiple gesture actions contain an audio file.
289. The method of embodiment 285, wherein the multiple gesture actions contain both at least one image file and an audio file.
290. The method of embodiment 285, wherein the transaction is a payment transaction between the user and a second entity.
291. The method of embodiment 285, wherein the transaction is a payment transaction between the user's payment device and second payment device also owned by the user.
292. An apparatus comprising:
a processor; and
a memory disposed in communication with the processor and storing processor-issuable instructions to:
293. A system comprising:
means to receive from a wallet user multiple gesture actions within a specified temporal quantum;
means to determine composite constituent gestures, gesture manipulated objects, and user account information from the received multiple gesture actions;
means to determine a composite gesture action associated with the determined composite constituent gestures and gesture manipulated objects; and
means to execute the composite gesture action to perform a transaction with a user account specified by the user account information.
294. A processor-readable tangible medium storing processor-issuable instructions to:
This application claims priority under 35 U.S.C. §119 to U.S. provisional patent application Ser. No. 61/583,378 filed Jan. 5, 2012, attorney docket no. 196US01/VISA-177/00US, U.S. provisional patent application Ser. No. 61/594,957, filed Feb. 3, 2012, attorney docket no. 196US02|VISA-177/01US, and U.S. provisional patent application Ser. No. 61/620,365, filed Apr. 4, 2012, attorney docket no. 196US03|VISA-177/02US, all entitled “Augmented Retail Shopping Apparatuses, Methods and Systems.” This application claims priority under 35 USC §119 to U.S. provisional patent application Ser. No. 61/625,170, filed Apr. 17, 2012, attorney docket no. 268US01|VISA-189/00US, entitled “Payment Transaction Visual Capturing Apparatuses, Methods And Systems”; and U.S. provisional patent application Ser. No. 61/749,202, filed Jan. 4, 2013, attorney docket no. 316US01|VISA-196/00US, and entitled “Multi Disparate Gesture Actions And Transactions Apparatuses, Methods And Systems.” This application claims priority under 35 USC §§120, 365 to U.S. non-provisional patent application Ser. No. 13/434,818 filed Mar. 29, 2012 and titled “Graduated Security Seasoning Apparatuses, Methods and Systems”; and PCT international application serial no. PCT/US12/66898, filed Nov. 28, 2012, entitled “Transaction Security Graduated Seasoning And Risk Shifting Apparatuses, Methods And Systems.” This application also claims priority under 35 USC §§120, 365 to PCT International Application Serial No. PCT/US13/20411, filed Jan. 5, 2013, attorney docket no. 196W001|VISA-17/01WO, entitled “TRANSACTION VISUAL CAPTURING Apparatuses, Methods And Systems.” The aforementioned applications are all hereby expressly incorporated by reference. This application incorporates by reference, the entire contents of the following applications: (1) U.S. non-provisional patent application Ser. No. 13/327,740 filed on Dec. 15, 2011 and titled “Social Media Payment Platform Apparatuses, Methods and Systems.”
Number | Date | Country | |
---|---|---|---|
61583378 | Jan 2012 | US | |
61594957 | Feb 2012 | US | |
61620365 | Apr 2012 | US | |
61625170 | Apr 2012 | US | |
61749202 | Jan 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13434818 | Mar 2012 | US |
Child | 13735802 | US | |
Parent | PCT/US12/66898 | Nov 2012 | US |
Child | 13434818 | US | |
Parent | PCT/US13/20411 | Jan 2013 | US |
Child | PCT/US12/66898 | US |