This application for letters patent disclosure document describes inventive aspects that include various novel innovations (hereinafter “disclosure”) and contains material that is subject to copyright, mask work, and/or other intellectual property protection. The respective owners of such intellectual property have no objection to the facsimile reproduction of the disclosure by anyone as it appears in published Patent Office file/records, but otherwise reserve all rights.
The present innovations generally address gesture command analysis, and more particularly, include GESTURE RECOGNITION CLOUD COMMAND APPARATUSES, METHODS AND SYSTEMS (GRCCT).
However, in order to develop a reader's understanding of the innovations, disclosures have been compiled into a single description to illustrate and clarify how aspects of these innovations operate independently, interoperate as between individual innovations, and/or cooperate collectively. The application goes on to further describe the interrelations and synergies as between the various innovations; all of which is to further compliance with 35 U.S.C. § 112.
Consumers visiting brick and mortar stores (i.e., point of sales) typically have limited options for communicating with the stores and requesting actions to be performed (e.g., checking inventory or making a purchase). The available options typically are to speak directly with a live agent, such as a cashier or sales person, or to interact through a kiosk, such as a price-check machine or self-checkout terminal. In both cases, the consumer must first locate the live agent or kiosk, approach him/it, and only begin to communicate if he/it is unoccupied. While consumers may also use their mobile devices to interact with the store's online presence (e.g., via its website or app), the virtual interaction is typically not integrated with the consumer's in-store shopping experience. Moreover, the user interface afforded by mobile devices is limiting. Therefore, is an increased demand to streamline communication and command execution at the point of sales.
Processor-implemented systems and methods are described herein are for transmitting a command to a remote system. A processing system determines the identity of the user based on the unique identifier and the biometric information. Thereafter, a sensor detects a gesture performed by the user. The sensor is configured to detect the gesture performed by the user when the user is located within the detectable range of the wireless antenna. The processing system determines an action associated with the detected gesture based on the identity of the user and sends a command to a remote computer system to cause it to perform the action associated with the detected gesture.
As another example, processor-implemented systems and methods are disclosed for transmitting a command to a remote system wherein a first sensor detects a unique identifier for a user device that is associated with a user. When the user is within a detectable range of a wireless antenna, a second sensor detects biometric information for the user. A processing system determines the identity of the user based on the unique identifier and the biometric information. Thereafter, a third sensor detects a gesture performed by the user. The third sensor is configured to detect the gesture performed by the user when the user is located within the detectable range of the wireless antenna. The processing system determines an action associated with the detected gesture based on the identity of the user and sends a command to a remote computer system to cause it to perform the action associated with the detected gesture.
The accompanying appendices and/or drawings illustrate various non-limiting, example, innovative aspects in accordance with the present descriptions:
The accompanying appendices and/or drawings illustrate various non-limiting, example, inventive aspects in accordance with the present disclosure:
The detected gesture and the detected voice command are provided to a second entity, where the user has an account with the second entity. An action associated with the detected gesture and the detected voice command is determined. (
A sensor detects a first gesture that is performed by the user, where the first gesture is directed to an item that is included in the merchant store. The first gesture is detected after the providing of the check-in information to the merchant store. (
The sensor detects a second gesture that is performed by the user, where the second gesture is detected after the performing of the action associated with the detected first gesture. (
Based on the subset of data, an augmented reality display is generated and viewed by a user. The user is associated with the subset of data, and the user uses the visual device to obtain the visual capture. (
The visual device generates an augmented reality display that includes i) the image of the customer, and ii) additional image data that surrounds the image of the customer. The augmented reality display is viewed by personnel of the merchant store. (
The visual device generates an augmented reality display that includes a user interactive area, where the user interactive area is associated with the itemized expense. (
An image of a user is stored at the visual device, where the visual device is operated by the user or worn by the user. (
The detected gesture is provided to the visual device. An action associated with the gesture is determined, and the action is performed at the visual device. The performing of the action updates the interactive display based on the image of the first item or the image of the second item. The updating of the interactive display causes the image of the user to be modified based on the image of the first item or the image of the second item. (
An augmented reality display is generated at the visual device. The augmented reality display includes i) the image of the item sold by the merchant store, and ii) additional image data that surrounds the image of the item. (
An interactive display is generated at the visual device. The interactive display includes a user interactive area and a second image of the item. A gesture performed by a user is detected via a sensor, where the gesture is directed to the user interactive area of the interactive display. (
In some implementations, an exemplary XML-encoded command message 208 may take a form similar to the following:
In some implementations, the electronic device may reduce the size of the vocal file by cropping the audio file to when the user begins and ends the vocal command. In some implementations, the MDGAAT may process the gesture and audio data 210 in order to determine the type of gesture performed, as well as the words spoken by the user. In some implementations, a composite gesture generated from the processing of the gesture and audio data may be embodied in an XML-encoded data structure similar to the following:
In some implementations, fields in the composite gesture data structure may be left blank depending on whether the particular gesture type (e.g., finger gesture, object gesture, and/or the like) has been made. The MDGAAT may then match 211 the gesture and the words to the various possible gesture types stored in the MDGAAT database. In some implementations, the MGDAAT may query the database for particular disparate gestures in a manner similar to the following:
In some implementations, the result of each query in the above example may be used to search for the composite gesture in the Multi-Disparate Gesture Action (MDGA) table of the database. For example, if $fingerresult is “tap check,” $objectresult is “swipe,” and $voiceresult is “pay total of check with this payment device,” MDGAAT may search the MDGA table using these three results to narrow down the precise composite action that has been performed. If a match is found, the MDGAAT may request confirmation that the right action was found, and then may perform the action 212 using the user's account. In some implementations, the MDGAAT may access the user's financial information and account 213 in order to perform the action. In some implementations, MDGAAT may update a gesture table 214 in the MDGAAT database 215 to refine models for usable gestures based on the user's input, to add new gestures the user has invented, and/or the like. In some implementations, an update 214 for a finger gesture may be performed via a PHP/MySQL command similar to the following:
After successfully updating the table 216, the MDGAAT may send the user to a confirmation page 217 (or may provide an augmented reality (AR) overlay to the user) which may indicate that the action was successfully performed. In some implementations, the AR overlay may be provided to the user through use of smart glasses, contacts, and/or a like device (e.g. Google Glasses).
As shown in
The MDGAAT may then perform the action specified 221, accessing any information necessary to conduct the action 222, and may send a confirmation page or AR overlay to the user 223. In some implementations, the XML-encoded data structure for the AR overlay may take a form similar to the following:
In some implementations, MDGAAT may receive 305 the data from the user's device, which may include accelerometer and/or gyroscope data pertaining to the gesture, a video and/or images of the gesture, an audio file of the vocal command, and/or the like. In some implementations, MDGAAT may determine what sort of data was sent by the user's device in order to determine how to process it. For example, if the user's device provides accelerometer and/or gyroscope data 306, MDGAAT may determine the gesture performed by matching the accelerometer and/or gyroscope data points with pre-determined mathematical gesture models 309. For example, if a particular gesture would generate accelerometer and/or gyroscope data that would fit a linear gesture model, MDGAAT will determine whether the received accelerometer and/or gyroscope data matches a linear model.
If the user's device provides a video and/or images of the gesture 307, MDGAAT may use an image processing component in order to process the video and/or images 310 and determine what the gesture is. In some implementations, if a video is provided, the video may also be used to determine the vocal command provided by the user. As shown in
If the user's device provides an audio file 308, then MDGAAT may determine the vocal command given using an audio analytics component 311. In some implementations, the audio analytics component may process the audio file and produce a text translation of the vocal command. As discussed above, in some implementations, the audio analytics component may also use a video, if provided, as input to produce a text translation of the user's vocal command.
As shown in
If the action is a single-party payment-related action 318 (i.e., concerning one person and/or entity transferring funds to his/her/itself), MDGAAT may retrieve the account information of the one user 319, and may use it to access the relevant financial and/or other accounts associated in the transaction. For example, if one user is transferring funds from a bank account to a refillable gift card owned by the same user, then MDGAAT would access the user's account in order to obtain information about both the bank account and the gift card, and would use the information to transfer funds from the bank account to the gift card 320.
In either the multi-party or the single-party action, MDGAAT may update the data of the affected accounts (including: saving a record of the transaction, which may include to whom the money was given to, the date and time of the transaction, the size of the transaction, and/or the like), and may send a confirmation of this update 322 to the user.
If the action is related to obtaining information about a product and/or service 323, MDGAAT may send a request 324 to the relevant merchant database(s) in order to get information about the product and/or service the user would like to know more about. MDGAAT may provide any information obtained from the merchant to the user 325. In some implementations, MDGAAT may provide the information via an AR overlay, or via an information page or pop-up which displays all the retrieved information.
In some implementations, the user, while shopping through the store, may also scan 407 items with the user's electronic device, in order to obtain more information about them, in order to add them to the user's cart, and/or the like. In such implementations, the user's electronic device may send a scanned item message 408 to the MDGAAT server. In some implementations, an exemplary XML-encoded scanned item message 408 may take a form similar to the following:
In some implementations, MDGAAT may then determine the location 409 of the user based on the location of the scanned item, and may send a notification 410 to a sale's representative 411 indicating that a user has checked into the store and is browsing items in the store. In some implementations, an exemplary XML-encoded notification message 410 may comprise of the scanned item message of scanned item message 408.
The sale's representative may use the information in the notification message to determine products and/or services to recommend 412 to the user, based on the user's profile, location in the store, items scanned, and/or the like. Once the sale's representative has chosen at least one product and/or service to suggest, it may send the suggestion 413 to the MDGAAT server. In some implementations, an exemplary XML-encoded suggestion 413 may take a form similar to the following:
In some implementations, MDGAAT may conduct its own analysis of the user based on the photo 427, including analyzing the image to determine the user's body size, body shape, complexion, and/or the like. In some implementations, MDGAAT may use these attributes, along with any provided through the apparel preview request, to search the database 428 for clothing that matches the user's attributes and search criteria. In some implementations, MDGAAT may also update 429 the user's attributes stored in the database, based on the attributes provided in the apparel preview request or based on MDGAAT′ analysis of the user's photo. After MDGAAT receives confirmation that the update is successful 430, MDGAAT may send a virtual closet 431 to the user, comprising a user interface for previewing clothing, accessories, and/or the like chosen for the user based on the user's attributes and search criteria. In some implementations, the virtual closet may be implemented via HTML and Javascript.
In some implementations, as shown in
In some implementations, the user may provide a selection of at least one article of clothing to try on 521, prompting MDGAAT to determine body and/or joint locations and markers in the user photo 522, and to scale the image of the article of clothing to match the user image 523, based on those body and/or joint locations and markers. In some implementations, MDGAAT may also format the clothing image 524, including altering shadows in the image, blurring the image, and/or the like, in order to match the look of the clothing image to the look of the user image. MDGAAT may superimpose 525 the clothing image on the user image to allow the user to virtually preview the article of clothing on the user, and may allow the user to change options such as the clothing color, size, and/or the like while the article of clothing is being previewed on the user. In some implementations, MDGAAT may receive a request to purchase at least one article of clothing 526, and may retrieve user information 527, including the user's ID, shipping address, and/or the like. MDGAAT may further retrieve the user's payment information 528, including the user's preferred payment device or account, and/or the like, and may contact the user's issuer (and that of the merchant) 529 in order to process the transaction. MDGAAT may send a confirmation to the user when the transaction is completed 530.
As shown in
As shown in
Typically, users, e.g., 1133a, which may be people and/or other systems, may engage information technology systems (e.g., computers) to facilitate information processing. In turn, computers employ processors to process information; such processors 1103 may be referred to as central processing units (CPU). One form of processor is referred to as a microprocessor. CPUs use communicative circuits to pass binary encoded signals acting as instructions to enable various operations. These instructions may be operational and/or data instructions containing and/or referencing other instructions and data in various processor accessible and operable areas of memory 1129 (e.g., registers, cache memory, random access memory, etc.). Such communicative instructions may be stored and/or transmitted in batches (e.g., batches of instructions) as programs and/or data components to facilitate desired operations. These stored instruction codes, e.g., programs, may engage the CPU circuit components and other motherboard and/or system components to perform desired operations. One type of program is a computer operating system, which, may be executed by CPU on a computer; the operating system enables and facilitates users to access and operate computer information technology and resources. Some resources that may be employed in information technology systems include: input and output mechanisms through which data may pass into and out of a computer; memory storage into which data may be saved; and processors by which information may be processed. These information technology systems may be used to collect data for later retrieval, analysis, and manipulation, which may be facilitated through a database program. These information technology systems provide interfaces that allow users to access and operate various system components.
In one embodiment, the MDGAAT controller 1101 may be connected to and/or communicate with entities such as, but not limited to: one or more users from user input devices 1111; peripheral devices 112; an optional cryptographic processor device 1128; and/or a communications network 1113. For example, the MDGAAT controller oi may be connected to and/or communicate with users, e.g., 1133a, operating client device(s), e.g., 1133b, including, but not limited to, personal computer(s), server(s) and/or various mobile device(s) including, but not limited to, cellular telephone(s), smartphone(s) (e.g., iPhone®, Blackberry®, Android OS-based phones etc.), tablet computer(s) (e.g., Apple iPad™, HP Slate™, Motorola Xoom™, etc.), eBook reader(s) (e.g., Amazon Kindle™, Barnes and Noble's Nook™ eReader, etc.), laptop computer(s), notebook(s), netbook(s), gaming console(s) (e.g., XBOX Live™, Nintendo® DS, Sony PlayStation® Portable, etc.), portable scanner(s), and/or the like.
Networks are commonly thought to comprise the interconnection and interoperation of clients, servers, and intermediary nodes in a graph topology. It should be noted that the term “server” as used throughout this application refers generally to a computer, other device, program, or combination thereof that processes and responds to the requests of remote users across a communications network. Servers serve their information to requesting “clients.” The term “client” as used herein refers generally to a computer, program, other device, user and/or combination thereof that is capable of processing and making requests and obtaining and processing any responses from servers across a communications network. A computer, other device, program, or combination thereof that facilitates, processes information and requests, and/or furthers the passage of information from a source user to a destination user is commonly referred to as a “node.” Networks are generally thought to facilitate the transfer of information from source points to destinations. A node specifically tasked with furthering the passage of information from a source to a destination is commonly called a “router.” There are many forms of networks such as Local Area Networks (LANs), Pico networks, Wide Area Networks (WANs), Wireless Networks (WLANs), etc. For example, the Internet is generally accepted as being an interconnection of a multitude of networks whereby remote clients and servers may access and interoperate with one another.
The MDGAAT controller 1101 may be based on computer systems that may comprise, but are not limited to, components such as: a computer systemization 1102 connected to memory 1129.
A computer systemization 1102 may comprise a clock 1130, central processing unit (“CPU(s)” and/or “processor(s)” (these terms are used interchangeable throughout the disclosure unless noted to the contrary)) 1103, a memory 1129 (e.g., a read only memory (ROM) 1106, a random access memory (RAM) 1105, etc.), and/or an interface bus 1107, and most frequently, although not necessarily, are all interconnected and/or communicating through a system bus 1104 on one or more (mother)board(s) 1102 having conductive and/or otherwise transportive circuit pathways through which instructions (e.g., binary encoded signals) may travel to effectuate communications, operations, storage, etc. The computer systemization may be connected to a power source 1186; e.g., optionally the power source may be internal. Optionally, a cryptographic processor 1126 and/or transceivers (e.g., ICs) 1174 may be connected to the system bus. In another embodiment, the cryptographic processor and/or transceivers may be connected as either internal and/or external peripheral devices 1112 via the interface bus I/O. In turn, the transceivers may be connected to antenna(s) 1175, thereby effectuating wireless transmission and reception of various communication and/or sensor protocols; for example the antenna(s) may connect to: a Texas Instruments WiLink WL1283 transceiver chip (e.g., providing 802.11n, Bluetooth 3.0, FM, global positioning system (GPS) (thereby allowing MDGAAT controller to determine its location)); Broadcom BCM4329FKUBG transceiver chip (e.g., providing 802.11n, Bluetooth 2.1+EDR, FM, etc.); a Broadcom BCM4750IUB8 receiver chip (e.g., GPS); an Infineon Technologies X-Gold 618-PMB9800 (e.g., providing 2G/3G HSDPA/HSUPA communications); and/or the like. The system clock typically has a crystal oscillator and generates a base signal through the computer systemization's circuit pathways. The clock is typically coupled to the system bus and various clock multipliers that will increase or decrease the base operating frequency for other components interconnected in the computer systemization. The clock and various components in a computer systemization drive signals embodying information throughout the system. Such transmission and reception of instructions embodying information throughout a computer systemization may be commonly referred to as communications. These communicative instructions may further be transmitted, received, and the cause of return and/or reply communications beyond the instant computer systemization to: communications networks, input devices, other computer systemizations, peripheral devices, and/or the like. It should be understood that in alternative embodiments, any of the above components may be connected directly to one another, connected to the CPU, and/or organized in numerous variations employed as exemplified by various computer systems.
The CPU comprises at least one high-speed data processor adequate to execute program components for executing user and/or system-generated requests. Often, the processors themselves will incorporate various specialized processing units, such as, but not limited to: integrated system (bus) controllers, memory management control units, floating point units, and even specialized processing sub-units like graphics processing units, digital signal processing units, and/or the like. Additionally, processors may include internal fast access addressable memory, and be capable of mapping and addressing memory 1129 beyond the processor itself; internal memory may include, but is not limited to: fast registers, various levels of cache memory (e.g., level 1, 2, 3, etc.), RAM, etc. The processor may access this memory through the use of a memory address space that is accessible via instruction address, which the processor can construct and decode allowing it to access a circuit path to a specific memory address space having a memory state. The CPU may be a microprocessor such as: AMD's Athlon, Duron and/or Opteron; ARM's application, embedded and secure processors; IBM and/or Motorola's DragonBall and PowerPC; IBM's and Sony's Cell processor; Intel's Celeron, Core (2) Duo, Itanium, Pentium, Xeon, and/or XScale; and/or the like processor(s). The CPU interacts with memory through instruction passing through conductive and/or transportive conduits (e.g., (printed) electronic and/or optic circuits) to execute stored instructions (i.e., program code) according to conventional data processing techniques. Such instruction passing facilitates communication within the MDGAAT controller and beyond through various interfaces. Should processing requirements dictate a greater amount speed and/or capacity, distributed processors (e.g., Distributed MDGAAT), mainframe, multi-core, parallel, and/or super-computer architectures may similarly be employed. Alternatively, should deployment requirements dictate greater portability, smaller Personal Digital Assistants (PDAs) may be employed.
Depending on the particular implementation, features of the MDGAAT may be achieved by implementing a microcontroller such as CAST's R8051XC2 microcontroller; Intel's MCS 51 (i.e., 8051 microcontroller); and/or the like. Also, to implement certain features of the MDGAAT, some feature implementations may rely on embedded components, such as: Application-Specific Integrated Circuit (“ASIC”), Digital Signal Processing (“DSP”), Field Programmable Gate Array (“FPGA”), and/or the like embedded technology. For example, any of the MDGAAT component collection (distributed or otherwise) and/or features may be implemented via the microprocessor and/or via embedded components; e.g., via ASIC, coprocessor, DSP, FPGA, and/or the like. Alternately, some implementations of the MDGAAT may be implemented with embedded components that are configured and used to achieve a variety of features or signal processing.
Depending on the particular implementation, the embedded components may include software solutions, hardware solutions, and/or some combination of both hardware/software solutions. For example, MDGAAT features discussed herein may be achieved through implementing FPGAs, which are a semiconductor devices containing programmable logic components called “logic blocks”, and programmable interconnects, such as the high performance FPGA Virtex series and/or the low cost Spartan series manufactured by Xilinx. Logic blocks and interconnects can be programmed by the customer or designer, after the FPGA is manufactured, to implement any of the MDGAAT features. A hierarchy of programmable interconnects allow logic blocks to be interconnected as needed by the MDGAAT system designer/administrator, somewhat like a one-chip programmable breadboard. An FPGA's logic blocks can be programmed to perform the operation of basic logic gates such as AND, and XOR, or more complex combinational operators such as decoders or simple mathematical operations. In most FPGAs, the logic blocks also include memory elements, which may be circuit flip-flops or more complete blocks of memory. In some circumstances, the MDGAAT may be developed on regular FPGAs and then migrated into a fixed version that more resembles ASIC implementations. Alternate or coordinating implementations may migrate MDGAAT controller features to a final ASIC instead of or in addition to FPGAs. Depending on the implementation all of the aforementioned embedded components and microprocessors may be considered the “CPU” and/or “processor” for the MDGAAT.
The power source 1186 may be of any standard form for powering small electronic circuit board devices such as the following power cells: alkaline, lithium hydride, lithium ion, lithium polymer, nickel cadmium, solar cells, and/or the like. Other types of AC or DC power sources may be used as well. In the case of solar cells, in one embodiment, the case provides an aperture through which the solar cell may capture photonic energy. The power cell 1186 is connected to at least one of the interconnected subsequent components of the MDGAAT thereby providing an electric current to all subsequent components. In one example, the power source 1186 is connected to the system bus component 1104. In an alternative embodiment, an outside power source 1186 is provided through a connection across the I/O 1108 interface. For example, a USB and/or IEEE 1394 connection carries both data and power across the connection and is therefore a suitable source of power.
Interface bus(ses) 1107 may accept, connect, and/or communicate to a number of interface adapters, conventionally although not necessarily in the form of adapter cards, such as but not limited to: input output interfaces (I/O) 1108, storage interfaces 1109, network interfaces 1110, and/or the like. Optionally, cryptographic processor interfaces 1127 similarly may be connected to the interface bus. The interface bus provides for the communications of interface adapters with one another as well as with other components of the computer systemization. Interface adapters are adapted for a compatible interface bus. Interface adapters conventionally connect to the interface bus via a slot architecture. Conventional slot architectures may be employed, such as, but not limited to: Accelerated Graphics Port (AGP), Card Bus, (Extended) Industry Standard Architecture ((E)ISA), Micro Channel Architecture (MCA), NuBus, Peripheral Component Interconnect (Extended) (PCI(X)), PCI Express, Personal Computer Memory Card International Association (PCMCIA), and/or the like.
Storage interfaces 1109 may accept, communicate, and/or connect to a number of storage devices such as, but not limited to: storage devices 1114, removable disc devices, and/or the like. Storage interfaces may employ connection protocols such as, but not limited to: (Ultra) (Serial) Advanced Technology Attachment (Packet Interface) ((Ultra) (Serial) ATA(PI)), (Enhanced) Integrated Drive Electronics ((E)IDE), Institute of Electrical and Electronics Engineers (IEEE) 1394, fiber channel, Small Computer Systems Interface (SCSI), Universal Serial Bus (USB), and/or the like.
Network interfaces 1110 may accept, communicate, and/or connect to a communications network 1113. Through a communications network 1113, the MDGAAT controller is accessible through remote clients 1133b (e.g., computers with web browsers) by users 1133a. Network interfaces may employ connection protocols such as, but not limited to: direct connect, Ethernet (thick, thin, twisted pair 10/100/1000 Base T, and/or the like), Token Ring, wireless connection such as IEEE 802.11a-x, and/or the like. Should processing requirements dictate a greater amount speed and/or capacity, distributed network controllers (e.g., Distributed MDGAAT), architectures may similarly be employed to pool, load balance, and/or otherwise increase the communicative bandwidth required by the MDGAAT controller. A communications network may be any one and/or the combination of the following: a direct interconnection; the Internet; a Local Area Network (LAN); a Metropolitan Area Network (MAN); an Operating Missions as Nodes on the Internet (OMNI); a secured custom connection; a Wide Area Network (WAN); a wireless network (e.g., employing protocols such as, but not limited to a Wireless Application Protocol (WAP), I-mode, and/or the like); and/or the like. A network interface may be regarded as a specialized form of an input output interface. Further, multiple network interfaces 1110 may be used to engage with various communications network types 1113. For example, multiple network interfaces may be employed to allow for the communication over broadcast, multicast, and/or unicast networks.
Input Output interfaces (I/O) 1108 may accept, communicate, and/or connect to user input devices 1111, peripheral devices 1112, cryptographic processor devices 1128, and/or the like. I/O may employ connection protocols such as, but not limited to: audio: analog, digital, monaural, RCA, stereo, and/or the like; data: Apple Desktop Bus (ADB), IEEE 1394a-b, serial, universal serial bus (USB); infrared; joystick; keyboard; midi; optical; PC AT; PS/2; parallel; radio; video interface: Apple Desktop Connector (ADC), BNC, coaxial, component, composite, digital, Digital Visual Interface (DVI), high-definition multimedia interface (HDMI), RCA, RF antennae, S-Video, VGA, and/or the like; wireless transceivers: 802.11a/b/g/n/x; Bluetooth; cellular (e.g., code division multiple access (CDMA), high speed packet access (HSPA(+)), high-speed downlink packet access (HSDPA), global system for mobile communications (GSM), long term evolution (LTE), WiMax, etc.); and/or the like. One typical output device may include a video display, which typically comprises a Cathode Ray Tube (CRT) or Liquid Crystal Display (LCD) based monitor with an interface (e.g., DVI circuitry and cable) that accepts signals from a video interface, may be used. The video interface composites information generated by a computer systemization and generates video signals based on the composited information in a video memory frame. Another output device is a television set, which accepts signals from a video interface. Typically, the video interface provides the composited video information through a video connection interface that accepts a video display interface (e.g., an RCA composite video connector accepting an RCA composite video cable; a DVI connector accepting a DVI display cable, etc.).
User input devices 1111 often are a type of peripheral device 1112 (see below) and may include: card readers, dongles, finger print readers, gloves, graphics tablets, joysticks, keyboards, microphones, mouse (mice), remote controls, retina readers, touch screens (e.g., capacitive, resistive, etc.), trackballs, trackpads, sensors (e.g., accelerometers, ambient light, GPS, gyroscopes, proximity, etc.), styluses, and/or the like.
Peripheral devices 1112 may be connected and/or communicate to I/O and/or other facilities of the like such as network interfaces, storage interfaces, directly to the interface bus, system bus, the CPU, and/or the like. Peripheral devices may be external, internal and/or part of the MDGAAT controller. Peripheral devices may include: antenna, audio devices (e.g., line-in, line-out, microphone input, speakers, etc.), cameras (e.g., still, video, webcam, etc.), dongles (e.g., for copy protection, ensuring secure transactions with a digital signature, and/or the like), external processors (for added capabilities; e.g., crypto devices 1128), force-feedback devices (e.g., vibrating motors), network interfaces, printers, scanners, storage devices, transceivers (e.g., cellular, GPS, etc.), video devices (e.g., goggles, monitors, etc.), video sources, visors, and/or the like. Peripheral devices often include types of input devices (e.g., cameras).
It should be noted that although user input devices and peripheral devices may be employed, the MDGAAT controller may be embodied as an embedded, dedicated, and/or monitor-less (i.e., headless) device, wherein access would be provided over a network interface connection.
Cryptographic units such as, but not limited to, microcontrollers, processors 1126, interfaces 1127, and/or devices 1128 may be attached, and/or communicate with the MDGAAT controller. A MC68HC16 microcontroller, manufactured by Motorola Inc., may be used for and/or within cryptographic units. The MC68HC16 microcontroller utilizes a 16-bit multiply-and-accumulate instruction in the 16 MHz configuration and requires less than one second to perform a 512-bit RSA private key operation. Cryptographic units support the authentication of communications from interacting agents, as well as allowing for anonymous transactions. Cryptographic units may also be configured as part of the CPU. Equivalent microcontrollers and/or processors may also be used. Other commercially available specialized cryptographic processors include: the Broadcom's CryptoNetX and other Security Processors; nCipher's nShield, SafeNet's Luna PCI (e.g., 7100) series; Semaphore Communications' 40 MHz Roadrunner 184; Sun's Cryptographic Accelerators (e.g., Accelerator 6000 PCIe Board, Accelerator 500 Daughtercard); Via Nano Processor (e.g., L2100, L2200, U2400) line, which is capable of performing 500+MB/s of cryptographic instructions; VLSI Technology's 33 MHz 6868; and/or the like.
Generally, any mechanization and/or embodiment allowing a processor to affect the storage and/or retrieval of information is regarded as memory 1129. However, memory is a fungible technology and resource, thus, any number of memory embodiments may be employed in lieu of or in concert with one another. It is to be understood that the MDGAAT controller and/or a computer systemization may employ various forms of memory 1129. For example, a computer systemization may be configured wherein the operation of on-chip CPU memory (e.g., registers), RAM, ROM, and any other storage devices are provided by a paper punch tape or paper punch card mechanism; however, such an embodiment would result in an extremely slow rate of operation. In a typical configuration, memory 1129 will include ROM 1106, RAM 1105, and a storage device 1114. A storage device 1114 may be any conventional computer system storage. Storage devices may include a drum; a (fixed and/or removable) magnetic disk drive; a magneto-optical drive; an optical drive (i.e., Blueray, CD ROM/RAM/Recordable (R)/ReWritable (RW), DVD R/RW, HD DVD R/RW etc.); an array of devices (e.g., Redundant Array of Independent Disks (RAID)); solid state memory devices (USB memory, solid state drives (SSD), etc.); other processor-readable storage mediums; and/or other devices of the like. Thus, a computer systemization generally requires and makes use of memory.
The memory 1129 may contain a collection of program and/or database components and/or data such as, but not limited to: operating system component(s) 1115 (operating system); information server component(s) 1116 (information server); user interface component(s) 1117 (user interface); Web browser component(s) 1118 (Web browser); database(s) 1119; mail server component(s) 1121; mail client component(s) 1122; cryptographic server component(s) 1120 (cryptographic server); the MDGAAT component(s) 1135; and/or the like (i.e., collectively a component collection). These components may be stored and accessed from the storage devices and/or from storage devices accessible through an interface bus. Although non-conventional program components such as those in the component collection, typically, are stored in a local storage device 1114, they may also be loaded and/or stored in memory such as: peripheral devices, RAM, remote storage facilities through a communications network, ROM, various forms of memory, and/or the like.
The operating system component 1115 is an executable program component facilitating the operation of the MDGAAT controller. Typically, the operating system facilitates access of I/O, network interfaces, peripheral devices, storage devices, and/or the like. The operating system may be a highly fault tolerant, scalable, and secure system such as: Apple Macintosh OS X (Server); AT&T Plan 9; Be OS; Unix and Unix-like system distributions (such as AT&T's UNIX; Berkley Software Distribution (BSD) variations such as FreeBSD, NetBSD, OpenBSD, and/or the like; Linux distributions such as Red Hat, Ubuntu, and/or the like); and/or the like operating systems. However, more limited and/or less secure operating systems also may be employed such as Apple Macintosh OS, IBM OS/2, Microsoft DOS, Microsoft Windows 2000/2003/3.1/95/98/CE/Millenium/NT/Vista/XP (Server), Palm OS, and/or the like. An operating system may communicate to and/or with other components in a component collection, including itself, and/or the like. Most frequently, the operating system communicates with other program components, user interfaces, and/or the like. For example, the operating system may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses. The operating system, once executed by the CPU, may enable the interaction with communications networks, data, I/O, peripheral devices, program components, memory, user input devices, and/or the like. The operating system may provide communications protocols that allow the MDGAAT controller to communicate with other entities through a communications network 1113. Various communication protocols may be used by the MDGAAT controller as a subcarrier transport mechanism for interaction, such as, but not limited to: multicast, TCP/IP, UDP, unicast, and/or the like.
An information server component 1116 is a stored program component that is executed by a CPU. The information server may be a conventional Internet information server such as, but not limited to Apache Software Foundation's Apache, Microsoft's Internet Information Server, and/or the like. The information server may allow for the execution of program components through facilities such as Active Server Page (ASP), ActiveX, (ANSI) (Objective-) C (++), C# and/or .NET, Common Gateway Interface (CGI) scripts, dynamic (D) hypertext markup language (HTML), FLASH, Java, JavaScript, Practical Extraction Report Language (PERL), Hypertext Pre-Processor (PHP), pipes, Python, wireless application protocol (WAP), WebObjects, and/or the like. The information server may support secure communications protocols such as, but not limited to, File Transfer Protocol (FTP); HyperText Transfer Protocol (HTTP); Secure Hypertext Transfer Protocol (HTTPS), Secure Socket Layer (SSL), messaging protocols (e.g., America Online (AOL) Instant Messenger (AIM), Application Exchange (APEX), ICQ, Internet Relay Chat (IRC), Microsoft Network (MSN) Messenger Service, Presence and Instant Messaging Protocol (PRIM), Internet Engineering Task Force's (IETF's) Session Initiation Protocol (SIP), SIP for Instant Messaging and Presence Leveraging Extensions (SIMPLE), open XML-based Extensible Messaging and Presence Protocol (XMPP) (i.e., Jabber or Open Mobile Alliance's (OMA's) Instant Messaging and Presence Service (IMPS)), Yahoo! Instant Messenger Service, and/or the like. The information server provides results in the form of Web pages to Web browsers, and allows for the manipulated generation of the Web pages through interaction with other program components. After a Domain Name System (DNS) resolution portion of an HTTP request is resolved to a particular information server, the information server resolves requests for information at specified locations on the MDGAAT controller based on the remainder of the HTTP request. For example, a request such as http://123.124.125.126/myInformation.html might have the IP portion of the request “123.124.125.126” resolved by a DNS server to an information server at that IP address; that information server might in turn further parse the http request for the “/myInformation.html” portion of the request and resolve it to a location in memory containing the information “myInformation.html.” Additionally, other information serving protocols may be employed across various ports, e.g., FTP communications across port 21, and/or the like. An information server may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the information server communicates with the MDGAAT database 1119, operating systems, other program components, user interfaces, Web browsers, and/or the like.
Access to the MDGAAT database may be achieved through a number of database bridge mechanisms such as through scripting languages as enumerated below (e.g., CGI) and through inter-application communication channels as enumerated below (e.g., CORBA, WebObjects, etc.). Any data requests through a Web browser are parsed through the bridge mechanism into appropriate grammars as required by the MDGAAT. In one embodiment, the information server would provide a Web form accessible by a Web browser. Entries made into supplied fields in the Web form are tagged as having been entered into the particular fields, and parsed as such. The entered terms are then passed along with the field tags, which act to instruct the parser to generate queries directed to appropriate tables and/or fields. In one embodiment, the parser may generate queries in standard SQL by instantiating a search string with the proper join/select commands based on the tagged text entries, wherein the resulting command is provided over the bridge mechanism to the MDGAAT as a query. Upon generating query results from the query, the results are passed over the bridge mechanism, and may be parsed for formatting and generation of a new results Web page by the bridge mechanism. Such a new results Web page is then provided to the information server, which may supply it to the requesting Web browser.
Also, an information server may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
Computer interfaces in some respects are similar to automobile operation interfaces. Automobile operation interface elements such as steering wheels, gearshifts, and speedometers facilitate the access, operation, and display of automobile resources, and status. Computer interaction interface elements such as check boxes, cursors, menus, scrollers, and windows (collectively and commonly referred to as widgets) similarly facilitate the access, capabilities, operation, and display of data and computer hardware and operating system resources, and status. Operation interfaces are commonly called user interfaces. Graphical user interfaces (GUIs) such as the Apple Macintosh Operating System's Aqua, IBM's OS/2, Microsoft's Windows 2000/2003/3.1/95/98/CE/Millenium/NT/XP/Vista/7 (i.e., Aero), Unix's X-Windows (e.g., which may include additional Unix graphic interface libraries and layers such as K Desktop Environment (KDE), mythTV and GNU Network Object Model Environment (GNOME)), web interface libraries (e.g., ActiveX, AJAX, (D)HTML, FLASH, Java, JavaScript, etc. interface libraries such as, but not limited to, Dojo, jQuery(UI), MooTools, Prototype, script.aculo.us, SWFObject, Yahoo! User Interface, any of which may be used and) provide a baseline and means of accessing and displaying information graphically to users.
A user interface component 1117 is a stored program component that is executed by a CPU. The user interface may be a conventional graphic user interface as provided by, with, and/or atop operating systems and/or operating environments such as already discussed. The user interface may allow for the display, execution, interaction, manipulation, and/or operation of program components and/or system facilities through textual and/or graphical facilities. The user interface provides a facility through which users may affect, interact, and/or operate a computer system. A user interface may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the user interface communicates with operating systems, other program components, and/or the like. The user interface may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
A Web browser component 1118 is a stored program component that is executed by a CPU. The Web browser may be a conventional hypertext viewing application such as Microsoft Internet Explorer or Netscape Navigator. Secure Web browsing may be supplied with 128 bit (or greater) encryption by way of HTTPS, SSL, and/or the like. Web browsers allowing for the execution of program components through facilities such as ActiveX, AJAX, (D)HTML, FLASH, Java, JavaScript, web browser plug-in APIs (e.g., FireFox, Safari Plug-in, and/or the like APIs), and/or the like. Web browsers and like information access tools may be integrated into PDAs, cellular telephones, and/or other mobile devices. A Web browser may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the Web browser communicates with information servers, operating systems, integrated program components (e.g., plug-ins), and/or the like; e.g., it may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses. Also, in place of a Web browser and information server, a combined application may be developed to perform similar operations of both. The combined application would similarly affect the obtaining and the provision of information to users, user agents, and/or the like from the MDGAAT enabled nodes. The combined application may be nugatory on systems employing standard Web browsers.
A mail server component 1121 is a stored program component that is executed by a CPU 1103. The mail server may be a conventional Internet mail server such as, but not limited to sendmail, Microsoft Exchange, and/or the like. The mail server may allow for the execution of program components through facilities such as MDGAAT, ActiveX, (ANSI) (Objective-) C (++), C# and/or .NET, CGI scripts, Java, JavaScript, PERL, PHP, pipes, Python, WebObjects, and/or the like. The mail server may support communications protocols such as, but not limited to: Internet message access protocol (IMAP), Messaging Application Programming Interface (MAPI)/Microsoft Exchange, post office protocol (POP3), simple mail transfer protocol (SMTP), and/or the like. The mail server can route, forward, and process incoming and outgoing mail messages that have been sent, relayed and/or otherwise traversing through and/or to the MDGAAT.
Access to the MDGAAT mail may be achieved through a number of APIs offered by the individual Web server components and/or the operating system.
Also, a mail server may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, information, and/or responses.
A mail client component 1122 is a stored program component that is executed by a CPU 1103. The mail client may be a conventional mail viewing application such as Apple Mail, Microsoft Entourage, Microsoft Outlook, Microsoft Outlook Express, Mozilla, Thunderbird, and/or the like. Mail clients may support a number of transfer protocols, such as: IMAP, Microsoft Exchange, POP3, SMTP, and/or the like. A mail client may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the mail client communicates with mail servers, operating systems, other mail clients, and/or the like; e.g., it may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, information, and/or responses. Generally, the mail client provides a facility to compose and transmit electronic mail messages.
A cryptographic server component 1120 is a stored program component that is executed by a CPU 1103, cryptographic processor 1126, cryptographic processor interface 1127, cryptographic processor device 1128, and/or the like. Cryptographic processor interfaces will allow for expedition of encryption and/or decryption requests by the cryptographic component; however, the cryptographic component, alternatively, may run on a conventional CPU. The cryptographic component allows for the encryption and/or decryption of provided data. The cryptographic component allows for both symmetric and asymmetric (e.g., Pretty Good Protection (PGP)) encryption and/or decryption. The cryptographic component may employ cryptographic techniques such as, but not limited to: digital certificates (e.g., X.509 authentication framework), digital signatures, dual signatures, enveloping, password access protection, public key management, and/or the like. The cryptographic component will facilitate numerous (encryption and/or decryption) security protocols such as, but not limited to: checksum, Data Encryption Standard (DES), Elliptical Curve Encryption (ECC), International Data Encryption Algorithm (IDEA), Message Digest 5 (MD5, which is a one way hash operation), passwords, Rivest Cipher (RC5), Rijndael, RSA (which is an Internet encryption and authentication system that uses an algorithm developed in 1977 by Ron Rivest, Adi Shamir, and Leonard Adleman), Secure Hash Algorithm (SHA), Secure Socket Layer (SSL), Secure Hypertext Transfer Protocol (HTTPS), and/or the like. Employing such encryption security protocols, the MDGAAT may encrypt all incoming and/or outgoing communications and may serve as node within a virtual private network (VPN) with a wider communications network. The cryptographic component facilitates the process of “security authorization” whereby access to a resource is inhibited by a security protocol wherein the cryptographic component effects authorized access to the secured resource. In addition, the cryptographic component may provide unique identifiers of content, e.g., employing and MD5 hash to obtain a unique signature for an digital audio file. A cryptographic component may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. The cryptographic component supports encryption schemes allowing for the secure transmission of information across a communications network to enable the MDGAAT component to engage in secure transactions if so desired. The cryptographic component facilitates the secure accessing of resources on the MDGAAT and facilitates the access of secured resources on remote systems; i.e., it may act as a client and/or server of secured resources. Most frequently, the cryptographic component communicates with information servers, operating systems, other program components, and/or the like. The cryptographic component may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
The MDGAAT database component 1119 may be embodied in a database and its stored data. The database is a stored program component, which is executed by the CPU; the stored program component portion configuring the CPU to process the stored data. The database may be a conventional, fault tolerant, relational, scalable, secure database such as Oracle or Sybase. Relational databases are an extension of a flat file. Relational databases consist of a series of related tables. The tables are interconnected via a key field. Use of the key field allows the combination of the tables by indexing against the key field; i.e., the key fields act as dimensional pivot points for combining information from various tables. Relationships generally identify links maintained between tables by matching primary keys. Primary keys represent fields that uniquely identify the rows of a table in a relational database. More precisely, they uniquely identify rows of a table on the “one” side of a one-to-many relationship.
Alternatively, the MDGAAT database may be implemented using various standard data-structures, such as an array, hash, (linked) list, struct, structured text file (e.g., XML), table, and/or the like. Such data-structures may be stored in memory and/or in (structured) files. In another alternative, an object-oriented database may be used, such as Frontier, ObjectStore, Poet, Zope, and/or the like. Object databases can include a number of object collections that are grouped and/or linked together by common attributes; they may be related to other object collections by some common attributes. Object-oriented databases perform similarly to relational databases with the exception that objects are not just pieces of data but may have other types of capabilities encapsulated within a given object. If the MDGAAT database is implemented as a data-structure, the use of the MDGAAT database 1119 may be integrated into another component such as the MDGAAT component 1135. Also, the database may be implemented as a mix of data structures, objects, and relational structures. Databases may be consolidated and/or distributed in countless variations through standard data processing techniques. Portions of databases, e.g., tables, may be exported and/or imported and thus decentralized and/or integrated.
In one embodiment, the database component 1119 includes several tables 1119a-e. A user accounts table 1119a includes fields such as, but not limited to: a user_id, user_wallet_id, user_device_id, user_created, user_firstname, user_lastname, user_email, user_address, user_birthday, user_clothing_size, user_body_type, user_gender, user_payment_devices, user_eye_color, user_hair_color, user_complexion, user_personalized_gesture_models, user_recommended_items, user_image, user_image_date, user_body_joint_location, and/or the like. The user accounts table may support and/or track multiple user accounts on a MDGAAT. A merchant accounts table 1119b includes fields such as, but not limited to: merchant_id, merchant_created, merchant_name, merchant_email, merchant_address, merchant_products, and/or the like. The merchant accounts table may support and/or track multiple merchant accounts on a MDGAAT. An MDGA table 1119c includes fields such as, but not limited to: MDGA_id, MDGA_name, MDGA_touch_gestures, MDGA_finger_gestures, MDGA_QR_gestures, MDGA_object_gestures, MDGA_vocal_commands, MDGA_merchant, and/or the like. The MDGA table may support and/or track multiple possible composite actions on a MDGAAT. A products table 1119d includes fields such as, but not limited to: product_id, product_name, product_date_added, product_image, product_merchant, product_qr, product_manufacturer, product_model, product_price, product_aisle, product_stack, product_shelf, product_type, product_attributes, and/or the like. The products table may support and/or track multiple merchants' products on a MDGAAT. A payment device table 1119e includes fields such as, but not limited to: pd_id, pd_user, pd_type, pd_issuer, pd_issuer_id, pd_qr, pd_date_added, and/or the like. The payment device table may support and/or track multiple payment devices used on a MDGAAT. A transaction table 1119f includes fields such as, but not limited to: transaction_id, transaction_entity1, transaction_entity2, transaction_amount, transaction_date, transaction_receipt_copy, transaction_products, transaction_notes, and/or the like. The transaction table may support and/or track multiple transactions performed on a MDGAAT. An object gestures table 1119g includes fields such as, but not limited to: object_gesture_id, object_gesture_type, object_gesture_x, object_gesture_x, object_gesture_merchant, and/or the like. The object gesture table may support and/or track multiple object gestures performed on a MDGAAT. A finger gesture table 1119h includes fields such as, but not limited to: finger_gesture_id, finger_gesture_type, finger_gesture_x, finger_gesture_x, finger_gesture_merchant, and/or the like. The finger gestures table may support and/or track multiple finger gestures performed on MDGAAT. A touch gesture table 1119i includes fields such as, but not limited to touch_gesture_id, touch_gesturetype, touch_gesture_x, touch_gesture_x, touch_gesturemerchant, and/or the like. The touch gestures table may support and/or track multiple touch gestures performed on a MDGAAT. A QR gesture table 1119j includes fields such as, but not limited to: QR_gesture_id, QR_gesture_type, QR_gesture_x, QR_gesture_x, QR_gesture_merchant, and/or the like. The QUADRATIC RESAMPLING gestures table may support and/or track multiple QR gestures performed on a MDGAAT. A vocal command table 1119k includes fields such as, but not limited to: vc_id, vc_name, vc_command_list, and/or the like. The vocal command gestures table may support and/or track multiple vocal commands performed on a MDGAAT. In one embodiment, the MDGAAT database may interact with other database systems. For example, employing a distributed database system, queries and data access by search MDGAAT component may treat the combination of the MDGAAT database, an integrated data security layer database as a single database entity.
In one embodiment, the MDGAAT database may interact with other database systems. For example, employing a distributed database system, queries and data access by search MDGAAT component may treat the combination of the MDGAAT database, an integrated data security layer database as a single database entity.
In one embodiment, user programs may contain various user interface primitives, which may serve to update the MDGAAT. Also, various accounts may require custom database tables depending upon the environments and the types of clients the MDGAAT may need to serve. It should be noted that any unique fields may be designated as a key field throughout. In an alternative embodiment, these tables have been decentralized into their own databases and their respective database controllers (i.e., individual database controllers for each of the above tables). Employing standard data processing techniques, one may further distribute the databases over several computer systemizations and/or storage devices. Similarly, configurations of the decentralized database controllers may be varied by consolidating and/or distributing the various database components 1141-1145. The Audio/Gesture Conversion Component 1141 handles translating audio and gesture data into actions. The Virtual Store Previewing Component 1142 handles virtual previews of store products. The Action Processing Component 1143 handles carrying out actions translated from the Audio/Gesture Conversion Component. The Image Processing 1144 handles processing images and videos for the purpose of locating information and/or determining gestures. The Audio Processing 1145 handles processing audio files and videos for the purpose of locating information and/or determining vocal commands. The MDGAAT may be configured to keep track of various settings, inputs, and parameters via database controllers.
The MDGAAT database may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the MDGAAT database communicates with the MDGAAT component, other program components, and/or the like. The database may contain, retain, and provide information regarding other nodes and data.
The MDGAAT component 1135 is a stored program component that is executed by a CPU. In one embodiment, the MDGAAT component incorporates any and/or all combinations of the aspects of the MDGAAT discussed in the previous figures. As such, the MDGAAT affects accessing, obtaining and the provision of information, services, transactions, and/or the like across various communications networks.
The MDGAAT component may transform reality scene visual captures (e.g., see 213 in
The MDGAAT component enabling access of information between nodes may be developed by employing standard development tools and languages such as, but not limited to: Apache components, Assembly, ActiveX, binary executables, (ANSI) (Objective-) C (++), C# and/or .NET, database adapters, CGI scripts, Java, JavaScript, mapping tools, procedural and object oriented development tools, PERL, PHP, Python, shell scripts, SQL commands, web application server extensions, web development environments and libraries (e.g., Microsoft's ActiveX; Adobe AIR, FLEX & FLASH; AJAX; (D)HTML; Dojo, Java; JavaScript; jQuery(UI); MooTools; Prototype; script.aculo.us; Simple Object Access Protocol (SOAP); SWFObject; Yahoo! User Interface; and/or the like), WebObjects, and/or the like. In one embodiment, the MDGAAT server employs a cryptographic server to encrypt and decrypt communications. The MDGAAT component may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the MDGAAT component communicates with the MDGAAT database, operating systems, other program components, and/or the like. The MDGAAT may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
The structure and/or operation of any of the MDGAAT node controller components may be combined, consolidated, and/or distributed in any number of ways to facilitate development and/or deployment. Similarly, the component collection may be combined in any number of ways to facilitate deployment and/or development. To accomplish this, one may integrate the components into a common code base or in a facility that can dynamically load the components on demand in an integrated fashion.
The component collection may be consolidated and/or distributed in countless variations through standard data processing and/or development techniques. Multiple instances of any one of the program components in the program component collection may be instantiated on a single node, and/or across numerous nodes to improve performance through load-balancing and/or data-processing techniques. Furthermore, single instances may also be distributed across multiple controllers and/or storage devices; e.g., databases. All program component instances and controllers working in concert may do so through standard data processing communication techniques.
The configuration of the MDGAAT controller will depend on the context of system deployment. Factors such as, but not limited to, the budget, capacity, location, and/or use of the underlying hardware resources may affect deployment requirements and configuration. Regardless of if the configuration results in more consolidated and/or integrated program components, results in a more distributed series of program components, and/or results in some combination between a consolidated and distributed configuration, data may be communicated, obtained, and/or provided. Instances of components consolidated into a common code base from the program component collection may communicate, obtain, and/or provide data. This may be accomplished through intra-application data processing communication techniques such as, but not limited to: data referencing (e.g., pointers), internal messaging, object instance variable communication, shared memory space, variable passing, and/or the like.
If component collection components are discrete, separate, and/or external to one another, then communicating, obtaining, and/or providing data with and/or to other components may be accomplished through inter-application data processing communication techniques such as, but not limited to: Application Program Interfaces (API) information passage; (distributed) Component Object Model ((D)COM), (Distributed) Object Linking and Embedding ((D)OLE), and/or the like), Common Object Request Broker Architecture (CORBA), Jini local and remote application program interfaces, JavaScript Object Notation (JSON), Remote Method Invocation (RMI), SOAP, process pipes, shared files, and/or the like. Messages sent between discrete component components for inter-application communication or within memory spaces of a singular component for intra-application communication may be facilitated through the creation and parsing of a grammar. A grammar may be developed by using development tools such as lex, yacc, XML, and/or the like, which allow for grammar generation and parsing capabilities, which in turn may form the basis of communication messages within and between components.
For example, a grammar may be arranged to recognize the tokens of an HTTP post command, e.g.:
where Value1 is discerned as being a parameter because “http://” is part of the grammar syntax, and what follows is considered part of the post value. Similarly, with such a grammar, a variable “Value1” may be inserted into an “http://” post command and then sent. The grammar syntax itself may be presented as structured data that is interpreted and/or otherwise used to generate the parsing mechanism (e.g., a syntax description text file as processed by lex, yacc, etc.). Also, once the parsing mechanism is generated and/or instantiated, it itself may process and/or parse structured data such as, but not limited to: character (e.g., tab) delineated text, HTML, structured text streams, XML, and/or the like structured data. In another embodiment, inter-application data processing protocols themselves may have integrated and/or readily available parsers (e.g., JSON, SOAP, and/or like parsers) that may be employed to parse (e.g., communications) data. Further, the parsing grammar may be used beyond message parsing, but may also be used to parse: databases, data collections, data stores, structured data, and/or the like. Again, the desired configuration will depend upon the context, environment, and requirements of system deployment.
For example, in some implementations, the MDGAAT controller may be executing a PHP script implementing a Secure Sockets Layer (“SSL”) socket server via the information server, which listens to incoming communications on a server port to which a client may send data, e.g., data encoded in JSON format. Upon identifying an incoming communication, the PHP script may read the incoming message from the client device, parse the received JSON-encoded text data to extract information from the JSON-encoded text data into PHP script variables, and store the data (e.g., client identifying information, etc.) and/or extracted information in a relational database accessible using the Structured Query Language (“SQL”). An exemplary listing, written substantially in the form of PHP/SQL commands, to accept JSON-encoded input data from a client device via a SSL connection, parse the data to extract variables, and store the data to a database, is provided below:
Also, the following resources may be used to provide example embodiments regarding SOAP parser implementation:
and other parser implementations:
all of which are hereby expressly incorporated by reference herein.
In order to address various issues and advance the art, the entirety of this application (including the Cover Page, Title, Headings, Field, Background, Summary, Brief Description of the Drawings, Detailed Description, Claims, Abstract, Figures, Appendices and/or otherwise) shows by way of illustration various embodiments in which the claimed innovations may be practiced. The advantages and features of the application are of a representative sample of embodiments only, and are not exhaustive and/or exclusive. They are presented only to assist in understanding and teach the claimed principles. It should be understood that they are not representative of all claimed innovations. As such, certain aspects of the disclosure have not been discussed herein. That alternate embodiments may not have been presented for a specific portion of the innovations or that further undescribed alternate embodiments may be available for a portion is not to be considered a disclaimer of those alternate embodiments. It will be appreciated that many of those undescribed embodiments incorporate the same principles of the innovations and others are equivalent. Thus, it is to be understood that other embodiments may be utilized and functional, logical, operational, organizational, structural and/or topological modifications may be made without departing from the scope and/or spirit of the disclosure. As such, all examples and/or embodiments are deemed to be non-limiting throughout this disclosure. Also, no inference should be drawn regarding those embodiments discussed herein relative to those not discussed herein other than it is as such for purposes of reducing space and repetition. For instance, it is to be understood that the logical and/or topological structure of any combination of any program components (a component collection), other components and/or any present feature sets as described in the figures and/or throughout are not limited to a fixed operating order and/or arrangement, but rather, any disclosed order is exemplary and all equivalents, regardless of order, are contemplated by the disclosure. Furthermore, it is to be understood that such features are not limited to serial execution, but rather, any number of threads, processes, services, servers, and/or the like that may execute asynchronously, concurrently, in parallel, simultaneously, synchronously, and/or the like are contemplated by the disclosure. As such, some of these features may be mutually contradictory, in that they cannot be simultaneously present in a single embodiment. Similarly, some features are applicable to one aspect of the innovations, and inapplicable to others. In addition, the disclosure includes other innovations not presently claimed. Applicant reserves all rights in those presently unclaimed innovations, including the right to claim such innovations, file additional applications, continuations, continuations in part, divisions, and/or the like thereof. As such, it should be understood that advantages, embodiments, examples, functional, features, logical, operational, organizational, structural, topological, and/or other aspects of the disclosure are not to be considered limitations on the disclosure as defined by the claims or limitations on equivalents to the claims. It is to be understood that, depending on the particular needs and/or characteristics of a MDGAAT individual and/or enterprise user, database configuration and/or relational model, data type, data transmission and/or network framework, syntax structure, and/or the like, various embodiments of the MDGAAT may be implemented that enable a great deal of flexibility and customization. For example, aspects of the MDGAAT may be adapted for (electronic/financial) trading systems, financial planning systems, and/or the like.
The AUGMENTED REALITY VISION DEVICE APPARATUSES, METHODS AND SYSTEMS (hereinafter “V-GLASSES”) transform mobile device location coordinate information transmissions, real-time reality visual capturing, and mixed gesture capturing, via V-GLASSES components, into real-time behavior-sensitive product purchase related information, shopping purchase transaction notifications, and electronic receipts. In one embodiment, a V-GLASSES device may take a form similar to a pair of eyeglasses, which may provide an enhanced view with virtual information labels atop the captured reality scene to a consumer who wears the V-GLASSES device.
Within embodiments, the V-GLASSES device may have a plurality of sensors and mechanisms including, but not limited to: front facing camera to capture a wearer's line of sight; rear facing camera to track the wearer's eye movement, dilation, retinal pattern; an infrared object distance sensor (e.g., such may be found in a camera allowing for auto-focus image range detection, etc.); EEG sensor array along the top inner periphery of the glasses so as to place the EEG sensors in contact with the wearers brow, temple, skin; dual microphones, one having a conical listening position pointing towards the wearer's mouth, a second external and front facing for noise cancellation and acquiring audio in the wearer's field of perception; accelerometers; gyroscopes; infrared/laser projector in the upper portion of the glasses distally placed from a screen element and usable for projecting rich media; a flip down transparent/semi-transparent/opaque LED screen element within the wearer's field of view; a speaker having an outward position towards those in the field of perception of the wearer; integrated headphones that may be connected by wire towards the armatures of the glasses such that they are proximate to the wearer's ears and may be placed into the wearer's ears; a plurality of removable and replaceable visors/filters that may be used for providing different types of enhanced views; and/or the like.
For example, in one implementation, a consumer wearing a pair of V-GLASSES device may obtain a view similar to the example augmented reality scenes illustrated in
In one implementation, merchants may utilize the V-GLASSES mechanisms to create new V-GLASSES shopping experiences for their customers. For example, V-GLASSES may integrate with alert mechanisms (e.g., V.me wallet push systems, vNotify, etc.) for fraud preventions, and/or the like. As another example, V-GLASSES may provide/integrate with merchant-specific loyalty programs (e.g., levels, points, notes, etc.), facilitate merchants to provide personal shopping assistance to VIP customers. In further implementations, via the V-GLASSES merchant UI platform, merchants may integrate and/or synchronize a consumer's wish list, shopping cart, referrals, loyalty, merchandise delivery options, and other shopping preference settings between online and in-store purchase.
Within implementations, V-GLASSES may employ a virtual wallet alert mechanisms (e.g., vNotify) to allow merchants to communicate with their customers without sharing customer's personal information (e.g., e-mail, mobile phone number, residential addresses, etc.). In one implementation, the consumer may engage a virtual wallet applications (e.g., Visa® V.me wallet) to complete purchases at the merchant PoS without revealing the consumer's payment information (e.g., a PAN number) to the merchant.
Integration of an electronic wallet, a desktop application, a plug-in to existing applications, a standalone mobile application, a web based application, a smart prepaid card, and/or the like in capturing payment transaction related objects such as purchase labels, payment cards, barcodes, receipts, and/or the like reduces the number of network transactions and messages that fulfill a transaction payment initiation and procurement of payment information (e.g., a user and/or a merchant does not need to generate paper bills or obtain and send digital images of paper bills, hand in a physical payment card to a cashier, etc., to initiate a payment transaction, fund transfer, and/or the like). In this way, with the reduction of network communications, the number of transactions that may be processed per day is increased, i.e., processing efficiency is improved, and bandwidth and network latency is reduced.
It should be noted that although a mobile wallet platform is depicted (e.g., see
In one implementation, the V-GLASSES may determine proximity 135b of the target payee John 141. For example, V-GLASSES may form a query to a remote server, a cloud, etc., to inquire about John's current location via V-GLASSES GPS tracking. As another example, V-GLASSES may track John's current location via John's wallet activities (e.g., scanning an item, check-in at a merchant store, as discussed in
In another implementation, if John 120b is within proximity to Jen 120a, Jen may send a communication message 135c “Jen sends $50.00 to John” to John 120b via various means, e.g., SMS, wallet messages, Bluetooth, Wi-Fi, and/or the like. In one implementation, Jen may communicate with John in proximity via an optical message, e.g., Jen's V-GLASSES device may be equipped with a blinking light 136a, the glasses may produce on/off effects, etc., to generate a binary optical sequence, which may encode the fund transfer message (e.g., Morse code, etc.). For example, such blinking light may be generated by the V-GLASSES glass turning black or white 136b, etc. In one implementation, John's V-GLASSES device, which is in proximity to Jen's, may capture the optical message, and decode it to extract the fund transfer request. In one implementation, John's V-GLASSES device may generate an optical message in a similar manner, to acknowledge receipt of Jen's message, e.g., “John accepts $50.00 transfer from Jen.” In further implementations, such optical message may be adopted to encode and/or encrypt various information, e.g., contact information, biometrics information, transaction information, and/or the like.
In one implementation, V-GLASSES may verify the transaction through integrated layers of information to prevent fraud, including verification such as facial recognition (e.g., whether the recipient is John Smith himself, etc.), geographical proximity (e.g., whether John Smith's is currently located at Jen's location, etc.), local proximity (e.g., whether John Smith successfully receives and returns an optical message “blinked” from Jen, etc.), and/or the like.
In one implementation, if the transaction verification 135d is positive, V-GLASSES may transfer $50.00 from Jen's account to John. Further implementations of transaction processing with regard to P2P transfer may be found in United States nonprovisional patent application Ser. No. 13/520,481, filed Jul. 3, 2012, entitled “Universal Electronic Payment Apparatuses, Methods and Systems,” attorney docket no. P-42051US02|VISA-19/02US, which is herein expressly incorporated by reference.
In alternative implementations, V-GLASSES may facilitate scanning an in-store scene and generate an inventory map based on visual capturing of inventory information of a merchant store and generate an inventory map based on image content detection. For example, as shown in
In one implementation, such cameras may be configured to scan the shelves periodically (e.g., every hour, etc.), and may form a camera social network to generate real-time updates of inventory information. For example, product items may be frequently taken off from a shelf by consumers, and such change in inventory may be captured by camera scanning, and reflected in the inventory updates. As another example, product items may be picked up by consumers and randomly placed at a wrong shelf, e.g., a can of “Organic Diced Tomato 16 OZ” being placed at the beauty product shelf, etc., and such inventory change may be captured and transmitted to the merchant store for correction. In further implementations, the camera scanning may facilitate security monitoring for the merchant store.
In further implementations, as shown in
In a further implementation, V-GLASSES may facilitate obtain an estimate of the shelf height, width, e.g., based on the angle of the vision, etc. In a similar manner, consumer John's 120b V-GLASSES may capture a “High Speed Internet Router” 132b in the electronics aisle 121b, and transmit such information for store inventory map updating. Multiple consumers' V-GLASSES capturing may generate various contributions for real-time store inventory updating.
In further implementations, V-GLASSES may project a QR code on a surface to facilitate a transaction. For example, as shown in
In further implementations, the V-GLASSES projection may be used for signature capture for security challenge (e.g., a consumer may sign with finger on a projected “signature area,” etc.)
In further implementations, the virtual “information wall” may be generated based on consumer interests, geo-location, and various atmospherics factors. For example, a V-GLASSES analytics component may determine a consumer may be interested in food, shoes, and electronics based on the consumer's purchasing history, browsing history, QR code scanning history, social media activities, and/or the like. V-GLASSES may generate an “information wall” including news feeds, social media feeds, ads, etc. related to the consumer's interested item categories, e.g., food, shoes and electronics, etc. V-GLASSES may further determine that when the consumer is at an office location, the consumer tends to browse “electronics” more often; as such, when V-GLASSES detects the consumer is at the office location, e.g., via GPS tracking, IP address, cell tower triangular positioning, etc., V-GLASSES may place “electronic” information to the consumer's “information wall.”
As another example, when a consumer is detected to be at an office location, V-GLASSES may fill an “information wall” with business related information labels, e.g., meeting reminders, stock banners, top business contacts, missing calls, new emails, and/or the like. In a further implementation, a consumer may set up and/or customize the “information wall” with interested items. For example, a consumer may choose to “display” a favorite oil painting, family picture, wedding photo on the “information wall,” so that the consumer may be able to see the personalized decoration item displayed via the V-GLASSES in an office setting, without having to physically hang or stitch the real picture/photo on a physical wall.
In one implementation, V-GLASSES may provide “layers” of “information walls.” For example, a consumer may “look” at an empty real wall via a V-GLASSES device and choose an “information wall” that the consumer would like to see, e.g., by articulating the name of the “wall” (e.g., “12 months electricity bills,” “my office wall,” etc.), by a mixed gesture command (e.g., waving leftward or rightward to proceed with another previously saved “information wall,” etc.), and/or the like. In another implementation, V-GLASSES may save and identify an “information wall” by generating a QR code 136, and display it at the corner of the “information wall.” A consumer may take a snap shot of the QR code via V-GLASSES device to identify the “information wall,” and/or to transmit information of the “information wall.” For example, a consumer may snap the QR code and project such QR code on a surface, and use a Smartphone to capture the QR code; in this way, the virtual “information wall” that is visible via a V-GLASSES device may be reproduced within the Smartphone based on the captured QR code.
In one implementation, the V-GLASSES device 130 may store, or retrieve information of an “information wall” from the QR code 136. For example, an example listing of an information wall record, substantially in the form of XML, is provided below:
As another example, V-GLASSES may place speech scripts 146b on Jen's hand to help Jen prepare a speech, e.g., when Jen looks down at her hand, she may see the speech script.
As another example, V-GLASSES may project stock banners on a trader's desk 146c, so that a trader may be able to expand the view of market data.
In a further implementation, V-GLASSES may generate a“virtual game” 146d. For example, when a consumer is waiting in a line, V-GLASSES may provide a virtual gaming option to entertain the consumer. When consumer Jen 120a looks down at her feet, V-GLASSES may generate virtual “walking bugs” in the scene, and if Jen 120a moves her feet to “squash the bug,” she may win a gaming point. In one implementation, when Jen 120a shift her focus from the ground (e.g., looking up, etc.), the “snatch the bug” game may automatically pause, and may resume when Jen stands still and looks down at the ground again.
With reference to
In one implementation, Jen 120a may look at her mobile phone which may have instantiated a mobile wallet component, and obtain a view of a list of virtual cards overlaying the reality scene 137. In one implementation, Jen 120a may point to a virtual card overlay 138 and articulate “Pay with this card” 151b. In one implementation, the virtual card overlay may be highlighted 139 upon Jen's fingertip pointing, and V-GLASSES may capture the verbal command to proceed a bill payment. For example, V-GLASSES may generate a payment transaction message paying Jen's October bill with Jen's PNC account.
With reference to
In further implementations, a V-GLASSES device may be accompanied with accessories such as various visors/filters for different layers of overlay labels. In one implementation, V-GLASSES may provide layers of information labels (e.g., similar to layers in augmented reality overlay as shown in
Alternatively, as shown in
In further implementations, consumer Jen 120a and John 120b may synchronized their view through the V-GLASSES devices. For example, Jen 120a may view a wall of virtually “stitched” utility bills, and may command 158b to synchronize the view with John 120b. In one implementation, Jen's V-GLASSES device may send a synchronization view message to John's, so that John will obtain the same view of virtually “stitched” utility bills when he looks at the wall 158c.
In one embodiment, V-GLASSES may generate social predictive purchase item recommendations based on a consumer's social atmospherics. For example, in one implementation, V-GLASSES may track a consumer's social media connections' social activities (e.g., Facebook status, posts, photos, comments, Tweets, Google+ status, Google+ messages, etc.) and generate heuristics of a possible gift recommendation. For example, if a consumer's Facebook friend has posted a “baby shower” event invitation, or a Facebook status updating indicating she is expecting a baby, V-GLASSES may generate a purchase recommendation for a baby gift to the consumer. As another example, if a consumer's Facebook friend's birthday is coming up, V-GLASSES may analyze the Facebook connection's social activities, purchasing history, etc. to determine the connection's interests (e.g., Facebook comments with regard to a brand, a product item, etc.; “likes”; posted photos related to a product category; hash tags of Tweets; published purchase history on social media; followed pages; followed social media celebrities; etc.). For example, if the consumer's connection follows a celebrity makeup artist on YouTube, and “likes” the page “Sephora,” V-GLASSES may recommend beauty products to the consumer as a gift for the consumer's connection when the connection's birthday is coming up.
In one implementation, such social “gifting” recommendations may be provided to the consumer via a Facebook ads, banner ads, cookie ads within a browser, messages, email, SMS, instant messages, wallet push messages, and/or the like. In further implementations, V-GLASSES may generate a recommendation via augmented reality information overlays. In the above social “birthday gifting” example, in one implementation, a consumer may view an augmented reality label “Gift idea for Jen!” overlaying a cosmetics product via the consumer's V-GLASSES.
In one implementation, the V-GLASSES social predictive gift component may obtain social history information via a virtual wallet component, e.g., the social publications related to purchase transactions of the consumer and/or the consumer's social connections. Further implementations of social publications may be found in U.S. nonprovisional patent application Ser. No. 13/520,481, filed Jul. 3, 2012, entitled “Universal Electronic Payment Apparatuses, Methods and Systems,” attorney docket no. P-42051US02|VISA-109/02US, which is herein expressly incorporated by reference. In another implementation, the V-GLASSES may obtain such social information and purchasing transaction information via an information aggregation platform, which aggregates, stores, and categories various consumer information across different platforms (e.g., transaction records at a transaction processing network, social media data, browsing history, purchasing history stored at a merchant, and/or the like). Further implementations of the information aggregation platform are discussed in U.S. provisional Ser. No. 61/594,063, entitled “Centralized Personal Information Platform Apparatuses, Methods And Systems,” filed Feb. 2, 2012, which is herein expressly incorporated by reference.
In further implementations, V-GLASSES may generate social predictive ads to the consumer, e.g., based on the consumer's purchasing patterns, seasonal purchases, and/or the like. For example, V-GLASSES may capture a consumer's habitual grocery purchases, e.g., one gallon of organic non-fat milk every two weeks, etc., and may generate a seasonal ads related to products, offers/rewards for organic milk every two weeks. Further implementations of the social predictive advertising component are discussed in U.S. non-provisional application Ser. No. 13/543,825, entitled “Bidirectional Bandwidth Reducing Notifications And Targeted Incentive Platform Apparatuses, Methods And Systems,” filed Jul. 7, 2012, which is herein expressly incorporated by reference.
In further implementations, V-GLASSES may submit information to a server for processing power saving. For example, V-GLASSES may pass on pattern recognition (e.g., store inventory map aggregation, facial recognition, etc.) requests to a server, a cloud, and/or the like. In one implementation, V-GLASSES may determine a distributed server to route such requests based on server availability, server geo-location, server specialty (e.g., a processor component dedicated for facial recognition, etc.).
In further implementations, the V-GLASSES device 130 may be adopted for security detection (e.g., retina scanning, etc.). A consumer may interact with V-GLASSES device via voice, gesture, brain waves, and/or the like.
In further implementations, the V-GLASSES may establish an image databases for pattern recognition. Such image database may include graphic content for image capture, maps, purchase, etc. For example, in one implementation, when a consumer sees an “iPad” via the V-GLASSES device, such image may be processed and compared to images previously stored in the image database to identify that the rectangular object is an “iPad.”
In further implementations, the consumer may operate a Smartphone as a remote control for the V-GLASSES device.
For example, the CSR app may include features such as described below in the discussion with reference to
Some embodiments of the V-GLASSES may feature a more streamlined login option for the consumer. For example, using a mobile device such as iPhone, the consumer may initially enter a device ID such as an Apple ID to get into the device. In one implementation, the device ID may be the ID used to gain access to the V-GLASSES application. As such, the V-GLASSES may use the device ID to identify the consumer and the consumer need not enter another set of credentials. In another implementation, the V-GLASSES application may identify the consumer using the device ID via federation. Again, the consumer may not need to enter his credentials to launch the V-GLASSES application. In some implementations, the consumer may also use their wallet credentials (e.g., V.me credentials) to access the V-GLASSES application. In such situations, the wallet credentials may be synchronized with the device credentials.
Once in the V-GLASSES application, the consumer may see some graphics that provide the consumer various options such as checking in and for carrying items in the store. In one implementation, as shown in
Some embodiments of the V-GLASSES application may include a dynamic key lock configuration. For example, the V-GLASSES application may include a dynamic keyboard that displays numbers or other characters in different configuration every time. Such a dynamic keypad would generate a different key entry pattern every time such that the consumer would need to enter their PIN every time. Such dynamic keypad may be used, for example, for entry of device ID, wallet PIN, and/or the like, and may provide an extra layer of security. In some embodiments, the dial and scrambled keypad may be provided based on user preference and settings. In other embodiments, the more cumbersome and intricate authentication mechanisms can be supplied based on increased seasoning and security requirements discussed in greater detail in U.S. patent application Ser. No. 13/434,818 filed Mar. 29, 2012 and titled “Graduated Security Seasoning Apparatuses, Methods and Systems,” and PCT international application serial no. PCT/US12/66898, filed Nov. 28, 2012, entitled “Transaction Security Graduated Seasoning And Risk Shifting Apparatuses, Methods And Systems,” which are all herein expressly incorporated by reference. These dynamic seasoned PIN authentication mechanisms may be used to authorize a purchase, and also to gain access to a purchasing application (e.g., wallet), to gain access to the device, and/or the like. In one embodiment, the GPS location of the device and/or discerned merchant may be used to determine a risk assessment of any purchasing made at such location and/or merchant, and as such may ratchet up or down the type of mechanism to be used for authentication/authorization.
In some embodiments, the V-GLASSES may also facilitate an outsourced customer service model wherein the customer service provider (e.g., sales associate) is remote, and the consumer may request help from the remote customer service provider by opening a communication channel from their mobile device application. The remote customer service provider may then guide the requesting user through the store and/or purchase.
With reference to
In an alternative implementation, a merchant 220 may optionally provide a store check-in information 206 so that the consumer may snap a picture of the provided store check-in information. The store check-in information 206 may include barcodes (e.g., UPC, 2D, QR code, etc.), a trademark logo, a street address plaque, and/or the like, displayed at the merchant store 220. The consumer mobile device may then generate a check-in request 208 including the snapped picture of store check-in information 206 to the V-GLASSES server 210. In further implementations, the store check-in information 206 may include a store floor plan transmitted to the consumer via MMS, wallet push messages, email, and/or the like.
For example, the store information 206 to the V-GLASSES consumer, substantially in the form of XML-formatted data, is provided below:
As another example, the consumer mobile device 203 may generate a (Secure) Hypertext Transfer Protocol (“HTTP(S)”) POST message including the consumer check-in information for the V-GLASSES server 210 in the form of data formatted according to the XML. An example listing of a checkout request 208 to the V-GLASSES server, substantially in the form of a HTTP(S) POST message including XML-formatted data, is provided below:
d rXYZ x gXYZ
bXYZ rTRC
{acute over ( )} aarg À vcgt ...
The above exemplary check-in request message includes a snapped image (e.g., QR code, trademark logo, storefront, etc.) for the V-GLASSES server 210 to process and extract merchant information 209. In another implementation, the mobile device 203 may snap and extract merchant information from the snapped QR code, and include such merchant information into the consumer check-in information 208.
In another implementation, the check-in message 208 may further include the consumer's GPS coordinates for the V-GLASSES server 210 to associate a merchant store with the consumer's location. In further implementations, the check-in message 208 may include additional information, such as, but not limited to biometrics (e.g., voice, fingerprint, facial, etc.), e.g., a consumer provides biometric information to a merchant PoS terminal, etc., mobile device identity (e.g., IMEI, ESN, SIMid, etc.), mobile component security identifying information, trusted execution environment (e.g., Intel TXT, TrustZone, etc.), and/or the like.
In one implementation, upon V-GLASSES server obtaining merchant information 209 from the consumer check-in request message 208, V-GLASSES server 210 may query for related consumer loyalty profile 218 from a database 219. In one implementation, the consumer profile query 218 may be performed at the V-GLASSES server 210, and/or at the merchant 220 based on merchant previously stored consumer loyalty profile database. For example, the V-GLASSES database 219 may be a relational database responsive to Structured Query Language (“SQL”) commands. The V-GLASSES server may execute a hypertext preprocessor (“PHP”) script including SQL commands to query a database table (such as
In one implementation, the V-GLASSES may obtain the query result including the consumer loyalty offers profile (e.g., loyalty points with the merchant, with related merchants, product items the consumer previously purchased, product items the consumer previously scanned, locations of such items, etc.) 220, and may optionally provide the consumer profile information 223 to the merchant. For example, in one implementation, the queried consumer loyalty profile 220 and/or the profile information provided to the merchant CSR 223, substantially in the form of XML-formatted data, is provided below:
In the above example, V-GLASSES may optionally provide information on the consumer's previously viewed or purchased items to the merchant. For example, the consumer has previously scanned the QR code of a product “Michael Kors Flat Pants” and such information including the inventory availability, SKU location, etc. may be provided to the merchant CSR, so that the merchant CSR may provide a recommendation to the consumer. In one implementation, the consumer loyalty message 223 may not include sensitive information such as consumer's wallet account information, contact information, purchasing history, and/or the like, so that the consumer's private financial information is not exposed to the merchant.
Alternatively, the merchant 220 may query its local database for consumer loyalty profile associated with the merchant, and retrieve consumer loyalty profile information similar to message 223. For example, in one implementation, at the merchant 220, upon receiving consumer check-in information, the merchant may determine a CSR for the consumer 212. For example, the merchant may query a local consumer loyalty profile database to determine the consumer's status, e.g., whether the consumer is a returning customer, or a new customer, whether the consumer has been treated with a particular CSR, etc., to assign a CSR to the consumer. In one implementation, the CSR 230 may receive a consumer assignment 224 notification at a CSR terminal 240 (e.g., a PoS terminal, a mobile device, etc.). In one implementation, the consumer assignment notification message 224 may include consumer loyalty profile with the merchant, consumer's previous viewed or purchased item information, and/or the like (e.g., similar to that in message 223), and may be sent via email, SMS, instant messenger, PoS transmission, and/or the like. For example, in one implementation, the consumer assignment notification 224, substantially in the form of XML-formatted data, is provided below:
In the above example, the consumer assignment notification 224 includes basic consumer information, and CSR profile information (e.g., CSR specialty, availability, language support skills, etc.). Additionally, the consumer assignment notification 224 may include consumer loyalty profile that may take a form similar to that in 223.
In one implementation, the consumer may optionally submit in-store scanning information 225a to the CSR (e.g., the consumer may interact with the CSR so that the CSR may assist the scanning of an item, etc.), which may provide consumer interest indications to the CSR, and update the consumer's in-store location with the CSR. For example, in one implementation, the consumer scanning item message 225a, substantially in the form of XML-formatted data, is provided below:
Additionally, the consumer scanning information 225a may be provided to the V-GLASSES server to update consumer interests and location information.
Upon receiving consumer loyalty information and updated location information, the CSR terminal 240 may retrieve a list of complementary items for recommendations 225b, e.g., items close to the consumer's in-store location, items related to the consumer's previous viewed items, etc. In one implementation, the CSR may submit a selection of the retrieved items to recommend to the consumer 226, wherein such selection may be based on the real-time communication between the consumer and the CSR, e.g., in-person communication, SMS, video chat, V-GLASSES push messages (e.g., see 416a-b in
In one implementation, upon receiving the consumer assignment notification, CSR may interact with the consumer 202 to assist shopping. For example, the CSR 230 may present recommended item/offer information 227 (e.g., see 434d-3 in
In the above example, the location information included in the message 227 may be used to provide a store map, and directions to find the product item in the store floor plan (e.g., see
Continuing on with
In one implementation, the consumer may continue to checkout with a virtual wallet instantiated on the mobile device 203, e.g., see 444b
In one implementation, the V-GLASSES server 210 may optionally send a transaction confirmation message 241 to the merchant 220, wherein the transaction confirmation message 241 may have a data structure similar to the purchase receipt 245. The merchant 220 may confirm the completion of the purchase 242. In another implementation, as shown in
In one implementation, merchant store clerks 230a may be notified to their iPad 240 with the customer's loyalty profile. For example, in one implementation, the V-GLASSES server 210 may communicate with the merchant payment system 220a (e.g., PoS terminal) via a wallet API 251b to load consumer profile. In one implementation, the V-GLASSES server 210 may keep private consumer information anonymous from the merchant, e.g., consumer payment account information, address, telephone number, email addresses, and/or the like. In one implementation, the merchant payment system 220a may retrieve product inventory information from the merchant inventory system 220b, and provide such information to the PoS application of the sales clerk 230a. For example, the sales clerk may assist customer in shopping and adding items to iPad shopping cart (e.g., see 439 in
With reference to
In one implementation, the consumer may snap the QR code and generate a check-in message to the V-GLASSES server 310, which may receive the consumer check-in message 309 (e.g., see 208 in
In one implementation, if the consumer visit a merchant shopping site at 303, the consumer may similarly check-in with the merchant by snapping a QR code presented at the merchant site in a similar manner in 308-312. Alternatively, the consumer may log into a consumer account, e.g., a consumer account with the merchant, a consumer wallet account (e.g., V.me wallet payment account, etc.), to check-in with the merchant.
In one implementation, the merchant may receive consumer information from the V-GLASSES server (e.g., see 223 in
In one implementation, if a desired CSR match is not locally available 319 (e.g., not available at the merchant store, etc.), the V-GLASSES may expand the query to look for a remote CSR 321 which may communicate with the consumer via SMS, video chat, V-GLASSES push messages, etc., and allocate the CSR to the consumer based 322.
Alternatively, a pool of remote CSRs may be used to serve consumers and reduce overhead costs. In an alternative embodiment, online consumers may experience a store virtually by receiving a store floor plan for a designated location; and moving a consumer shopper avatar through the store floor plan to experience product offerings virtually, and the remote CSR may assist the virtual consumer, e.g., see
In one implementation, the consumer 302 may receive a check-in confirmation 324 (e.g., see 407 in
In another implementation, if the consumer has a transaction payment request (e.g., see 434g in
In one implementation, V-GLASSES server may receive the payment request from the consumer and may request PIN verification 341. For example, the V-GLASSES server may provide a PIN security challenge UI for the consumer to enter a PIN number 342, e.g., see 464 in
Continuing on with
In one implementation, the consumer may instantiate the shop 402c option, and check-in with a merchant store. For example, the consumer may operate the wallet application 403 to scan the merchant check-in QR code 404. Continuing on with
For example, with reference to
With reference to
In a further implementation, the consumer wallet may allow a consumer to include an image in the message with CSRs. In one implementation, the consumer may tap a camera icon 423 to snap a picture of an in-store advertisement, a front window display, a poster, etc., and submit the picture to the CSR to indicate the consumer's shopping interests. For example, the consumer may express interests in “Jeans” 427a, and may snap a picture of an in-store commercial poster of “men's jeans” 427b, and ask the CSR about “where to find” the jeans in display 427c.
With reference to
With reference to
With reference to
With reference to
In one implementation, upon submitting a “Pay” request 444b, the V-GLASSES may provide a PIN security challenge prior to payment processing to verify the consumer's identity. For example, the V-GLASSES may request a user to enter a PIN number 454 via a dial lock panel 455. In alternative implementations, as shown in
With reference to
With reference to
With reference to
With reference to
In another implementation, with reference to
In an alternative implementation, every aisle and shelving stack may include a numerous, wide-angle cameras having a specified accelerometer gyroscopic, positional/directional orientation, periodically taking a photograph of the opposing aisle/area, which may be submitted to the V-GLASSES server, so that the virtual store map may be continually updated and be kept up to date. For example, as shown in
In an alternative embodiment, such cameras may provide a continuous live video feed and still photos may be obtained from the live video frame grabs, which may be used to generate virtual store maps. In one implementation, a motion detection component may be used as a trigger to take still photos out of a live videos when the motion detection component detects no motion in the video and thereby provides unobstructed views for virtual map composition. In addition, when a consumer focuses on a particular shelf, aisle, stack, and/or region, e.g., a consumer turns their avatars parallel to a camera directional view, the consumer's view may then become filled with the live video feed of the camera closest to the consumer avatar's location.
In another implementation, as shown in
In further implementations, the consumer may be navigating a merchant's shopping site, having a shopping cart filled with product items, and the remote CSR may join the consumer's shopping session and provide assistance, allowing the CSR to provide the consumer with links to product items that may be of interests to the consumer; this may be achieved by having a CSR help/request button that may generate a pop-up window for audio/video chat with the CSR, and a dialogue box into which the CSR may place a link to the products. The consumer may click on the link provided by the CSR to be directed to a product page to view product details.
In a different embodiment, the users 611a-b may launch a V-GLASSES component instantiated on a camera-enabled mobile device 613a-103b to capture a view of the table, e.g., including the received invoice/bill 615 having a quick response (QR) code or barcode printed thereon, and a plurality of payment cards 619a-109b that the users 611a-b are going to pay for the bill. The users 611a-b may view virtual overlaid labels on top of the captured scene, so that they can tap on the option labels to split a bill equally, proportionally, and/or the like.
Within implementations, users 611a-b may facilitate payment from their payment cards upon V-GLASSES augmented reality capturing at the same mobile device/wallet. For example, user 611a may operate her mobile device 613a to capture a scene of the two payment cards 619a-b, while card 619b belongs to user 611b. In one implementation, the V-GLASSES component instantiated on the mobile device 613a may send an authorization request to a processing server, or a wallet management server to authorize split payment transaction on the payment card 613b. In such scenarios, users 611a-b may conduct a transaction including payments from two wallets on the same mobile device, without user 611b independently initiates a transaction using his mobile device 613b. Further implementations of restaurant bill payment scenarios are illustrated in
Within embodiments, the different layers 715a-d may comprise interdependent information. For example, merchant layer 715a and/or retailer layer 715b may provide information of related products based on user reviews from the social payer 715d. A variety of commerce participants, such as, but not limited to manufacturers, merchants, retailers, distributors, transaction processing networks, issuers, acquirers, payment gateway servers, and/or the like, may bid for layer space in the augmented reality shopping experience.
In one implementation, as shown at 719 in
As another example, a consumer 811 may walk into a merchant store and the mobile device 813 may capture the consumer's GPS coordinates 826. The V-GLASSES may then determine the consumer is located at a retailer shop based on the GPS coordinates 827, and may provide a retailer layer of augmented reality overlay labels 829 to the mobile device captured in-store scenes, e.g., including retailer discounts, in-store map, related products inventories, and/or the like.
For example, a user may move a sliding bar 907a to enable or disable a smart finger tip component 903a, e.g., when the smart finger tip component is enabled, the V-GLASSES may capture a human finger point within a captured reality scene (e.g., see also 912, etc.), etc. In one implementation, the smart finger tip component 903a may engage fingertip motion detection component (e.g., see
In another example, a user may move the sliding bar 907b to enable or disable auto card detection 903b, e.g., when the auto card detection component is enabled, the V-GLASSES may automatically detect and identify whether any rectangular object in a captured reality scene comprise a payment card, etc. In another example, a user may move the sliding bar 907c to enable or disable facial recognition 903c, e.g., when the facial recognition component is enabled, the V-GLASSES may automatically recognize human faces (e.g., including a human, a printed facial image on a magazine, a friend's picture displayed on a digital screen, etc.) that are presented in the reality scene and identify whether the human face matches with any of previously stored contacts. In another example, a user may move the sliding bar 907d to enable or disable smart bill tender component 903d, e.g., when the smart bill tender component is enabled, the V-GLASSES may provide option labels based on a type of the bill. When the bill is a restaurant bill, the V-GLASSES may provide options to facilitate tip calculation, bill splitting per actual consumption, and/or the like. In another example, a user may move the sliding bar 907e to enable or barcode reading component 903e, e.g., the V-GLASSES may read a barcode, and/or a QR code printed on a purchase label, invoice or bill to provide payment information via overlaid labels on the captured reality scene.
In one implementation, the user may configure a maximum one-time payment amount 904 via the V-GLASSES initiated transaction, e.g., by sliding the bar 905 to select a maximum amount of $500.00. In another implementation, a user may select to include social connections 906 into the V-GLASSES capturing component, e.g., the V-GLASSES may obtain social data such as user reviews, ratings with regard to a capture purchase item in the reality scene (see 1435 in
Within implementations, when a user places a camera-enabled mobile device (e.g., 913) to capture a reality scene, a user may view a plurality of virtual labels overlaid on top of the captured reality scene. For example, the user may view a sliding bar 910 to control whether to enable the smart finger tip component. As shown in
Continuing on with
With reference to
In one implementation, upon enrolling the card, the V-GLASSES may switch back to the visual capturing scene, with an overlaid notification showing the card is ready to use 926, and provide a plurality of overlaid option labels beneath the card 911, such as, but not limited to view balance 927a (e.g., a user may tap and see the current balance of the card), view history 927b (e.g., the user may tap and view recent transaction history associated with the card), transfer money from 927c (e.g., the user may select to transfer money from the card to another account), transfer money to 927d (e.g., the user may transfer money to the card from another account, etc.), pay shopping cart 927e (e.g., the user may engage the card to pay the current shopping cart 908a), and/or the like. Various other option labels related to the card may be contemplated.
In one implementation, if the user selects to tap on the “transfer $$ to” button 927d, with reference to
In one implementation, the user may move his finger to point to another card in the real scene so that the smart finger tip component may capture the payee card. In another implementation, as shown in
With reference to 9E, upon user confirming fund transfer, the V-GLASSES may provide a message notifying completion of the transaction 937, and the user may select to view the transaction receipt 938. In one implementation, the V-GLASSES may provide a virtual receipt 939 including a barcode 940 summarizing the transaction. In one implementation, the user may email 941 the virtual receipt (e.g., for reimbursement, etc.), or to earn points 942 from the transaction.
In one implementation, upon the facial recognition, the V-GLASSES may provide a plurality of option labels overlaid on top of the reality scene, so that the user may select to call the contact 1008a, send a SMS 1008b, email the contact 1008c, transfer funds to the contact 1008d, connect to the contact on social media 1008e, view the contact's published purchasing history 1008f, and/or the like. In one implementation, if the user selects to transfer money to the contact, the V-GLASSES may retrieve a previously stored account associated with the contact, or prompt the user to enter account information to facilitate the transfer.
With reference to
In another implementation, when the V-GLASSES determines the user tapped portion of the screen comprises a user's DMV license, 1113, the V-GLASSES may provide a plurality of option labels, such as view DMV profile 1114a, view pending tickets 1114b, pay ticket 1114c, file a dispute request 1114d, and/or the like.
With reference to
In another implementation, when the V-GLASSES determines the user tapped portion comprises a store membership card 1220, e.g., a PF Chang's card, the V-GLASSES may provide a plurality of labels including viewpoints 1221a, pay with the card 1221b, buy points 1221d-e, call to order 1221e, and/or the like.
With reference to
In another implementation, when the V-GLASSES determines the user tapped portion comprises a bill including a barcode 1326, e.g., a purchase invoice, a restaurant bill, a utility bill, a medical bill, etc., the V-GLASSES may provide a plurality of labels including view bill details 1327a, pay the bill 1327b, request extension 1327c, dispute bill 1327d, insurance reimbursement 1327e (e.g., for medical bills, etc.), and/or the like.
With reference to
In one implementation, if the user selects view social rating 1433f of the product, the V-GLASSES may retrieve social data from various social media platforms (e.g., Facebook, Twitter, Tumblr, etc.) related to the featured product, so that the user may review other users' comments related to the product.
In one implementation, when the user selects action consumption 1505c, the PVTC may provide tags of the consumed items 1507a-b, e.g., by reading the bill barcode 1502, or by performing OCR on the bill image, etc. In one implementation, a user may drag the item 1507a, e.g., a “bloody Mary” 1508 into the “I Pay” bowl 1510. The user may tap on the plus sign 1509 to increase quantity of the consumed item. In one implementation, the user may tap on a card 1511 to indicate pay with this card for the item in the “I Pay” bowl 1510 as summarized in label 1512. In one implementation, the V-GLASSES may provide option labels for tips, including suggested tip percentage (e.g., 15% or 20%) 1513 or enter tip amount 1514.
Continuing on with
With reference to
With reference to
Continuing on with
In one implementation, if the consumer elects “add card” 1542, the consumer may proceed with card enrollment in a similar manner as 215 in
Continuing on with
In one implementation, a user may place two payment cards in the scene so that the V-GLASSES may capture the cards. For example, the V-GLASSES may capture the type of the card, e.g., Visa 1608a and MasterCard 1608b, and provide labels to show rebate/rewards policy associated with each card for such a transaction 1609a-b. As such, the user may select to pay with a card to gain the provided rebate/rewards.
In an alternative embodiment, as shown in
With reference to
In another implementation, a consumer may slide the information layer 1611a to obtain another layer, e.g., retail information 1611b, social information 1611c, item information 1611d, and/or the like. For example, PVTC may capture a receipt and/or certificate in the scene, and provide information including other Cartier products 1618, purchase item description and price information 1615, retail store inventory information (e.g., stores where the purchase item is available) including physical stores 1623 and online shopping sites 1625, and/or the like.
In further embodiments, a consumer may tap on the provided virtual label of a “Cartier” store, e.g., 1613, 1623, etc., and be directed to a store map including inventory information, e.g., as shown in
With reference to
In another implementation, for additional captured objects 1630 in the scene (e.g., objects without textual contents, etc.), V-GLASSES may perform a pattern recognition to provide information of the recognized object 1630. For example, the pattern recognition may be correlated with other contexts within the scene to determine what the captured object is, e.g., the ring shaped object 1630 may be a piece of “Cartier” branded jewelry as the “Cartier” logo is captured in the same scene. In one implementation, the V-GLASSES may provide identified item information 1631 in a virtual label, and alternative item recognition information 1632, 1633, 1634. For example, for the ring-shaped product 1630, the V-GLASSES may recognize it as a “Cartier” branded bracelet 1631/1632, or ring shaped jewelry products of related brands 1633, 1634, and/or provide an option to the consumer to see more similar products 1635.
In one implementation, the V-GLASSES may automatically identify goods that are eligible for restricted-use accounts in a merchant store. For example, the V-GLASSES may allow a user to place a camera enabled device at a merchant store (e.g., scanning), and view a camera scene with augmented reality labels to indicate possible items eligible for a restricted-use account.
For example, in one implementation, when the user operate the camera enabled device to obtain a view inside the merchant store 1750, the user may also obtain augmented reality labels 1751 which identifies various products/items on the shelf, and show one or more possible eligible restricted-use accounts 1752. For example, over the counter drugs may be labeled as eligible for “FSA, HSA, HRA,” etc., 1752; grocery products may be eligible for food stamp usage; and infant food may be eligible for a children nutrition benefit account, and/or the like.
In further implementation, if the V-GLASSES does not automatically determine an item as eligible for any restricted-use accounts, e.g., an “Ester-C” supplement, a user may tap on the screen to select it, and may view a list of accounts 1863 to select a user desired reallocation account, e.g., any restricted-use account, loyalty account, and/or the like.
In further implementations, the V-GLASSES may identify a payment account that has been used to fulfill the transaction associated with the receipt, e.g., a Visa account 1866a, and/or obtain account information from the barcode printed on the receipt 1866b. In one implementation, the V-GLASSES may match the “*1234” Visa account with any of user's enrolled account in the wallet, and recommend the user to reimburse funds into an identified “Visa *1234” account if such account is identified from the wallet 1865. In another implementation, the V-GLASSES may prompt the user to select other accounts for depositing reimbursement funds 1865.
Continuing on with
In another implementation, if the user selects to tap on 1963 in
In one implementation, upon receiving user finger indication, the V-GLASSES may obtain an image of the scene (or the user finger pointed portion) 2006, e.g., grabbing a video frame, etc. In one implementation, the V-GLASSES may detect fingertip position within the video frame, and determine an object around the fingertip position for recognition 2007. The V-GLASSES may then perform OCR and/or pattern recognition on the obtained image (e.g., around the fingertip position) 2008 to determine a type of the object in the image 2010. For example, in one implementation, the V-GLASSES may start from the finger point and scan outwardly to perform edge detection so as to determine a contour of the object. The V-GLASSES may then perform OCR within the determined contour to determine a type of the object, e.g., whether there is card number presented 2011, whether there is a barcode or QR code presented 2012, whether there is a human face 2013, and/or the like.
In one implementation, if there is a payment card in the reality scene 2011, the V-GLASSES may determine a type of the card 2015 and the card number 2017. For example, the V-GLASSES may determine whether the card is a payment card (e.g., a credit card, a debit card, etc.), a membership card (e.g., a metro card, a store points card, a library card, etc.), a personal ID (e.g., a driver's license, etc.), an insurance card, and/or the like, based on the obtained textual content via OCR from the card. In one implementation, the V-GLASSES may query the user wallet for the card information to determine whether the card matches with any enrolled user account, and may generate and present overlay labels 2030 based on the type of the card (e.g., see overlay labels 927a-e for an identified Visa credit card 911 in
In another implementation, if there is a barcode and/or QR code detected within the reality scene 2012, the V-GLASSES may extract information from the barcode/QR code 2022, and determine a type of the object 2023, e.g., the barcode information may indicate whether the object comprises a purchase item, a bill, an invoice, and/or the like. In one implementation, the V-GLASSES may retrieve merchant information when the object comprises a purchase item, and/or biller information when the object comprises a bill 2028, and generate overlay labels accordingly, e.g., see overlay labels 1327a-e for an identified invoice 1326 in
In another implementation, if there is a human face detected from the reality scene 2013, the V-GLASSES may perform facial recognition to identify whether the presented human face matches with an existing contact 2024. In one implementation, the V-GLASSES may retrieve contact information if the contact is located from a contact list 2026, and/or add a new contact 2027 per user selection if the human face does not match with any existing contact record. The V-GLASSES may then generate and present overlay labels for the detected human face, e.g., see overlay labels 1008a-f for an identified face 1002 in
Upon user selection of the overlay labels, the V-GLASSES may proceed to transfer funds to an identified card, identified contact, and/or the like. The V-GLASSES may send financial transaction requests to an issuer network for processing, which may be performed in a similar manner as in
In one implementation, a consumer may engage in user interests indicative activities (e.g., web searches, wallet check-in, etc) 2031. For example, as shown in
In one implementation, when a consumer uses a mobile device to capture a reality scene (e.g., 2003/2004), V-GLASSES may determine a type of the object in the captured visual scene 2036, e.g., an item, card, barcode, receipt, etc. In one implementation, the V-GLASSES may retrieve stored user interest record 2038, and obtain information in the stored record. If the user interests record comprise a search term 2041, V-GLASSES may correlate the search term with product information 2044 (e.g., include price comparison information if the user is interested in finding the lowest price of a product, etc.), and generate an information layer for the virtual overlay 2049. In one implementation, the V-GLASSES may optionally capture mixed gestures within the captured reality scene 2029, e.g., consumer motion gestures, verbal gestures by articulating a command, etc. (see
In another implementation, if the user interests record comprise a real-time wallet check-in information 2042 of the consumer checking in at a retail store, the V-GLASSES may insert a retailer layer of virtual labels 2046 to the consumer device. In another implementation, the V-GLASSES may parse the user activity record for user interests indicators 2048 for other types of user activity data, e.g., browsing history, recent purchases, and/or the like, and determine an information layer of virtual overlay 2047. The consumer may obtain an automatically recommended injected layer of virtual label overlays 2050, and may switch to another layer of information labels by sliding on the layer, e.g., see 1611a-d in
As shown in
In one implementation, the V-GLASSES may determine whether the difference region has a “pointer” shape 2082, e.g., a fingertip, a pencil, etc. If not, e.g., the difference region may be noise caused by camera movement, etc., the V-GLASSES may determine whether the time lapse has exceeded a threshold. For example, if the V-GLASSES has been capturing the video scene for more than 10 seconds and detects no “pointer” shapes or “fingertip,” V-GLASSES may proceed to OCR/pattern recognition of the entire image 2087. Otherwise, the V-GLASSES may re-generate video frames at 2071.
In one implementation, if a “fingertip” or a “pointer” is detected at 2082, the V-GLASSES may determine a center point of the fingertip, e.g., by taking a middle point of the X and Y coordinates of the “fingertip.” The V-GLASSES may perform edge detection starting from the determined center point to determine the boundary of a consumer pointed object 2085. For example, the V-GLASSES may employ edge detection components such as, but not limited to Adobe Photoshop edge detection, Java edge detection package, and/or the like. Within implementations, upon V-GLASSES has defined boundaries of an object, the V-GLASSES may perform OCR and pattern recognition of the defined area 2088 to determine a type of the object.
Within implementations, V-GLASSES may receive information related to the determined object 2057 (e.g., 2018, 2027, 2028 in
For example, a data structure of a generated virtual label, substantially in the form of XML-formatted data, is provided below:
In the above example, the generated virtual label data structure includes fields such as size of the video frame, the captured object (e.g., the object is a barcode, etc.), information to be included in the virtual label, orientation of the label, format of the virtual label (e.g., template, font, background, transparency, etc.), injection position of the label, and/or the like. In one implementation, the virtual label may contain an informational link, e.g., for the product information in the above example, an Amazon link may be provided, etc. In one implementation, the injection position may be determined based on the position of the object (e.g., X, Y coordinates of the area on the image, determined by a barcode detector, etc.).
In some implementations, an exemplary XML-encoded command message 2208 may take a form similar to the following:
d rXYZ x gXYZ
bXYZ rTRC
{acute over ( )} aarg À vcgt ...
In some implementations, the electronic device may reduce the size of the vocal file by cropping the audio file to when the user begins and ends the vocal command. In some implementations, the V-GLASSES may process the gesture and audio data 2210 in order to determine the type of gesture performed, as well as the words spoken by the user. In some implementations, a composite gesture generated from the processing of the gesture and audio data may be embodied in an XML-encoded data structure similar to the following:
mntrRGB XYZ Ü $ acspAPPL öÖÓ-appl
d rXYZ x gXYZ
bXYZ rTRC
{acute over ( )} aarg À vcgt ...
mntrRGB XYZ Ü $ acspAPPL öÖÓ-appl
d rXYZ x gXYZ
bXYZ rTRC
{acute over ( )} aarg À vcgt ...
In some implementations, fields in the composite gesture data structure may be left blank depending on whether the particular gesture type (e.g., finger gesture, object gesture, and/or the like) has been made. The V-GLASSES may then match 2211 the gesture and the words to the various possible gesture types stored in the V-GLASSES database. In some implementations, the V-GLASSES may query the database for particular disparate gestures in a manner similar to the following:
In some implementations, the result of each query in the above example may be used to search for the composite gesture in the Multi-Disparate Gesture Action (MDGA) table of the database. For example, if $fingerresult is “tap check,” $objectresult is “swipe,” and $voiceresult is “pay total of check with this payment device,” V-GLASSES may search the MDGA table using these three results to narrow down the precise composite action that has been performed. If a match is found, the V-GLASSES may request confirmation that the right action was found, and then may perform the action 2212 using the user's account. In some implementations, the V-GLASSES may access the user's financial information and account 2213 in order to perform the action. In some implementations, V-GLASSES may update a gesture table 2214 in the V-GLASSES database 2215 to refine models for usable gestures based on the user's input, to add new gestures the user has invented, and/or the like. In some implementations, an update 2214 for a finger gesture may be performed via a PHP/MySQL command similar to the following:
After successfully updating the table 2216, the V-GLASSES may send the user to a confirmation page 2217 (or may provide an augmented reality (AR) overlay to the user) which may indicate that the action was successfully performed. In some implementations, the AR overlay may be provided to the user through use of smart glasses, contacts, and/or a like device (e.g. Google Glasses).
As shown in
The V-GLASSES may then perform the action specified 2221, accessing any information necessary to conduct the action 2222, and may send a confirmation page or AR overlay to the user 2223. In some implementations, the XML-encoded data structure for the AR overlay may take a form similar to the following:
In some implementations, V-GLASSES may receive 2305 the data from the user's device, which may include accelerometer and/or gyroscope data pertaining to the gesture, a video and/or images of the gesture, an audio file of the vocal command, and/or the like. In some implementations, V-GLASSES may determine what sort of data was sent by the user's device in order to determine how to process it. For example, if the user's device provides accelerometer and/or gyroscope data 2306, V-GLASSES may determine the gesture performed by matching the accelerometer and/or gyroscope data points with pre-determined mathematical gesture models 2309. For example, if a particular gesture would generate accelerometer and/or gyroscope data that would fit a linear gesture model, V-GLASSES will determine whether the received accelerometer and/or gyroscope data matches a linear model.
If the user's device provides a video and/or images of the gesture 2307, V-GLASSES may use an image processing component in order to process the video and/or images 2310 and determine what the gesture is. In some implementations, if a video is provided, the video may also be used to determine the vocal command provided by the user. As shown in
If the user's device provides an audio file 2308, then V-GLASSES may determine the vocal command given using an audio analytics component 2311. In some implementations, the audio analytics component may process the audio file and produce a text translation of the vocal command. As discussed above, in some implementations, the audio analytics component may also use a video, if provided, as input to produce a text translation of the user's vocal command.
As shown in
If the action is a single-party payment-related action 2318 (i.e., concerning one person and/or entity transferring funds to his/her/itself), V-GLASSES may retrieve the account information of the one user 2319, and may use it to access the relevant financial and/or other accounts associated in the transaction. For example, if one user is transferring funds from a bank account to a refillable gift card owned by the same user, then V-GLASSES would access the user's account in order to obtain information about both the bank account and the gift card, and would use the information to transfer funds from the bank account to the gift card 2320.
In either the multi-party or the single-party action, V-GLASSES may update 2321 the data of the affected accounts (including: saving a record of the transaction, which may include to whom the money was given to, the date and time of the transaction, the size of the transaction, and/or the like), and may send a confirmation of this update 2322 to the user.
If the action is related to obtaining information about a product and/or service 2323, V-GLASSES may send a request 2324 to the relevant merchant database(s) in order to get information about the product and/or service the user would like to know more about. V-GLASSES may provide any information obtained from the merchant to the user 2325. In some implementations, V-GLASSES may provide the information via an AR overlay, or via an information page or pop-up which displays all the retrieved information.
mntrRGB XYZ Ü $ acspAPPL öÖÓ-appl
d rXYZ x gXYZ
bXYZ rTRC
{acute over ( )} aarg À vcgt ...
In some implementations, the user, while shopping through the store, may also scan 2407 items with the user's electronic device, in order to obtain more information about them, in order to add them to the user's cart, and/or the like. In such implementations, the user's electronic device may send a scanned item message 2408 to the V-GLASSES server. In some implementations, an exemplary XML-encoded scanned item message 2408 may take a form similar to the following:
mntrRGB XYZ Ü $ acspAPPL öÖÓ-appl
d rXYZ x gXYZ
bXYZ rTRC
{acute over ( )} aarg À vcgt ...
In some implementations, V-GLASSES may then determine the location 2409 of the user based on the location of the scanned item, and may send a notification 2410 to a sale's representative 2411 indicating that a user has checked into the store and is browsing items in the store. In some implementations, an exemplary XML-encoded notification message 2410 may comprise of the scanned item message of scanned item message 2408.
The sale's representative may use the information in the notification message to determine products and/or services to recommend 2412 to the user, based on the user's profile, location in the store, items scanned, and/or the like. Once the sale's representative has chosen at least one product and/or service to suggest, it may send the suggestion 2413 to the V-GLASSES server. In some implementations, an exemplary XML-encoded suggestion 2413 may take a form similar to the following:
mntrRGB XYZ Ü $ acspAPPL öÖÓ-appl
d rXYZ x gXYZ
bXYZ rTRC
{acute over ( )} aarg À vcgt ...
In some implementations, V-GLASSES may also use the user's profile information, location, scanned items, and/or the like to determine its own products and/or services to recommend 2414 to the user. In some implementations, V-GLASSES may determine where in the store any suggested product and/or service is 2415, based on aisle information in the item data structure, and may generate a map from the user's location to the location of the suggested product and/or service. In some implementations, the map overlays a colored path on a store map from the user's location to the suggested product and/or service. V-GLASSES may send 2416 this map, along with the suggested product and/or item, to the user, who may use it to find the suggested item, and add the suggested item to its shopping cart 2440 if the user would like to purchase it.
@ $wtpt d rXYZ
x gXYZ bXYZ
rTRC {acute over ( )} aarg À vcgt ...
In some implementations, V-GLASSES may conduct its own analysis of the user based on the photo 2427, including analyzing the image to determine the user's body size, body shape, complexion, and/or the like. In some implementations, V-GLASSES may use these attributes, along with any provided through the apparel preview request, to search the database 2428 for clothing that matches the user's attributes and search criteria. In some implementations, V-GLASSES may also update 2429 the user's attributes stored in the database, based on the attributes provided in the apparel preview request or based on V-GLASSES' analysis of the user's photo. After V-GLASSES receives confirmation that the update is successful 2430, V-GLASSES may send a virtual closet 2431 to the user, comprising a user interface for previewing clothing, accessories, and/or the like chosen for the user based on the user's attributes and search criteria. In some implementations, the virtual closet may be implemented via HTML and Javascript.
In some implementations, as shown in
In some implementations, the user may provide a selection of at least one article of clothing to try on 2521, prompting V-GLASSES to determine body and/or joint locations and markers in the user photo 2522, and to scale the image of the article of clothing to match the user image 2523, based on those body and/or joint locations and markers. In some implementations, V-GLASSES may also format the clothing image 2524, including altering shadows in the image, blurring the image, and/or the like, in order to match the look of the clothing image to the look of the user image. V-GLASSES may superimpose 2525 the clothing image on the user image to allow the user to virtually preview the article of clothing on the user, and may allow the user to change options such as the clothing color, size, and/or the like while the article of clothing is being previewed on the user. In some implementations, V-GLASSES may receive a request to purchase at least one article of clothing 2526, and may retrieve user information 2527, including the user's ID, shipping address, and/or the like. V-GLASSES may further retrieve the user's payment information 2528, including the user's preferred payment device or account, and/or the like, and may contact the user's issuer (and that of the merchant) 2529 in order to process the transaction. V-GLASSES may send a confirmation to the user when the transaction is completed 2530.
As shown in
As shown in
In one embodiment, for example, a user may select the option current items 3215, as shown in the left most user interface of
With reference to
With reference to
In one implementation, a user may select Joe P. for payment. Joe P., as shown in the user interface, has an email icon 3217g next to his name indicating that Joe P. accepts payment via email. When his name is selected, the user interface may display his contact information such as email, phone, etc. If a user wishes to make a payment to Joe P. by a method other than email, the user may add another transfer mode 3217j to his contact information and make a payment transfer. With reference to
With reference to
With reference to
With reference to
In one implementation, the user may combine funds from multiple sources to pay for the transaction. The amount 3315 displayed on the user interface may provide an indication of the amount of total funds covered so far by the selected forms of payment (e.g., Discover card and rewards points). The user may choose another form of payment or adjust the amount to be debited from one or more forms of payment until the amount 3315 matches the amount payable 3314. Once the amounts to be debited from one or more forms of payment are finalized by the user, payment authorization may begin.
In one implementation, the user may select a secure authorization of the transaction by selecting the cloak button 3322 to effectively cloak or anonymize some (e.g., pre-configured) or all identifying information such that when the user selects pay button 3321, the transaction authorization is conducted in a secure and anonymous manner. In another implementation, the user may select the pay button 3321 which may use standard authorization techniques for transaction processing. In yet another implementation, when the user selects the social button 3323, a message regarding the transaction may be communicated to one of more social networks (set up by the user) which may post or announce the purchase transaction in a social forum such as a wall post or a tweet. In one implementation, the user may select a social payment processing option 3323. The indicator 3324 may show the authorizing and sending social share data in progress.
In another implementation, a restricted payment mode 3325 may be activated for certain purchase activities such as prescription purchases. The mode may be activated in accordance with rules defined by issuers, insurers, merchants, payment processor and/or other entities to facilitate processing of specialized goods and services. In this mode, the user may scroll down the list of forms of payments 3326 under the funds tab to select specialized accounts such as a flexible spending account (FSA) 3327, health savings account (HAS), and/or the like and amounts to be debited to the selected accounts. In one implementation, such restricted payment mode 1925 processing may disable social sharing of purchase information.
In one embodiment, the wallet mobile application may facilitate importing of funds via the import funds user interface 3328. For example, a user who is unemployed may obtain unemployment benefit fund 3329 via the wallet mobile application. In one implementation, the entity providing the funds may also configure rules for using the fund as shown by the processing indicator message 3330. The wallet may read and apply the rules prior, and may reject any purchases with the unemployment funds that fail to meet the criteria set by the rules. Example criteria may include, for example, merchant category code (MCC), time of transaction, location of transaction, and/or the like. As an example, a transaction with a grocery merchant having MCC 5411 may be approved, while a transaction with a bar merchant having an MCC 5813 may be refused.
With reference to
Similarly, when a German user operates a wallet in Germany, the mobile wallet application user interface may be dynamically updated to reflect the country of operation 3332 and the currency 3334. In a further implementation, the wallet application may rearrange the order in which different forms of payment 3336 are listed based on their acceptance level in that country. Of course, the order of these forms of payments may be modified by the user to suit his or her own preferences.
With reference to
With reference to
With reference to
With reference to
In one implementation, the user may select a transaction, for example transaction 3415, to view the details of the transaction. For example, the user may view the details of the items associated with the transaction and the amounts 3416 of each item. In a further implementation, the user may select the show option 3417 to view actions 3418 that the user may take in regards to the transaction or the items in the transaction. For example, the user may add a photo to the transaction (e.g., a picture of the user and the iPad the user bought). In a further implementation, if the user previously shared the purchase via social channels, a post including the photo may be generated and sent to the social channels for publishing. In one implementation, any sharing may be optional, and the user, who did not share the purchase via social channels, may still share the photo through one or more social channels of his or her choice directly from the history mode of the wallet application. In another implementation, the user may add the transaction to a group such as company expense, home expense, travel expense or other categories set up by the user. Such grouping may facilitate year-end accounting of expenses, submission of work expense reports, submission for value added tax (VAT) refunds, personal expenses, and/or the like. In yet another implementation, the user may buy one or more items purchased in the transaction. The user may then execute a transaction without going to the merchant catalog or site to find the items. In a further implementation, the user may also cart one or more items in the transaction for later purchase.
The history mode, in another embodiment, may offer facilities for obtaining and displaying ratings 3419 of the items in the transaction. The source of the ratings may be the user, the user's friends (e.g., from social channels, contacts, etc.), reviews aggregated from the web, and/or the like. The user interface in some implementations may also allow the user to post messages to other users of social channels (e.g., TWITTER or FACEBOOK). For example, the display area 3420 shows FACEBOOK message exchanges between two users. In one implementation, a user may share a link via a message 3421. Selection of such a message having embedded link to a product may allow the user to view a description of the product and/or purchase the product directly from the history mode.
In one embodiment, the history mode may also include facilities for exporting receipts. The export receipts pop up 3422 may provide a number of options for exporting the receipts of transactions in the history. For example, a user may use one or more of the options 3425, which include save (to local mobile memory, to server, to a cloud account, and/or the like), print to a printer, fax, email, and/or the like. The user may utilize his or her address book 3423 to look up email or fax number for exporting. The user may also specify format options 3424 for exporting receipts. Example format options may include, without limitation, text files (.doc, .txt, .rtf, iif, etc.), spreadsheet (.csv, .xls, etc.), image files (.jpg, .tff, .png, etc.), portable document format (.pdf), postscript (.ps), and/or the like. The user may then click or tap the export button 3427 to initiate export of receipts.
With reference to
As shown, the user may enter a search term (e.g., bills) in the search bar 2121. The user may then identify in the tab 3522 the receipt 3523 the user wants to reallocate. Alternatively, the user may directly snap a picture of a barcode on a receipt, and the snap mode may generate and display a receipt 3523 using information from the barcode. The user may now reallocate 3525. In some implementations, the user may also dispute the transaction 3524 or archive the receipt 3526.
In one implementation, when the reallocate button 3525 is selected, the wallet application may perform optical character recognition (OCR) of the receipt. Each of the items in the receipt may then be examined to identify one or more items which could be charged to which payment device or account for tax or other benefits such as cash back, reward points, etc. In this example, there is a tax benefit if the prescription medication charged to the user's Visa card is charged to the user's FSA. The wallet application may then perform the reallocation as the back end. The reallocation process may include the wallet contacting the payment processor to credit the amount of the prescription medication to the Visa card and debit the same amount to the user's FSA account. In an alternate implementation, the payment processor (e.g., Visa or MasterCard) may obtain and OCR the receipt, identify items and payment accounts for reallocation and perform the reallocation. In one implementation, the wallet application may request the user to confirm reallocation of charges for the selected items to another payment account. The receipt 3527 may be generated after the completion of the reallocation process. As discussed, the receipt shows that some charges have been moved from the Visa account to the FSA.
With reference to
In one implementation, the user may decide to pay with default 3534. The wallet application may then use the user's default method of payment, in this example the wallet, to complete the purchase transaction. Upon completion of the transaction, a receipt may be automatically generated for proof of purchase. The user interface may also be updated to provide other options for handling a completed transaction. Example options include social 3537 to share purchase information with others, reallocate 3538 as discussed with regard to
With reference to
In one implementation, after the offer or coupon 3546 is applied, the user may have the option to find qualifying merchants and/or products using find, the user may go to the wallet using 3548, and the user may also save the offer or coupon 3546 for later use.
With reference to
For example, a user may go to doctor's office and desire to pay the co-pay for doctor's appointment. In addition to basic transactional information such as account number and name, the app may provide the user the ability to select to transfer medical records, health information, which may be provided to the medical provider, insurance company, as well as the transaction processor to reconcile payments between the parties. In some implementations, the records may be sent in a Health Insurance Portability and Accountability Act (HIPAA)-compliant data format and encrypted, and only the recipients who are authorized to view such records may have appropriate decryption keys to decrypt and view the private user information.
With reference to
In some implementations, the V-GLASSES may utilize a text challenge procedure to verify the authenticity of the user, e.g., 3725. For example, the V-GLASSES may communicate with the user via text chat, SMS messages, electronic mail, Facebook® messages, Twitter™ tweets, and/or the like. The V-GLASSES may pose a challenge question, e.g., 3726, for the user. The app may provide a user input interface element(s) (e.g., virtual keyboard 3728) to answer the challenge question posed by the V-GLASSES. In some implementations, the challenge question may be randomly selected by the V-GLASSES automatically; in some implementations, a customer service representative may manually communicate with the user. In some implementations, the user may not have initiated the transaction, e.g., the transaction is fraudulent. In such implementations, the user may cancel the text challenge. The V-GLASSES may cancel the transaction, and/or initiate fraud investigation on behalf of the user.
In some embodiments, the merchant server may obtain the checkout request from the client, and extract the checkout detail (e.g., XML data) from the checkout request. For example, the merchant server may utilize a parser such as the example parsers described below in the discussion with reference to
In some embodiments, in response to obtaining the product data, the merchant server may generate, e.g., 3816, checkout data to provide for the PoS client. In some embodiments, such checkout data, e.g., 3817, may be embodied, in part, in a HyperText Markup Language (“HTML”) page including data for display, such as product detail, product pricing, total pricing, tax information, shipping information, offers, discounts, rewards, value-added service information, etc., and input fields to provide payment information to process the purchase transaction, such as account holder name, account number, billing address, shipping address, tip amount, etc. In some embodiments, the checkout data may be embodied, in part, in a Quick Response (“QR”) code image that the PoS client can display, so that the user may capture the QR code using a user's device to obtain merchant and/or product data for generating a purchase transaction processing request. In some embodiments, a user alert mechanism may be built into the checkout data. For example, the merchant server may embed a URL specific to the transaction into the checkout data. In some embodiments, the alerts URL may further be embedded into optional level 3 data in card authorization requests, such as those discussed further below with reference to
Upon obtaining the checkout data, e.g., 3817, the PoS client may render and display, e.g., 3818, the checkout data for the user.
In some embodiments, upon authenticating the user for access to virtual wallet features, the user wallet device may provide a transaction authorization input, e.g., 4014, to a point-of-sale (“PoS”) client, e.g., 4002. For example, the user wallet device may communicate with the PoS client via Bluetooth, Wi-Fi, cellular communication, one- or two-way near-field communication (“NFC”), and/or the like. In embodiments where the user utilizes a plastic card instead of the user wallet device, the user may swipe the plastic card at the PoS client to transfer information from the plastic card into the PoS client. For example, the PoS client may obtain, as transaction authorization input 4014, track 1 data from the user's plastic card (e.g., credit card, debit card, prepaid card, charge card, etc.), such as the example track 1 data provided below:
In embodiments where the user utilizes a user wallet device, the user wallet device may provide payment information to the PoS client, formatted according to a data formatting protocol appropriate to the communication mechanism employed in the communication between the user wallet device and the PoS client. An example listing of transaction authorization input 4014, substantially in the form of XML-formatted data, is provided below:
In some embodiments, the PoS client may generate a card authorization request, e.g., 4015, using the obtained transaction authorization input from the user wallet device, and/or product/checkout data (see, e.g.,
In some embodiments, the card authorization request generated by the user device may include a minimum of information required to process the purchase transaction. For example, this may improve the efficiency of communicating the purchase transaction request, and may also advantageously improve the privacy protections provided to the user and/or merchant. For example, in some embodiments, the card authorization request may include at least a session ID for the user's shopping session with the merchant. The session ID may be utilized by any component and/or entity having the appropriate access authority to access a secure site on the merchant server to obtain alerts, reminders, and/or other data about the transaction(s) within that shopping session between the user and the merchant. In some embodiments, the PoS client may provide the generated card authorization request to the merchant server, e.g., 4016. The merchant server may forward the card authorization request to a pay gateway server, e.g., 4004a, for routing the card authorization request to the appropriate payment network for payment processing. For example, the pay gateway server may be able to select from payment networks, such as Visa, Mastercard, American Express, Paypal, etc., to process various types of transactions including, but not limited to: credit card, debit card, prepaid card, B2B and/or like transactions. In some embodiments, the merchant server may query a database, e.g., merchant/acquirer database 4003b, for a network address of the payment gateway server, for example by using a portion of a user payment card number, or a user ID (such as an email address) as a keyword for the database query. For example, the merchant server may issue PHP/SQL commands to query a database table (such as
In response, the merchant/acquirer database may provide the requested payment gateway address, e.g., 4018. The merchant server may forward the card authorization request to the pay gateway server using the provided address, e.g., 4019. In some embodiments, upon receiving the card authorization request from the merchant server, the pay gateway server may invoke a component to provide one or more services associated with purchase transaction authorization. For example, the pay gateway server may invoke components for fraud prevention, loyalty and/or rewards, and/or other services for which the user-merchant combination is authorized. The pay gateway server may forward the card authorization request to a pay network server, e.g., 4005a, for payment processing. For example, the pay gateway server may be able to select from payment networks, such as Visa, Mastercard, American Express, Paypal, etc., to process various types of transactions including, but not limited to: credit card, debit card, prepaid card, B2B and/or like transactions. In some embodiments, the pay gateway server may query a database, e.g., pay gateway database 4004b, for a network address of the payment network server, for example by using a portion of a user payment card number, or a user ID (such as an email address) as a keyword for the database query. For example, the pay gateway server may issue PHP/SQL commands to query a database table (such as
In response, the payment gateway database may provide the requested payment network address, e.g., 4022. The pay gateway server may forward the card authorization request to the pay network server using the provided address, e.g., 4023.
With reference to
In some embodiments, the pay network server may generate a query, e.g., 4024, for issuer server(s) corresponding to the user-selected payment options. For example, the user's account may be linked to one or more issuer financial institutions (“issuers”), such as banking institutions, which issued the account(s) for the user. For example, such accounts may include, but not be limited to: credit card, debit card, prepaid card, checking, savings, money market, certificates of deposit, stored (cash) value accounts and/or the like. Issuer server(s), e.g., 4006a, of the issuer(s) may maintain details of the user's account(s). In some embodiments, a database, e.g., pay network database 4005b, may store details of the issuer server(s) associated with the issuer(s). In some embodiments, the pay network server may query a database, e.g., pay network database 4005b, for a network address of the issuer(s) server(s), for example by using a portion of a user payment card number, or a user ID (such as an email address) as a keyword for the database query. For example, the merchant server may issue PHP/SQL commands to query a database table (such as
In response to obtaining the issuer server query, e.g., 4024, the pay network database may provide, e.g., 4025, the requested issuer server data to the pay network server. In some embodiments, the pay network server may utilize the issuer server data to generate funds authorization request(s), e.g., 4026, for each of the issuer server(s) selected based on the pre-defined payment settings associated with the user's virtual wallet, and/or the user's payment options input, and provide the funds authorization request(s) to the issuer server(s). In some embodiments, the funds authorization request(s) may include details such as, but not limited to: the costs to the user involved in the transaction, card account details of the user, user billing and/or shipping information, and/or the like. An example listing of a funds authorization request 4026, substantially in the form of a HTTPS(S) POST message including XML-formatted data, is provided below:
In some embodiments, an issuer server may parse the authorization request(s), and based on the request details may query a database, e.g., user profile database 4006b, for data associated with an account linked to the user. For example, the merchant server may issue PHP/SQL commands to query a database table (such as
In some embodiments, on obtaining the user account(s) data, e.g., 4028, the issuer server may determine whether the user can pay for the transaction using funds available in the account, 4029. For example, the issuer server may determine whether the user has a sufficient balance remaining in the account, sufficient credit associated with the account, and/or the like. Based on the determination, the issuer server(s) may provide a funds authorization response, e.g., 4030, to the pay network server. For example, the issuer server(s) may provide a HTTP(S) POST message similar to the examples above. In some embodiments, if at least one issuer server determines that the user cannot pay for the transaction using the funds available in the account, the pay network server may request payment options again from the user (e.g., by providing an authorization fail message to the user device and requesting the user device to provide new payment options), and re-attempt authorization for the purchase transaction. In some embodiments, if the number of failed authorization attempts exceeds a threshold, the pay network server may abort the authorization process, and provide an “authorization fail” message to the merchant server, user device and/or client.
In some embodiments, the pay network server may obtain the funds authorization response including a notification of successful authorization, and parse the message to extract authorization details. Upon determining that the user possesses sufficient funds for the transaction, e.g., 4031, the pay network server may invoke a component to provide value-add services for the user.
In some embodiments, the pay network server may generate a transaction data record from the authorization request and/or authorization response, and store the details of the transaction and authorization relating to the transaction in a transactions database. For example, the pay network server may issue PHP/SQL commands to store the data to a database table (such as
In some embodiments, the pay network server may forward a transaction authorization response, e.g., 4032, to the user wallet device, PoS client, and/or merchant server. The merchant may obtain the transaction authorization response, and determine from it that the user possesses sufficient funds in the card account to conduct the transaction. The merchant server may add a record of the transaction for the user to a batch of transaction data relating to authorized transactions. For example, the merchant may append the XML data pertaining to the user transaction to an XML data file comprising XML data for transactions that have been authorized for various users, e.g., 4033, and store the XML data file, e.g., 4034, in a database, e.g., merchant database 404. For example, a batch XML data file may be structured similar to the example XML data structure template provided below:
In some embodiments, the server may also generate a purchase receipt, e.g., 4033, and provide the purchase receipt to the client, e.g., 4035. The client may render and display, e.g., 4036, the purchase receipt for the user. In some embodiments, the user's wallet device may also provide a notification of successful authorization to the user. For example, the PoS client/user device may render a webpage, electronic message, text/SMS message, buffer a voicemail, emit a ring tone, and/or play an audio message, etc., and provide output including, but not limited to: sounds, music, audio, video, images, tactile feedback, vibration alerts (e.g., on vibration-capable client devices such as a smartphone etc.), and/or the like.
In some embodiments, upon authenticating the user for access to virtual wallet features, the user wallet device may provide a transaction authorization input, e.g., 4104, to a point-of-sale (“PoS”) client. For example, the user wallet device may communicate with the PoS client via Bluetooth, Wi-Fi, cellular communication, one- or two-way near-field communication (“NFC”), and/or the like. In embodiments where the user utilizes a plastic card instead of the user wallet device, the user may swipe the plastic card at the PoS client to transfer information from the plastic card into the PoS client. In embodiments where the user utilizes a user wallet device, the user wallet device may provide payment information to the PoS client, formatted according to a data formatting protocol appropriate to the communication mechanism employed in the communication between the user wallet device and the PoS client.
In some embodiments, the PoS client may obtain the transaction authorization input, and parse the input to extract payment information from the transaction authorization input, e.g., 4105. For example, the PoS client may utilize a parser, such as the example parsers provided below in the discussion with reference to
In some embodiments, the PoS client may provide the generated card authorization request to the merchant server. The merchant server may forward the card authorization request to a pay gateway server, for routing the card authorization request to the appropriate payment network for payment processing. For example, the pay gateway server may be able to select from payment networks, such as Visa, Mastercard, American Express, Paypal, etc., to process various types of transactions including, but not limited to: credit card, debit card, prepaid card, B2B and/or like transactions. In some embodiments, the merchant server may query a database, e.g., 4108, for a network address of the payment gateway server, for example by using a portion of a user payment card number, or a user ID (such as an email address) as a keyword for the database query. In response, the merchant/acquirer database may provide the requested payment gateway address, e.g., 4110. The merchant server may forward the card authorization request to the pay gateway server using the provided address. In some embodiments, upon receiving the card authorization request from the merchant server, the pay gateway server may invoke a component to provide one or more service associated with purchase transaction authorization, e.g., 4111. For example, the pay gateway server may invoke components for fraud prevention (see e.g., VerifyChat,
The pay gateway server may forward the card authorization request to a pay network server for payment processing, e.g., 4114. For example, the pay gateway server may be able to select from payment networks, such as Visa, Mastercard, American Express, Paypal, etc., to process various types of transactions including, but not limited to: credit card, debit card, prepaid card, B2B and/or like transactions. In some embodiments, the pay gateway server may query a database, e.g., 4112, for a network address of the payment network server, for example by using a portion of a user payment card number, or a user ID (such as an email address) as a keyword for the database query. In response, the payment gateway database may provide the requested payment network address, e.g., 4113. The pay gateway server may forward the card authorization request to the pay network server using the provided address, e.g., 4114.
With reference to
In response to obtaining the issuer server query, the pay network database may provide, e.g., 4116, the requested issuer server data to the pay network server. In some embodiments, the pay network server may utilize the issuer server data to generate funds authorization request(s), e.g., 4117, for each of the issuer server(s) selected based on the pre-defined payment settings associated with the user's virtual wallet, and/or the user's payment options input, and provide the funds authorization request(s) to the issuer server(s). In some embodiments, the funds authorization request(s) may include details such as, but not limited to: the costs to the user involved in the transaction, card account details of the user, user billing and/or shipping information, and/or the like. In some embodiments, an issuer server may parse the authorization request(s), e.g., 4118, and based on the request details may query a database, e.g., 4119, for data associated with an account linked to the user.
In some embodiments, on obtaining the user account(s) data, e.g., 4120, the issuer server may determine whether the user can pay for the transaction using funds available in the account, e.g., 4121. For example, the issuer server may determine whether the user has a sufficient balance remaining in the account, sufficient credit associated with the account, and/or the like. Based on the determination, the issuer server(s) may provide a funds authorization response, e.g., 4122, to the pay network server. In some embodiments, if at least one issuer server determines that the user cannot pay for the transaction using the funds available in the account, the pay network server may request payment options again from the user (e.g., by providing an authorization fail message to the user device and requesting the user device to provide new payment options), and re-attempt authorization for the purchase transaction. In some embodiments, if the number of failed authorization attempts exceeds a threshold, the pay network server may abort the authorization process, and provide an “authorization fail” message to the merchant server, user device and/or client.
In some embodiments, the pay network server may obtain the funds authorization response including a notification of successful authorization, and parse the message to extract authorization details. Upon determining that the user possesses sufficient funds for the transaction, e.g., 4123, the pay network server may invoke a component to provide value-add services for the user, e.g., 4123.
In some embodiments, the pay network server may forward a transaction authorization response to the user wallet device, PoS client, and/or merchant server. The merchant may parse, e.g., 4124, the transaction authorization response, and determine from it that the user possesses sufficient funds in the card account to conduct the transaction, e.g., 4125, option “Yes.” The merchant server may add a record of the transaction for the user to a batch of transaction data relating to authorized transactions. For example, the merchant may append the XML data pertaining to the user transaction to an XML data file comprising XML data for transactions that have been authorized for various users, e.g., 4126, and store the XML data file, e.g., 4127, in a database. In some embodiments, the server may also generate a purchase receipt, e.g., 4128, and provide the purchase receipt to the client. The client may render and display, e.g., 4129, the purchase receipt for the user. In some embodiments, the user's wallet device may also provide a notification of successful authorization to the user. For example, the PoS client/user device may render a webpage, electronic message, text/SMS message, buffer a voicemail, emit a ring tone, and/or play an audio message, etc., and provide output including, but not limited to: sounds, music, audio, video, images, tactile feedback, vibration alerts (e.g., on vibration-capable client devices such as a smartphone etc.), and/or the like.
With reference to
In some embodiments, the issuer server may generate a payment command, e.g., 4227. For example, the issuer server may issue a command to deduct funds from the user's account (or add a charge to the user's credit card account). The issuer server may issue a payment command, e.g., 4227, to a database storing the user's account information, e.g., user profile database 4206b. The issuer server may provide an individual payment confirmation, e.g., 4228, to the pay network server, which may forward, e.g., 4229, the funds transfer message to the acquirer server. An example listing of an individual payment confirmation 4228, substantially in the form of a HTTP(S) POST message including XML-formatted data, is provided below:
In some embodiments, the acquirer server may parse the individual payment confirmation, and correlate the transaction (e.g., using the request_ID field in the example above) to the merchant. The acquirer server may then transfer the funds specified in the funds transfer message to an account of the merchant. For example, the acquirer server may query, e.g. 4230, an acquirer database 4207b for payment ledger and/or merchant account data, e.g., 4231. The acquirer server may utilize payment ledger and/or merchant account data from the acquirer database, along with the individual payment confirmation, to generate updated payment ledger and/or merchant account data, e.g., 4232. The acquirer server may then store, e.g., 4233, the updated payment ledger and/or merchant account data to the acquire database.
The pay network server may parse the batch payment request obtained from the acquirer server, and extract the transaction data for each transaction stored in the batch payment request, e.g., 4308. The pay network server may store the transaction data, e.g., 4309, for each transaction in a pay network database. In some embodiments, the pay network server may invoke a component, e.g., 4310, to provide analytics based on the transactions of the merchant for whom purchase transaction are being cleared.
With reference to
In some embodiments, the acquirer server may parse the individual payment confirmation, and correlate the transaction (e.g., using the request_ID field in the example above) to the merchant. The acquirer server may then transfer the funds specified in the funds transfer message to an account of the merchant. For example, the acquirer server may query, e.g. 4319, an acquirer database for payment ledger and/or merchant account data, e.g., 4320. The acquirer server may utilize payment ledger and/or merchant account data from the acquirer database, along with the individual payment confirmation, to generate updated payment ledger and/or merchant account data, e.g., 4321. The acquirer server may then store, e.g., 4322, the updated payment ledger and/or merchant account data to the acquire database.
Typically, users, e.g., 4433a, which may be people and/or other systems, may engage information technology systems (e.g., computers) to facilitate information processing. In turn, computers employ processors to process information; such processors 4403 may be referred to as central processing units (CPU). One form of processor is referred to as a microprocessor. CPUs use communicative circuits to pass binary encoded signals acting as instructions to enable various operations. These instructions may be operational and/or data instructions containing and/or referencing other instructions and data in various processor accessible and operable areas of memory 4429 (e.g., registers, cache memory, random access memory, etc.). Such communicative instructions may be stored and/or transmitted in batches (e.g., batches of instructions) as programs and/or data components to facilitate desired operations. These stored instruction codes, e.g., programs, may engage the CPU circuit components and other motherboard and/or system components to perform desired operations. One type of program is a computer operating system, which, may be executed by CPU on a computer; the operating system enables and facilitates users to access and operate computer information technology and resources. Some resources that may be employed in information technology systems include: input and output mechanisms through which data may pass into and out of a computer; memory storage into which data may be saved; and processors by which information may be processed. These information technology systems may be used to collect data for later retrieval, analysis, and manipulation, which may be facilitated through a database program. These information technology systems provide interfaces that allow users to access and operate various system components.
In one embodiment, the V-GLASSES controller 4401 may be connected to and/or communicate with entities such as, but not limited to: one or more users from user input devices 4411; peripheral devices 4412; an optional cryptographic processor device 4428; and/or a communications network 4413. For example, the V-GLASSES controller 4401 may be connected to and/or communicate with users, e.g., 4433a, operating client device(s), e.g., 4433b, including, but not limited to, personal computer(s), server(s) and/or various mobile device(s) including, but not limited to, cellular telephone(s), smartphone(s) (e.g., iPhone®, Blackberry®, Android OS-based phones etc.), tablet computer(s) (e.g., Apple iPad™, HP Slate™, Motorola Xoom™, etc.), eBook reader(s) (e.g., Amazon Kindle™, Barnes and Noble's Nook™ eReader, etc.), laptop computer(s), notebook(s), netbook(s), gaming console(s) (e.g., XBOX Live™, Nintendo® DS, Sony PlayStation® Portable, etc.), portable scanner(s), and/or the like.
Networks are commonly thought to comprise the interconnection and interoperation of clients, servers, and intermediary nodes in a graph topology. It should be noted that the term “server” as used throughout this application refers generally to a computer, other device, program, or combination thereof that processes and responds to the requests of remote users across a communications network. Servers serve their information to requesting “clients.” The term “client” as used herein refers generally to a computer, program, other device, user and/or combination thereof that is capable of processing and making requests and obtaining and processing any responses from servers across a communications network. A computer, other device, program, or combination thereof that facilitates, processes information and requests, and/or furthers the passage of information from a source user to a destination user is commonly referred to as a “node.” Networks are generally thought to facilitate the transfer of information from source points to destinations. A node specifically tasked with furthering the passage of information from a source to a destination is commonly called a “router.” There are many forms of networks such as Local Area Networks (LANs), Pico networks, Wide Area Networks (WANs), Wireless Networks (WLANs), etc. For example, the Internet is generally accepted as being an interconnection of a multitude of networks whereby remote clients and servers may access and interoperate with one another.
The V-GLASSES controller 4401 may be based on computer systems that may comprise, but are not limited to, components such as: a computer systemization 4402 connected to memory 4429.
A computer systemization 4402 may comprise a clock 4430, central processing unit (“CPU(s)” and/or “processor(s)” (these terms are used interchangeable throughout the disclosure unless noted to the contrary)) 4403, a memory 4429 (e.g., a read only memory (ROM) 4406, a random access memory (RAM) 4405, etc.), and/or an interface bus 4407, and most frequently, although not necessarily, are all interconnected and/or communicating through a system bus 4404 on one or more (mother)board(s) 4402 having conductive and/or otherwise transportive circuit pathways through which instructions (e.g., binary encoded signals) may travel to effectuate communications, operations, storage, etc. The computer systemization may be connected to a power source 4486; e.g., optionally the power source may be internal. Optionally, a cryptographic processor 4426 and/or transceivers (e.g., ICs) 4474 may be connected to the system bus. In another embodiment, the cryptographic processor and/or transceivers may be connected as either internal and/or external peripheral devices 4412 via the interface bus I/O. In turn, the transceivers may be connected to antenna(s) 4475, thereby effectuating wireless transmission and reception of various communication and/or sensor protocols; for example the antenna(s) may connect to: a Texas Instruments WiLink WL1283 transceiver chip (e.g., providing 802.11n, Bluetooth 3.0, FM, global positioning system (GPS) (thereby allowing V-GLASSES controller to determine its location)); Broadcom BCM4329FKUBG transceiver chip (e.g., providing 802.11n, Bluetooth 2.1+EDR, FM, etc.); a Broadcom BCM4750IUB8 receiver chip (e.g., GPS); an Infineon Technologies X-Gold 618-PMB9800 (e.g., providing 2G/3G HSDPA/HSUPA communications); and/or the like. The system clock typically has a crystal oscillator and generates a base signal through the computer systemization's circuit pathways. The clock is typically coupled to the system bus and various clock multipliers that will increase or decrease the base operating frequency for other components interconnected in the computer systemization. The clock and various components in a computer systemization drive signals embodying information throughout the system. Such transmission and reception of instructions embodying information throughout a computer systemization may be commonly referred to as communications. These communicative instructions may further be transmitted, received, and the cause of return and/or reply communications beyond the instant computer systemization to: communications networks, input devices, other computer systemizations, peripheral devices, and/or the like. It should be understood that in alternative embodiments, any of the above components may be connected directly to one another, connected to the CPU, and/or organized in numerous variations employed as exemplified by various computer systems.
The CPU comprises at least one high-speed data processor adequate to execute program components for executing user and/or system-generated requests. Often, the processors themselves will incorporate various specialized processing units, such as, but not limited to: integrated system (bus) controllers, memory management control units, floating point units, and even specialized processing sub-units like graphics processing units, digital signal processing units, and/or the like. Additionally, processors may include internal fast access addressable memory, and be capable of mapping and addressing memory 4429 beyond the processor itself; internal memory may include, but is not limited to: fast registers, various levels of cache memory (e.g., level 1, 2, 3, etc.), RAM, etc. The processor may access this memory through the use of a memory address space that is accessible via instruction address, which the processor can construct and decode allowing it to access a circuit path to a specific memory address space having a memory state. The CPU may be a microprocessor such as: AMD's Athlon, Duron and/or Opteron; ARM's application, embedded and secure processors; IBM and/or Motorola's DragonBall and PowerPC; IBM's and Sony's Cell processor; Intel's Celeron, Core (2) Duo, Itanium, Pentium, Xeon, and/or XScale; and/or the like processor(s). The CPU interacts with memory through instruction passing through conductive and/or transportive conduits (e.g., (printed) electronic and/or optic circuits) to execute stored instructions (i.e., program code) according to conventional data processing techniques. Such instruction passing facilitates communication within the V-GLASSES controller and beyond through various interfaces. Should processing requirements dictate a greater amount speed and/or capacity, distributed processors (e.g., Distributed V-GLASSES), mainframe, multi-core, parallel, and/or super-computer architectures may similarly be employed. Alternatively, should deployment requirements dictate greater portability, smaller Personal Digital Assistants (PDAs) may be employed.
Depending on the particular implementation, features of the V-GLASSES may be achieved by implementing a microcontroller such as CAST's R8051XC2 microcontroller; Intel's MCS 51 (i.e., 8051 microcontroller); and/or the like. Also, to implement certain features of the V-GLASSES, some feature implementations may rely on embedded components, such as: Application-Specific Integrated Circuit (“ASIC”), Digital Signal Processing (“DSP”), Field Programmable Gate Array (“FPGA”), and/or the like embedded technology. For example, any of the V-GLASSES component collection (distributed or otherwise) and/or features may be implemented via the microprocessor and/or via embedded components; e.g., via ASIC, coprocessor, DSP, FPGA, and/or the like. Alternately, some implementations of the V-GLASSES may be implemented with embedded components that are configured and used to achieve a variety of features or signal processing.
Depending on the particular implementation, the embedded components may include software solutions, hardware solutions, and/or some combination of both hardware/software solutions. For example, V-GLASSES features discussed herein may be achieved through implementing FPGAs, which are a semiconductor devices containing programmable logic components called “logic blocks”, and programmable interconnects, such as the high performance FPGA Virtex series and/or the low cost Spartan series manufactured by Xilinx. Logic blocks and interconnects can be programmed by the customer or designer, after the FPGA is manufactured, to implement any of the V-GLASSES features. A hierarchy of programmable interconnects allow logic blocks to be interconnected as needed by the V-GLASSES system designer/administrator, somewhat like a one-chip programmable breadboard. An FPGA's logic blocks can be programmed to perform the operation of basic logic gates such as AND, and XOR, or more complex combinational operators such as decoders or simple mathematical operations. In most FPGAs, the logic blocks also include memory elements, which may be circuit flip-flops or more complete blocks of memory. In some circumstances, the V-GLASSES may be developed on regular FPGAs and then migrated into a fixed version that more resembles ASIC implementations. Alternate or coordinating implementations may migrate V-GLASSES controller features to a final ASIC instead of or in addition to FPGAs. Depending on the implementation all of the aforementioned embedded components and microprocessors may be considered the “CPU” and/or “processor” for the V-GLASSES.
The power source 4486 may be of any standard form for powering small electronic circuit board devices such as the following power cells: alkaline, lithium hydride, lithium ion, lithium polymer, nickel cadmium, solar cells, and/or the like. Other types of AC or DC power sources may be used as well. In the case of solar cells, in one embodiment, the case provides an aperture through which the solar cell may capture photonic energy. The power cell 4486 is connected to at least one of the interconnected subsequent components of the V-GLASSES thereby providing an electric current to all subsequent components. In one example, the power source 4486 is connected to the system bus component 4404. In an alternative embodiment, an outside power source 4486 is provided through a connection across the I/O 4408 interface. For example, a USB and/or IEEE 1394 connection carries both data and power across the connection and is therefore a suitable source of power.
Interface bus(ses) 4407 may accept, connect, and/or communicate to a number of interface adapters, conventionally although not necessarily in the form of adapter cards, such as but not limited to: input output interfaces (I/O) 4408, storage interfaces 4409, network interfaces 4410, and/or the like. Optionally, cryptographic processor interfaces 4427 similarly may be connected to the interface bus. The interface bus provides for the communications of interface adapters with one another as well as with other components of the computer systemization. Interface adapters are adapted for a compatible interface bus. Interface adapters conventionally connect to the interface bus via a slot architecture. Conventional slot architectures may be employed, such as, but not limited to: Accelerated Graphics Port (AGP), Card Bus, (Extended) Industry Standard Architecture ((E)ISA), Micro Channel Architecture (MCA), NuBus, Peripheral Component Interconnect (Extended) (PCI(X)), PCI Express, Personal Computer Memory Card International Association (PCMCIA), and/or the like.
Storage interfaces 4409 may accept, communicate, and/or connect to a number of storage devices such as, but not limited to: storage devices 4414, removable disc devices, and/or the like. Storage interfaces may employ connection protocols such as, but not limited to: (Ultra) (Serial) Advanced Technology Attachment (Packet Interface) ((Ultra) (Serial) ATA(PI)), (Enhanced) Integrated Drive Electronics ((E)IDE), Institute of Electrical and Electronics Engineers (IEEE) 1394, fiber channel, Small Computer Systems Interface (SCSI), Universal Serial Bus (USB), and/or the like.
Network interfaces 4410 may accept, communicate, and/or connect to a communications network 4413. Through a communications network 4413, the V-GLASSES controller is accessible through remote clients 4433b (e.g., computers with web browsers) by users 4433a. Network interfaces may employ connection protocols such as, but not limited to: direct connect, Ethernet (thick, thin, twisted pair 10/100/1000 Base T, and/or the like), Token Ring, wireless connection such as IEEE 802.11a-x, and/or the like. Should processing requirements dictate a greater amount speed and/or capacity, distributed network controllers (e.g., Distributed V-GLASSES), architectures may similarly be employed to pool, load balance, and/or otherwise increase the communicative bandwidth required by the V-GLASSES controller. A communications network may be any one and/or the combination of the following: a direct interconnection; the Internet; a Local Area Network (LAN); a Metropolitan Area Network (MAN); an Operating Missions as Nodes on the Internet (OMNI); a secured custom connection; a Wide Area Network (WAN); a wireless network (e.g., employing protocols such as, but not limited to a Wireless Application Protocol (WAP), I-mode, and/or the like); and/or the like. A network interface may be regarded as a specialized form of an input output interface. Further, multiple network interfaces 4410 may be used to engage with various communications network types 4413. For example, multiple network interfaces may be employed to allow for the communication over broadcast, multicast, and/or unicast networks.
Input Output interfaces (I/O) 4408 may accept, communicate, and/or connect to user input devices 4411, peripheral devices 4412, cryptographic processor devices 4428, and/or the like. I/O may employ connection protocols such as, but not limited to: audio: analog, digital, monaural, RCA, stereo, and/or the like; data: Apple Desktop Bus (ADB), IEEE 1394a-b, serial, universal serial bus (USB); infrared; joystick; keyboard; midi; optical; PC AT; PS/2; parallel; radio; video interface: Apple Desktop Connector (ADC), BNC, coaxial, component, composite, digital, Digital Visual Interface (DVI), high-definition multimedia interface (HDMI), RCA, RF antennae, S-Video, VGA, and/or the like; wireless transceivers: 802.11a/b/g/n/x; Bluetooth; cellular (e.g., code division multiple access (CDMA), high speed packet access (HSPA(+)), high-speed downlink packet access (HSDPA), global system for mobile communications (GSM), long term evolution (LTE), WiMax, etc.); and/or the like. One typical output device may include a video display, which typically comprises a Cathode Ray Tube (CRT) or Liquid Crystal Display (LCD) based monitor with an interface (e.g., DVI circuitry and cable) that accepts signals from a video interface, may be used. The video interface composites information generated by a computer systemization and generates video signals based on the composited information in a video memory frame. Another output device is a television set, which accepts signals from a video interface. Typically, the video interface provides the composited video information through a video connection interface that accepts a video display interface (e.g., an RCA composite video connector accepting an RCA composite video cable; a DVI connector accepting a DVI display cable, etc.).
User input devices 4411 often are a type of peripheral device 4412 (see below) and may include: card readers, dongles, finger print readers, gloves, graphics tablets, joysticks, keyboards, microphones, mouse (mice), remote controls, retina readers, touch screens (e.g., capacitive, resistive, etc.), trackballs, trackpads, sensors (e.g., accelerometers, ambient light, GPS, gyroscopes, proximity, etc.), styluses, and/or the like.
Peripheral devices 4412 may be connected and/or communicate to I/O and/or other facilities of the like such as network interfaces, storage interfaces, directly to the interface bus, system bus, the CPU, and/or the like. Peripheral devices may be external, internal and/or part of the V-GLASSES controller. Peripheral devices may include: antenna, audio devices (e.g., line-in, line-out, microphone input, speakers, etc.), cameras (e.g., still, video, webcam, etc.), dongles (e.g., for copy protection, ensuring secure transactions with a digital signature, and/or the like), external processors (for added capabilities; e.g., crypto devices 4428), force-feedback devices (e.g., vibrating motors), network interfaces, printers, scanners, storage devices, transceivers (e.g., cellular, GPS, etc.), video devices (e.g., goggles, monitors, etc.), video sources, visors, and/or the like. Peripheral devices often include types of input devices (e.g., cameras).
It should be noted that although user input devices and peripheral devices may be employed, the V-GLASSES controller may be embodied as an embedded, dedicated, and/or monitor-less (i.e., headless) device, wherein access would be provided over a network interface connection.
Cryptographic units such as, but not limited to, microcontrollers, processors 4426, interfaces 4427, and/or devices 4428 may be attached, and/or communicate with the V-GLASSES controller. A MC68HC16 microcontroller, manufactured by Motorola Inc., may be used for and/or within cryptographic units. The MC68HC16 microcontroller utilizes a 16-bit multiply-and-accumulate instruction in the 16 MHz configuration and requires less than one second to perform a 512-bit RSA private key operation. Cryptographic units support the authentication of communications from interacting agents, as well as allowing for anonymous transactions. Cryptographic units may also be configured as part of the CPU. Equivalent microcontrollers and/or processors may also be used. Other commercially available specialized cryptographic processors include: the Broadcom's CryptoNetX and other Security Processors; nCipher's nShield, SafeNet's Luna PCI (e.g., 7100) series; Semaphore Communications' 40 MHz Roadrunner 184; Sun's Cryptographic Accelerators (e.g., Accelerator 6000 PCIe Board, Accelerator 500 Daughtercard); Via Nano Processor (e.g., L2100, L2200, U2400) line, which is capable of performing 500+MB/s of cryptographic instructions; VLSI Technology's 33 MHz 6868; and/or the like.
Generally, any mechanization and/or embodiment allowing a processor to affect the storage and/or retrieval of information is regarded as memory 4429. However, memory is a fungible technology and resource, thus, any number of memory embodiments may be employed in lieu of or in concert with one another. It is to be understood that the V-GLASSES controller and/or a computer systemization may employ various forms of memory 4429. For example, a computer systemization may be configured wherein the operation of on-chip CPU memory (e.g., registers), RAM, ROM, and any other storage devices are provided by a paper punch tape or paper punch card mechanism; however, such an embodiment would result in an extremely slow rate of operation. In a typical configuration, memory 4429 will include ROM 4406, RAM 4405, and a storage device 4414. A storage device 4414 may be any conventional computer system storage. Storage devices may include a drum; a (fixed and/or removable) magnetic disk drive; a magneto-optical drive; an optical drive (i.e., Blueray, CD ROM/RAM/Recordable (R)/ReWritable (RW), DVD R/RW, HD DVD R/RW etc.); an array of devices (e.g., Redundant Array of Independent Disks (RAID)); solid state memory devices (USB memory, solid state drives (SSD), etc.); other processor-readable storage mediums; and/or other devices of the like. Thus, a computer systemization generally requires and makes use of memory.
The memory 4429 may contain a collection of program and/or database components and/or data such as, but not limited to: operating system component(s) 4415 (operating system); information server component(s) 4416 (information server); user interface component(s) 4417 (user interface); Web browser component(s) 4418 (Web browser); database(s) 4419; mail server component(s) 4421; mail client component(s) 4422; cryptographic server component(s) 4420 (cryptographic server); the V-GLASSES component(s) 4435; and/or the like (i.e., collectively a component collection). These components may be stored and accessed from the storage devices and/or from storage devices accessible through an interface bus. Although non-conventional program components such as those in the component collection, typically, are stored in a local storage device 4414, they may also be loaded and/or stored in memory such as: peripheral devices, RAM, remote storage facilities through a communications network, ROM, various forms of memory, and/or the like.
The operating system component 4415 is an executable program component facilitating the operation of the V-GLASSES controller. Typically, the operating system facilitates access of I/O, network interfaces, peripheral devices, storage devices, and/or the like. The operating system may be a highly fault tolerant, scalable, and secure system such as: Apple Macintosh OS X (Server); AT&T Plan 9; Be OS; Unix and Unix-like system distributions (such as AT&T's UNIX; Berkley Software Distribution (BSD) variations such as FreeBSD, NetBSD, OpenBSD, and/or the like; Linux distributions such as Red Hat, Ubuntu, and/or the like); and/or the like operating systems. However, more limited and/or less secure operating systems also may be employed such as Apple Macintosh OS, IBM OS/2, Microsoft DOS, Microsoft Windows 2000/2003/3.1/95/98/CE/Millenium/NT/Vista/XP (Server), Palm OS, and/or the like. An operating system may communicate to and/or with other components in a component collection, including itself, and/or the like. Most frequently, the operating system communicates with other program components, user interfaces, and/or the like. For example, the operating system may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses. The operating system, once executed by the CPU, may enable the interaction with communications networks, data, I/O, peripheral devices, program components, memory, user input devices, and/or the like. The operating system may provide communications protocols that allow the V-GLASSES controller to communicate with other entities through a communications network 4413. Various communication protocols may be used by the V-GLASSES controller as a subcarrier transport mechanism for interaction, such as, but not limited to: multicast, TCP/IP, UDP, unicast, and/or the like.
An information server component 4416 is a stored program component that is executed by a CPU. The information server may be a conventional Internet information server such as, but not limited to Apache Software Foundation's Apache, Microsoft's Internet Information Server, and/or the like. The information server may allow for the execution of program components through facilities such as Active Server Page (ASP), ActiveX, (ANSI) (Objective-) C (++), C# and/or .NET, Common Gateway Interface (CGI) scripts, dynamic (D) hypertext markup language (HTML), FLASH, Java, JavaScript, Practical Extraction Report Language (PERL), Hypertext Pre-Processor (PHP), pipes, Python, wireless application protocol (WAP), WebObjects, and/or the like. The information server may support secure communications protocols such as, but not limited to, File Transfer Protocol (FTP); HyperText Transfer Protocol (HTTP); Secure Hypertext Transfer Protocol (HTTPS), Secure Socket Layer (SSL), messaging protocols (e.g., America Online (AOL) Instant Messenger (AIM), Application Exchange (APEX), ICQ, Internet Relay Chat (IRC), Microsoft Network (MSN) Messenger Service, Presence and Instant Messaging Protocol (PRIM), Internet Engineering Task Force's (IETF's) Session Initiation Protocol (SIP), SIP for Instant Messaging and Presence Leveraging Extensions (SIMPLE), open XML-based Extensible Messaging and Presence Protocol (XMPP) (i.e., Jabber or Open Mobile Alliance's (OMA's) Instant Messaging and Presence Service (IMPS)), Yahoo! Instant Messenger Service, and/or the like. The information server provides results in the form of Web pages to Web browsers, and allows for the manipulated generation of the Web pages through interaction with other program components. After a Domain Name System (DNS) resolution portion of an HTTP request is resolved to a particular information server, the information server resolves requests for information at specified locations on the V-GLASSES controller based on the remainder of the HTTP request. For example, a request such as http://123.124.125.126/myInformation.html might have the IP portion of the request “123.124.125.126” resolved by a DNS server to an information server at that IP address; that information server might in turn further parse the http request for the “/myInformation.html” portion of the request and resolve it to a location in memory containing the information “myInformation.html.” Additionally, other information serving protocols may be employed across various ports, e.g., FTP communications across port 21, and/or the like. An information server may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the information server communicates with the V-GLASSES database 4419, operating systems, other program components, user interfaces, Web browsers, and/or the like.
Access to the V-GLASSES database may be achieved through a number of database bridge mechanisms such as through scripting languages as enumerated below (e.g., CGI) and through inter-application communication channels as enumerated below (e.g., CORBA, WebObjects, etc.). Any data requests through a Web browser are parsed through the bridge mechanism into appropriate grammars as required by the V-GLASSES. In one embodiment, the information server would provide a Web form accessible by a Web browser. Entries made into supplied fields in the Web form are tagged as having been entered into the particular fields, and parsed as such. The entered terms are then passed along with the field tags, which act to instruct the parser to generate queries directed to appropriate tables and/or fields. In one embodiment, the parser may generate queries in standard SQL by instantiating a search string with the proper join/select commands based on the tagged text entries, wherein the resulting command is provided over the bridge mechanism to the V-GLASSES as a query. Upon generating query results from the query, the results are passed over the bridge mechanism, and may be parsed for formatting and generation of a new results Web page by the bridge mechanism. Such a new results Web page is then provided to the information server, which may supply it to the requesting Web browser.
Also, an information server may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
Computer interfaces in some respects are similar to automobile operation interfaces. Automobile operation interface elements such as steering wheels, gearshifts, and speedometers facilitate the access, operation, and display of automobile resources, and status. Computer interaction interface elements such as check boxes, cursors, menus, scrollers, and windows (collectively and commonly referred to as widgets) similarly facilitate the access, capabilities, operation, and display of data and computer hardware and operating system resources, and status. Operation interfaces are commonly called user interfaces. Graphical user interfaces (GUIs) such as the Apple Macintosh Operating System's Aqua, IBM's OS/2, Microsoft's Windows 2000/2003/3.1/95/98/CE/Millenium/NT/XP/Vista/7 (i.e., Aero), Unix's X-Windows (e.g., which may include additional Unix graphic interface libraries and layers such as K Desktop Environment (KDE), mythTV and GNU Network Object Model Environment (GNOME)), web interface libraries (e.g., ActiveX, AJAX, (D)HTML, FLASH, Java, JavaScript, etc. interface libraries such as, but not limited to, Dojo, jQuery(UI), MooTools, Prototype, script.aculo.us, SWFObject, Yahoo! User Interface, any of which may be used and) provide a baseline and means of accessing and displaying information graphically to users.
A user interface component 4417 is a stored program component that is executed by a CPU. The user interface may be a conventional graphic user interface as provided by, with, and/or atop operating systems and/or operating environments such as already discussed. The user interface may allow for the display, execution, interaction, manipulation, and/or operation of program components and/or system facilities through textual and/or graphical facilities. The user interface provides a facility through which users may affect, interact, and/or operate a computer system. A user interface may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the user interface communicates with operating systems, other program components, and/or the like. The user interface may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
A Web browser component 4418 is a stored program component that is executed by a CPU. The Web browser may be a conventional hypertext viewing application such as Microsoft Internet Explorer or Netscape Navigator. Secure Web browsing may be supplied with 128 bit (or greater) encryption by way of HTTPS, SSL, and/or the like. Web browsers allowing for the execution of program components through facilities such as ActiveX, AJAX, (D)HTML, FLASH, Java, JavaScript, web browser plug-in APIs (e.g., FireFox, Safari Plug-in, and/or the like APIs), and/or the like. Web browsers and like information access tools may be integrated into PDAs, cellular telephones, and/or other mobile devices. A Web browser may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the Web browser communicates with information servers, operating systems, integrated program components (e.g., plug-ins), and/or the like; e.g., it may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses. Also, in place of a Web browser and information server, a combined application may be developed to perform similar operations of both. The combined application would similarly affect the obtaining and the provision of information to users, user agents, and/or the like from the V-GLASSES enabled nodes. The combined application may be nugatory on systems employing standard Web browsers.
A mail server component 4421 is a stored program component that is executed by a CPU 4403. The mail server may be a conventional Internet mail server such as, but not limited to sendmail, Microsoft Exchange, and/or the like. The mail server may allow for the execution of program components through facilities such as V-GLASSES, ActiveX, (ANSI) (Objective-) C (++), C# and/or .NET, CGI scripts, Java, JavaScript, PERL, PHP, pipes, Python, WebObjects, and/or the like. The mail server may support communications protocols such as, but not limited to: Internet message access protocol (IMAP), Messaging Application Programming Interface (MAPI)/Microsoft Exchange, post office protocol (POP3), simple mail transfer protocol (SMTP), and/or the like. The mail server can route, forward, and process incoming and outgoing mail messages that have been sent, relayed and/or otherwise traversing through and/or to the V-GLASSES.
Access to the V-GLASSES mail may be achieved through a number of APIs offered by the individual Web server components and/or the operating system.
Also, a mail server may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, information, and/or responses.
A mail client component 4422 is a stored program component that is executed by a CPU 4403. The mail client may be a conventional mail viewing application such as Apple Mail, Microsoft Entourage, Microsoft Outlook, Microsoft Outlook Express, Mozilla, Thunderbird, and/or the like. Mail clients may support a number of transfer protocols, such as: IMAP, Microsoft Exchange, POP3, SMTP, and/or the like. A mail client may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the mail client communicates with mail servers, operating systems, other mail clients, and/or the like; e.g., it may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, information, and/or responses. Generally, the mail client provides a facility to compose and transmit electronic mail messages.
A cryptographic server component 4420 is a stored program component that is executed by a CPU 4403, cryptographic processor 4426, cryptographic processor interface 4427, cryptographic processor device 4428, and/or the like. Cryptographic processor interfaces will allow for expedition of encryption and/or decryption requests by the cryptographic component; however, the cryptographic component, alternatively, may run on a conventional CPU. The cryptographic component allows for the encryption and/or decryption of provided data. The cryptographic component allows for both symmetric and asymmetric (e.g., Pretty Good Protection (PGP)) encryption and/or decryption. The cryptographic component may employ cryptographic techniques such as, but not limited to: digital certificates (e.g., X.509 authentication framework), digital signatures, dual signatures, enveloping, password access protection, public key management, and/or the like. The cryptographic component will facilitate numerous (encryption and/or decryption) security protocols such as, but not limited to: checksum, Data Encryption Standard (DES), Elliptical Curve Encryption (ECC), International Data Encryption Algorithm (IDEA), Message Digest 5 (MD5, which is a one way hash operation), passwords, Rivest Cipher (RC5), Rijndael, RSA (which is an Internet encryption and authentication system that uses an algorithm developed in 1977 by Ron Rivest, Adi Shamir, and Leonard Adleman), Secure Hash Algorithm (SHA), Secure Socket Layer (SSL), Secure Hypertext Transfer Protocol (HTTPS), and/or the like. Employing such encryption security protocols, the V-GLASSES may encrypt all incoming and/or outgoing communications and may serve as node within a virtual private network (VPN) with a wider communications network. The cryptographic component facilitates the process of “security authorization” whereby access to a resource is inhibited by a security protocol wherein the cryptographic component effects authorized access to the secured resource. In addition, the cryptographic component may provide unique identifiers of content, e.g., employing and MD5 hash to obtain a unique signature for an digital audio file. A cryptographic component may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. The cryptographic component supports encryption schemes allowing for the secure transmission of information across a communications network to enable the V-GLASSES component to engage in secure transactions if so desired. The cryptographic component facilitates the secure accessing of resources on the V-GLASSES and facilitates the access of secured resources on remote systems; i.e., it may act as a client and/or server of secured resources. Most frequently, the cryptographic component communicates with information servers, operating systems, other program components, and/or the like. The cryptographic component may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
The V-GLASSES database component 4419 may be embodied in a database and its stored data. The database is a stored program component, which is executed by the CPU; the stored program component portion configuring the CPU to process the stored data. The database may be a conventional, fault tolerant, relational, scalable, secure database such as Oracle or Sybase. Relational databases are an extension of a flat file. Relational databases consist of a series of related tables. The tables are interconnected via a key field. Use of the key field allows the combination of the tables by indexing against the key field; i.e., the key fields act as dimensional pivot points for combining information from various tables. Relationships generally identify links maintained between tables by matching primary keys. Primary keys represent fields that uniquely identify the rows of a table in a relational database. More precisely, they uniquely identify rows of a table on the “one” side of a one-to-many relationship.
Alternatively, the V-GLASSES database may be implemented using various standard data-structures, such as an array, hash, (linked) list, struct, structured text file (e.g., XML), table, and/or the like. Such data-structures may be stored in memory and/or in (structured) files. In another alternative, an object-oriented database may be used, such as Frontier, ObjectStore, Poet, Zope, and/or the like. Object databases can include a number of object collections that are grouped and/or linked together by common attributes; they may be related to other object collections by some common attributes. Object-oriented databases perform similarly to relational databases with the exception that objects are not just pieces of data but may have other types of capabilities encapsulated within a given object. If the V-GLASSES database is implemented as a data-structure, the use of the V-GLASSES database 4419 may be integrated into another component such as the V-GLASSES component 4435. Also, the database may be implemented as a mix of data structures, objects, and relational structures. Databases may be consolidated and/or distributed in countless variations through standard data processing techniques. Portions of databases, e.g., tables, may be exported and/or imported and thus decentralized and/or integrated.
In one embodiment, the database component 4419 includes several tables 4419a-q. A Users table 4419a may include fields such as, but not limited to: user_id, ssn, dob, first_name, last_name, age, state, address_firstline, address_secondline, zipcode, devices_list, contact_info, contact_type, alt_contact_info, alt_contact_type, user_gender, user_clothing_size, user_body_type, user_eye_color, user_hair_color, user_complexion, user_personalized_gesture_models, user_recmmended_items, user_image, user_image_date, user_body_joint_location, and/or the like. The Users table may support and/or track multiple entity accounts on a V-GLASSES. A Devices table 4419b may include fields such as, but not limited to: device_ID, device_name, device_IP, device_GPS, device_MAC, device_serial, device_ECID, device_UDID, device_browser, device_type, device_model, device_version, device_OS, device-apps_list. device_securekey, wallet-appinstalled flag, and/or the like. An Apps table 4419c may include fields such as, but not limited to: app_ID, app_name, app_type, app_dependencies, app_access_code, user_pin, and/or the like. An Accounts table 4419d may include fields such as, but not limited to: account_number, account_security_code, account_name, issuer_acquirer_flag, issuer_name, acquirer_name, account_address, routing_number, access_API_call, linked_wallets_list, and/or the like. A Merchants table 4419e may include fields such as, but not limited to: merchant_id, merchant_name, merchant_address, store_id, ip_address, mac_address, auth_key, port_num, security_settings_list, and/or the like. An Issuers table 4419f may include fields such as, but not limited to: issuer_id, issuer_name, issuer_address, ip_address, mac_address, auth_key, port_num, security_settings_list, and/or the like. An Acquirers table 4419g may include fields such as, but not limited to: account_firstname, account_lastname, account_type, account_num, account_balance_list, billingaddress_line1, billingaddress_line2, billing_zipcode, billing_state, shipping_preferences, shippingaddress_line1, shippingaddress_line2, shipping_zipcode, shipping_state, and/or the like. A Pay Gateways table 4419h may include fields such as, but not limited to: gateway_ID, gateway_IP, gateway_MAC, gateway_secure_key, gateway_access_list, gateway_API_call_list, gateway_services_list, and/or the like. A Shop Sessions table 4419i may include fields such as, but not limited to: user_id, session_id, alerts_URL, timestamp, expiry_lapse, merchant_id, store_id, device_type, device_ID, device_IP, device_MAC, device_browser, device_serial, device_ECID, device_model, device_OS, wallet_app_installed, total_cost, cart_ID_list, product_params_list, social_flag, social_message, social_networks_list, coupon_lists, accounts_list, CVV2_lists, charge_ratio_list, charge_priority_list, value_exchange_symbols_list, bill_address, ship_address, cloak_flag, pay_mode, alerts_rules_list, and/or the like. A Transactions table 4419j may include fields such as, but not limited to: order_id, user_id, timestamp, transaction_cost, purchase_details_list, num_products, products_list, product_type, product_params_list, product_title, product_summary, quantity, user_id, client_id, client_ip, client_type, client_model, operating_system, os_version, app_installed_flag, user_id, account_firstname, account_lastname, account_type, account_num, account_priority_account_ratio, billingaddress_line, billingaddress_line2, billing_zipcode, billing_state, shipping_preferences, shippingaddress_line1, shippingaddress_line2, shipping_zipcode, shipping_state, merchant_id, merchant_name, merchant_auth_key, and/or the like. A Batches table 4419k may include fields such as, but not limited to: batch_id, transaction_id_list, timestamp_list, cleared_flag_list, clearance_trigger settings, and/or the like. A Ledgers table 4419l may include fields such as, but not limited to: request_id, timestamp, deposit_amount, batch_id, transaction_id, clear_flag, deposit_account, transaction_summary, payor_name, payor_account, and/or the like. A Products table 4419m may include fields such as, but not limited to: product_ID, product_title, product_attributes_list, product_price, tax_info_list, related_products_list, offers_list, discounts_list, rewards_list, merchants_list, merchant_availability_list, product_date_added, product_image, product_qr, product_manufacturer, product_model, product_aisle, product_stack, product_shelf, product_type, and/or the like. An Offers table 4419n may include fields such as, but not limited to: offer_ID, offer_title, offer_attributes_list, offer_price, offer_expiry, related_products_list, discounts_list, rewards_list, merchants_list, merchant_availability_list, and/or the like. A Behavior Data table 4419o may include fields such as, but not limited to: user_id, timestamp, activity_type, activity_location, activity_attribute_list, activity_attribute_values_list, and/or the like. A Label Analytics table 4419p may include fields such as, but not limited to: label_id, label_name, label_format, label_account_type, label_session_id, label_session_type, label_product_id, label_product_type, Label_transaction_id, label_transaction_type, and/or the like. A Social table 4419q may include fields such as, but not limited to: social_id, social_name, social_server_id, social_server_ip, social_domain_id, social_source, social_feed_id, social_feed_source, social_comment, social_comment_time, social_comment_keyterms, social_comment_product_id, and/or the like. A MDGA table 4419r includes fields such as, but not limited to: MDGA_id, MDGA_name, MDGA_touch_gestures, MDGA_finger_gestures, MDGA_QR_gestures, MDGA_object_gestures, MDGA_vocal_commands, MDGA_merchant, and/or the like. The MDGA table may support and/or track multiple possible composite actions on a V-GLASSES. A payment device table 4419s includes fields such as, but not limited to: pd_id, pd_user, pd_type, pd_issuer, pd_issuer_id, pd_qr, pd_date_added, and/or the like. The payment device table may support and/or track multiple payment devices used on a V-GLASSES. An object gestures table 4419t includes fields such as, but not limited to: object_gesture_id, object_gesture_type, object_gesture_x, object_gesture_x, object_gesture_merchant, and/or the like. The object gesture table may support and/or track multiple object gestures performed on a V-GLASSES. A touch gesture table 4419u includes fields such as, but not limited to: touch_gesture_id, touch_gesture_type, touch_gesture_x, touch_gesture_x, touch_gesture_merchant, and/or the like. The touch gestures table may support and/or track multiple touch gestures performed on a V-GLASSES. A finger gesture table 4419v includes fields such as, but not limited to: finger_gesture_id, finger_gesture_type, finger_gesture_x, finger_gesture_x, finger_gesture_merchant, and/or the like. The finger gestures table may support and/or track multiple finger gestures performed on a V-GLASSES. A QR gesture table 4419w includes fields such as, but not limited to: QR_gesture_id, QR_gesture_type, QR_gesture_x, QR_gesture_x, QR_gesture_merchant, and/or the like. The QR gestures table may support and/or track multiple QR gestures performed on a V-GLASSES. A vocal command table 4419x includes fields such as, but not limited to: vc_id, vc_name, vc_command_list, and/or the like. The vocal command gestures table may support and/or track multiple vocal commands performed on a V-GLASSES.
In one embodiment, the V-GLASSES database may interact with other database systems. For example, employing a distributed database system, queries and data access by search V-GLASSES component may treat the combination of the V-GLASSES database, an integrated data security layer database as a single database entity.
In one embodiment, user programs may contain various user interface primitives, which may serve to update the V-GLASSES. Also, various accounts may require custom database tables depending upon the environments and the types of clients the V-GLASSES may need to serve. It should be noted that any unique fields may be designated as a key field throughout. In an alternative embodiment, these tables have been decentralized into their own databases and their respective database controllers (i.e., individual database controllers for each of the above tables). Employing standard data processing techniques, one may further distribute the databases over several computer systemizations and/or storage devices. Similarly, configurations of the decentralized database controllers may be varied by consolidating and/or distributing the various database components 4419a-x. The V-GLASSES may be configured to keep track of various settings, inputs, and parameters via database controllers.
The V-GLASSES database may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the V-GLASSES database communicates with the V-GLASSES component, other program components, and/or the like. The database may contain, retain, and provide information regarding other nodes and data.
The V-GLASSES component 4435 is a stored program component that is executed by a CPU. In one embodiment, the V-GLASSES component incorporates any and/or all combinations of the aspects of the V-GLASSES discussed in the previous figures. As such, the V-GLASSES affects accessing, obtaining and the provision of information, services, transactions, and/or the like across various communications networks.
The V-GLASSES component may transform reality scene visual captures (e.g., see 213 in
The V-GLASSES component enabling access of information between nodes may be developed by employing standard development tools and languages such as, but not limited to: Apache components, Assembly, ActiveX, binary executables, (ANSI) (Objective-) C (++), C# and/or .NET, database adapters, CGI scripts, Java, JavaScript, mapping tools, procedural and object oriented development tools, PERL, PHP, Python, shell scripts, SQL commands, web application server extensions, web development environments and libraries (e.g., Microsoft's ActiveX; Adobe AIR, FLEX & FLASH; AJAX; (D)HTML; Dojo, Java; JavaScript; jQuery(UI); MooTools; Prototype; script.aculo.us; Simple Object Access Protocol (SOAP); SWFObject; Yahoo! User Interface; and/or the like), WebObjects, and/or the like. In one embodiment, the V-GLASSES server employs a cryptographic server to encrypt and decrypt communications. The V-GLASSES component may communicate to and/or with other components in a component collection, including itself, and/or facilities of the like. Most frequently, the V-GLASSES component communicates with the V-GLASSES database, operating systems, other program components, and/or the like. The V-GLASSES may contain, communicate, generate, obtain, and/or provide program component, system, user, and/or data communications, requests, and/or responses.
The structure and/or operation of any of the V-GLASSES node controller components may be combined, consolidated, and/or distributed in any number of ways to facilitate development and/or deployment. Similarly, the component collection may be combined in any number of ways to facilitate deployment and/or development. To accomplish this, one may integrate the components into a common code base or in a facility that can dynamically load the components on demand in an integrated fashion.
The component collection may be consolidated and/or distributed in countless variations through standard data processing and/or development techniques. Multiple instances of any one of the program components in the program component collection may be instantiated on a single node, and/or across numerous nodes to improve performance through load-balancing and/or data-processing techniques. Furthermore, single instances may also be distributed across multiple controllers and/or storage devices; e.g., databases. All program component instances and controllers working in concert may do so through standard data processing communication techniques.
The configuration of the V-GLASSES controller will depend on the context of system deployment. Factors such as, but not limited to, the budget, capacity, location, and/or use of the underlying hardware resources may affect deployment requirements and configuration. Regardless of if the configuration results in more consolidated and/or integrated program components, results in a more distributed series of program components, and/or results in some combination between a consolidated and distributed configuration, data may be communicated, obtained, and/or provided. Instances of components consolidated into a common code base from the program component collection may communicate, obtain, and/or provide data. This may be accomplished through intra-application data processing communication techniques such as, but not limited to: data referencing (e.g., pointers), internal messaging, object instance variable communication, shared memory space, variable passing, and/or the like.
If component collection components are discrete, separate, and/or external to one another, then communicating, obtaining, and/or providing data with and/or to other components may be accomplished through inter-application data processing communication techniques such as, but not limited to: Application Program Interfaces (API) information passage; (distributed) Component Object Model ((D)COM), (Distributed) Object Linking and Embedding ((D)OLE), and/or the like), Common Object Request Broker Architecture (CORBA), Jini local and remote application program interfaces, JavaScript Object Notation (JSON), Remote Method Invocation (RMI), SOAP, process pipes, shared files, and/or the like. Messages sent between discrete component components for inter-application communication or within memory spaces of a singular component for intra-application communication may be facilitated through the creation and parsing of a grammar. A grammar may be developed by using development tools such as lex, yacc, XML, and/or the like, which allow for grammar generation and parsing capabilities, which in turn may form the basis of communication messages within and between components.
For example, a grammar may be arranged to recognize the tokens of an HTTP post command, e.g.:
where Value1 is discerned as being a parameter because “http://” is part of the grammar syntax, and what follows is considered part of the post value. Similarly, with such a grammar, a variable “Value1” may be inserted into an “http://” post command and then sent. The grammar syntax itself may be presented as structured data that is interpreted and/or otherwise used to generate the parsing mechanism (e.g., a syntax description text file as processed by lex, yacc, etc.). Also, once the parsing mechanism is generated and/or instantiated, it itself may process and/or parse structured data such as, but not limited to: character (e.g., tab) delineated text, HTML, structured text streams, XML, and/or the like structured data. In another embodiment, inter-application data processing protocols themselves may have integrated and/or readily available parsers (e.g., JSON, SOAP, and/or like parsers) that may be employed to parse (e.g., communications) data. Further, the parsing grammar may be used beyond message parsing, but may also be used to parse: databases, data collections, data stores, structured data, and/or the like. Again, the desired configuration will depend upon the context, environment, and requirements of system deployment.
For example, in some implementations, the V-GLASSES controller may be executing a PHP script implementing a Secure Sockets Layer (“SSL”) socket server via the information server, which listens to incoming communications on a server port to which a client may send data, e.g., data encoded in JSON format. Upon identifying an incoming communication, the PHP script may read the incoming message from the client device, parse the received JSON-encoded text data to extract information from the JSON-encoded text data into PHP script variables, and store the data (e.g., client identifying information, etc.) and/or extracted information in a relational database accessible using the Structured Query Language (“SQL”). An exemplary listing, written substantially in the form of PHP/SQL commands, to accept JSON-encoded input data from a client device via a SSL connection, parse the data to extract variables, and store the data to a database, is provided below:
Also, the following resources may be used to provide example embodiments regarding SOAP parser implementation:
and other parser implementations:
all of which are hereby expressly incorporated by reference herein.
In order to address various issues and advance the art, the entirety of this application for AUGMENTED REALITY VISION DEVICE APPARATUSES, METHODS AND SYSTEMS (including the Cover Page, Title, Headings, Field, Background, Summary, Brief Description of the Drawings, Detailed Description, Claims, Abstract, Figures, Appendices and/or otherwise) shows by way of illustration various embodiments in which the claimed innovations may be practiced. The advantages and features of the application are of a representative sample of embodiments only, and are not exhaustive and/or exclusive. They are presented only to assist in understanding and teach the claimed principles. It should be understood that they are not representative of all claimed innovations. As such, certain aspects of the disclosure have not been discussed herein. That alternate embodiments may not have been presented for a specific portion of the innovations or that further undescribed alternate embodiments may be available for a portion is not to be considered a disclaimer of those alternate embodiments. It will be appreciated that many of those undescribed embodiments incorporate the same principles of the innovations and others are equivalent. Thus, it is to be understood that other embodiments may be utilized and functional, logical, operational, organizational, structural and/or topological modifications may be made without departing from the scope and/or spirit of the disclosure. As such, all examples and/or embodiments are deemed to be non-limiting throughout this disclosure. Also, no inference should be drawn regarding those embodiments discussed herein relative to those not discussed herein other than it is as such for purposes of reducing space and repetition. For instance, it is to be understood that the logical and/or topological structure of any combination of any program components (a component collection), other components and/or any present feature sets as described in the figures and/or throughout are not limited to a fixed operating order and/or arrangement, but rather, any disclosed order is exemplary and all equivalents, regardless of order, are contemplated by the disclosure. Furthermore, it is to be understood that such features are not limited to serial execution, but rather, any number of threads, processes, services, servers, and/or the like that may execute asynchronously, concurrently, in parallel, simultaneously, synchronously, and/or the like are contemplated by the disclosure. As such, some of these features may be mutually contradictory, in that they cannot be simultaneously present in a single embodiment. Similarly, some features are applicable to one aspect of the innovations, and inapplicable to others. In addition, the disclosure includes other innovations not presently claimed. Applicant reserves all rights in those presently unclaimed innovations, including the right to claim such innovations, file additional applications, continuations, continuations in part, divisions, and/or the like thereof. As such, it should be understood that advantages, embodiments, examples, functional, features, logical, operational, organizational, structural, topological, and/or other aspects of the disclosure are not to be considered limitations on the disclosure as defined by the claims or limitations on equivalents to the claims. It is to be understood that, depending on the particular needs and/or characteristics of a V-GLASSES individual and/or enterprise user, database configuration and/or relational model, data type, data transmission and/or network framework, syntax structure, and/or the like, various embodiments of the V-GLASSES may be implemented that enable a great deal of flexibility and customization. For example, aspects of the V-GLASSES may be adapted for (electronic/financial) trading systems, financial planning systems, and/or the like. While various embodiments and discussions of the V-GLASSES have been directed to enhanced interactive user interface, however, it is to be understood that the embodiments described herein may be readily configured and/or customized for a wide variety of other applications and/or implementations.
The Gesture Recognition Cloud Computing Terminal (hereinafter “GRCCT”) allows users to send commands to remote systems (e.g., cloud systems) via gestures (e.g., hand or body movements) in open and public environments. With this system, a consumer shopping at a brick and mortar store can interact with the store and issue commands without needing to seek out sales agents or kiosks and thereby disrupt his shopping experience. Moreover, the consumer may personalize his gesture commands so that they are unique, private, and convenient for him.
Once the identity of the user 5700 is ascertained, the GRCCT system then tries to locate the user 5700 in the store in order to monitor his gestures. In one embodiment, biometrics are used to locate the user 5700. For example, the computer system 5704 may attempt to locate the user 5700 using conventional face-recognition techniques known in the art based on his facial parameters on file (e.g., the user's 5700 profile may include a picture of himself). The computer system 5704 may use sensor 5706 (e.g., cameras) to detect facial features 5705 of detectable customers in the store, and compare the detected facial features 5705 to the user's 5700 facial parameters on file. Once a match is found, the computer system 5704 would have located the user 5700. If a match cannot be found, the computer system 5704 may cause a message (e.g., SMS) to be sent to the identified user 5700. Once sent, the computer system 5704 may monitor the movements of every person within view of its cameras to detect any movement that may resemble picking up or looking at a mobile device or phone. Those who have moved as such may have a greater likelihood of being the user 5700 whom the computer system 5704 is trying to locate. Therefore, once such motion by a person is detected, the computer system 5704 may detect his/her biofeatures and reassess whether the person matches the profile of the user 5700.
Once located, the user's 5700 gestures 5707 may be detected by sensor 5708 (e.g., cameras). The detected gestures 5707 are provided to the computer system 5704, which identifies the gesture 5707 and compares it to the user's 5700 profile to determine the command 5711 intended by the user 5700. The command 5711 then causes a cloud system 5712 to perform actions 5710 intended by the user's 5700 gesture 5707. The particular action 5710 performed may be defined by the user 5700 through a user interface 5709.
In one embodiment, a user's 5902 associated personal device 5904 transmits its MAC address (or other personal identifiers, such as the user's email address or phone number) to the store's wireless antenna 410. The detected MAC address is then provided to the computer system 5906, which uses the MAC address to identify the user 5902. Once identified, the cameras 5908 collect biometric information, such as an image, of the user 5902. The collected biometric information is then transmitted to the computer system 5906, which checks whether the biometric information matches that of the identified user 5902 to locate the user 5902. Once located, the camera 5908 collects images of gestures performed by the user 5902 and provides the images to the computer system 5906. The computer system 5906 processes the images to identify the gesture and, based on the associated user's 5902 identity, determine the action intended by the user 5902.
The computer system 6004 may utilize a variety of means to identify the user 6002 who performed the gesture. In one embodiment, the computer system 6004 may use facial-recognition technology to match the facial features in the images to profile images of users who have checked into the system (e.g., users whose smartphone MAC addresses have been detected). In another embodiment, the camera tracking the user 6002 may also track an associated user ID, which is transmitted along with the images to the computer system 6004. In yet another embodiment, the computer system 6004 tracks a user ID with an associated camera or video feed.
After identifying the gesture type and the user 6002 who performed the gesture, the computer system 6004 identifies the command intended by the user's 6002 gesture. In one embodiment, user 6002 has a profile containing a list of predefined gestures with associated commands. The computer system 6004 accesses the user's 6002 list and compares it to the identified gesture type to determine the intended command. For example, user 6002 may have defined a shoulder shrug to trigger a command to send a product information email to himself.
Once the command is determined, the computer system 6004 sends the command to a remote computer 6008 via a cloud 6006. The remote computer 6008 then acts according to the command, such as sending an email to the user 6002 with product information.
Once the product has been identified, the computer system 6204 may process the command to initiate a purchase transaction of the product. The purchase command, along with the user 6202 identity and a product ID associated with the product, may then be transmitted to a remote computer system 6208 through a cloud 6206. The remote computer system 6208 in response to the purchase command may retrieve the user's 6202 payment account (e.g., credit card information, bank account, etc.) based on the user identity. Prior to finalizing the purchase, however, the remote computer system 6208 may send a message (e.g., SMS) to the user's 6202 mobile device to confirm that the user 6202 would like to complete the purchase. Based on the user's 6202 response (e.g., by sending a response using his mobile device, his computer, a public terminal at a store, etc.), the remote computer system 6208 may proceed to finalize the purchase.
The user interface 6402 may allow the user 6401 to select actions/commands from a list of system-defined command or macros (i.e., sets of command), or allow the user 6401 to define his own commands/macros (e.g., the user 6401 may do so by configuring parameters of a command or write computer programs/scripts/codes/instructions. The user interface 6402 may also enable the user 6401 to select a gesture from a list of predefined gestures (e.g., may be presented by textual description, images, animations, sounds, etc.) using, for example, a drag-and-drop interface. In another example, the user 6401 may record his gesture using any conventional image capturing device 6407, such as a camera (e.g., webcam) or a portable user device (e.g., smartphone). The captured gesture images may be transmitted to a gesture processing software/hardware 6406 (e.g., it may be a client-side software installed on the user's 6401 computer or smartphone) included as part of the user interface 6402. The captured gesture images may then be used to define a gesture command and stored in the database 6404. When the user 6401 next performs the gesture at a store, for example, images of the performance may be compared to the recorded gesture stored in the database 6404 to identify the performed gesture and determine the user-defined corresponding command.
The user 6401 may also test out the gesture commands via the user interface 6402. The user 6401 may use the image capturing devices 6307 of his mobile device or computer to capture images of his own gesture. The captured images may be transmitted to the user interface 6402, which may process the images using the gesture processing software/hardware 6406. Once the gesture has been identified, the user interface 6402 may access the gesture command database 6404 via the website 6405 and attempt to find an action/command defined by the user 6401 that is associated with the identified gesture. If one is found, the system may commence execution of the action/command. Alternatively or additionally, the system may display to the user 6401 that the corresponding action/command has been found based on the gesture.
The packaged gesture data may then be transmitted to the user interface, 6540, which may be a website. The user interface computer (e.g., the web server or an associated backend server) may first attempt to determine whether the received gesture data is accelerometer/gyroscope data 6545, in which case the computer may identify the gesture by matching the gesture data with predetermined gesture models 6550. If the gesture data is not of the accelerometer/gyroscope data variety, then it may be a video or image capture 6555, in which case the gesture type may be determine via image processing 6560. Once the gesture type is determined, a database query may then be issued to determine the action/command that corresponds to that gesture type 6570. If the gesture command is user-defined, the database query may be limited to defined gesture commands associated with the user. Based on the query results or any comparison processing, the computer may determine whether a corresponding action/command is found 6575. If it is found, the computer may perform the action/command 6580. If no match is found, then the computer may prompt the user (e.g., via the user interface, the user's mobile device, etc.) to specify an intended action/command 6590.
Once the computer receives the video/images of the gesture 6640, it identifies the gesture via an image processing component 6650. Based on the identified gesture and the identity of the user who performed the gesture (e.g., which may be determined based on face-recognition technology), the computer queries a gesture command database to determine the user-defined command/action that is associated with the identified gesture 6660. The computer then determines whether a command/action is found based on the query results and/or comparison processing 6670. If a corresponding command/action is found, then the command/action may be sent to a remote computer if necessitated by the command/action 6680, or otherwise perform the command/action locally. If no command/action corresponding to the gesture is found, then the computer may perform no action or perform one or more remedial actions (e.g., alert the user that the command is not recognized) 6690.
Additional examples will now be described with regard to additional exemplary aspects of implementation of the approaches described herein.
A disk controller 6960 interfaces one or more optional disk drives to the system bus 6952. These disk drives may be external or internal floppy disk drives such as 6962, external or internal CD-ROM, CD-R, CD-RW or DVD drives such as 6964, or external or internal hard drives 6966. As indicated previously, these various disk drives and disk controllers are optional devices.
Each of the element managers, real-time data buffer, conveyors, file input processor, database index shared access memory loader, reference data buffer and data managers may include a software application stored in one or more of the disk drives connected to the disk controller 6960, the ROM 6958 and/or the RAM 6959. Preferably, the processor 6954 may access each component as required.
A display interface 6987 may permit information from the bus 6952 to be displayed on a display 6980 in audio, graphic, or alphanumeric format. Communication with external devices may optionally occur using various communication ports 6982.
In addition to the standard computer-type components, the hardware may also include data input devices, such as a keyboard 6979, or other input device 6981, such as a microphone, remote control, pointer, mouse and/or joystick.
Additionally, the methods and systems described herein may be implemented on many different types of processing devices by program code comprising program instructions that are executable by the device processing subsystem. The software program instructions may include source code, object code, machine code, or any other stored data that is operable to cause a processing system to perform the methods and operations described herein and may be provided in any suitable language such as C, C++, JAVA, for example, or any other suitable programming language. Other implementations may also be used, however, such as firmware or even appropriately designed hardware configured to carry out the methods and systems described herein.
The systems' and methods' data (e.g., associations, mappings, data input, data output, intermediate data results, final data results, etc.) may be stored and implemented in one or more different types of computer-implemented data stores, such as different types of storage devices and programming constructs (e.g., RAM, ROM, Flash memory, flat files, databases, programming data structures, programming variables, IF-THEN (or similar type) statement constructs, etc.). It is noted that data structures describe formats for use in organizing and storing data in databases, programs, memory, or other computer-readable media for use by a computer program.
The computer components, software modules, functions, data stores and data structures described herein may be connected directly or indirectly to each other in order to allow the flow of data needed for their operations. It is also noted that a module or processor includes but is not limited to a unit of code that performs a software operation, and can be implemented for example as a subroutine unit of code, or as a software function unit of code, or as an object (as in an object-oriented paradigm), or as an applet, or in a computer script language, or as another type of computer code. The software components and/or functionality may be located on a single computer or distributed across multiple computers depending upon the situation at hand.
It should be understood that as used in the description herein and throughout the claims that follow, the meaning of “a,” “an,” and “the” includes plural reference unless the context clearly dictates otherwise. Also, as used in the description herein and throughout the claims that follow, the meaning of “in” includes “in” and “on” unless the context clearly dictates otherwise. Further, as used in the description herein and throughout the claims that follow, the meaning of “each” does not require “each and every” unless the context clearly dictates otherwise. Finally, as used in the description herein and throughout the claims that follow, the meanings of “and” and “or” include both the conjunctive and disjunctive and may be used interchangeably unless the context expressly dictates otherwise; the phrase “exclusive or” may be used to indicate situation where only the disjunctive meaning may apply.
This application is a continuation of U.S. patent application Ser. No. 15/862,323 filed Jan. 4, 2018, entitled “Gesture Recognition Cloud Command Platform, System, Method, and Apparatus,” which in turn is a continuation of U.S. patent application Ser. No. 14/715,105 filed May 18, 2015 entitled “Gesture Recognition Cloud Command Platform, System, Method, and Apparatus” which claims priority to U.S. provisional patent application Ser. No. 61/994,793 filed May 16, 2014, entitled “Gesture Recognition Cloud Command Platform, System, Method, and Apparatus.” This application is related to PCT International Application Serial No. PCT/US13/20411, filed Jan. 5, 2013, attorney docket no. 196 W001|VISA-177/01 WO, entitled “Transaction Visual Capturing Apparatuses, Methods And Systems,” which in turn claims priority under 35 U.SC § 119 to U.S. provisional patent application Ser. No. 61/583,378 filed Jan. 5, 2012, attorney docket no. 196US01|VISA-177/00US, U.S. provisional patent application Ser. No. 61/594,957, filed Feb. 3, 2012, attorney docket no. 196US02|VISA-177/01US, and U.S. provisional patent application Ser. No. 61/620,365, filed Apr. 4, 2012, attorney docket no. 196US03|VISA-177/02US, all entitled “Augmented Retail Shopping Apparatuses, Methods and Systems.” The aforementioned applications are all hereby expressly incorporated by reference.
Number | Date | Country | |
---|---|---|---|
61994793 | May 2014 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15862323 | Jan 2018 | US |
Child | 17084259 | US | |
Parent | 14715105 | May 2015 | US |
Child | 15862323 | US |