There are many issues with the above stated method of pairing devices. First, a special hardware is needed at both ends to effectuate such pairing. Second, such pairing can only be used for predetermined specific functions. Also, the Bluetooth™ signals have wider range, hence, without a proper security, unintended pairing may occur. Still further, the paired devices must stay within a particular range after the pairing.
In one embodiment, a method of pairing a first device with a second device is disclosed. An image that includes a unique identifier for identifying a first device and a security code is generated and displayed on the first device. A second device captures the image using an image sensing device. The unique identifier and the security code is then sent to a server that is communicatively connected to the first device and the second device. The server communicates with the first device, using the unique identifier, to verify the security code.
In another embodiment, a method of pairing a first device with a second device is disclosed. Accordingly, an image that include encoded data is generated by the first device. The encoded data includes a unique identifier for identifying the first device and an arbitrary security code. The first device displays the image on a display. The second device captures the image using an image sensing device. The encoded data is decoded to generate a decoded data. The second device sends the decoded data to a server that is communicatively connected to the first device and the second device. Upon receiving the decoded data and using the unique identifier, the server communicates with the first device to verify the arbitrary security code.
In yet another embodiment, a system for pairing a first device with a second device is disclosed. The system includes a first device connected to a network. The first device is configured to generate an image that include encoded data. The encoded data includes a unique identifier for identifying the first device and an arbitrary security code, and to display the image on a display of the first device. The system also includes a second device connected to the network. The second device is configured to capture the image and to decode the encoded data to generate a decoded data, and to send the decoded data to a server that is communicatively connected to the first device and the second device. The server is connected to the first device and the second device through the network and the server is configured to communicate, using the unique identifier, with the first device to verify the arbitrary security code.
In yet another embodiment, a computer readable storage medium containing a program which, when executed, performs an operation of pairing a first device with a second device, is disclosed. The operation comprises capturing an image using an image sensor. The image includes a unique identifier of another device and a security code. The operation further includes sending the unique identifier and the security code to a server via a network and instructing the server to communicate with the another device using the unique identifier to verify the security code with the another device.
Other embodiments include, without limitation, a computer-readable storage medium that includes instructions that enable a processing unit to implement one or more aspects of the disclosed methods as well as a system configured to implement one or more aspects of the disclosed methods.
So that the manner in which the above recited features can be understood in detail, a more particular description, briefly summarized above, may be had by reference to embodiments, some of which are illustrated in the appended drawings. It is to be noted, however, that the appended drawings illustrate only typical embodiments and are therefore not to be considered limiting of its scope, for the claimed subject matter may admit to other equally effective embodiments.
In the following description, numerous specific details are set forth to provide a more thorough understanding of the described embodiments. However, it will be apparent to one of skill in the art that various embodiments may be practiced without one or more of these specific details. In other instances, well-known features have not been described in order to avoid obscuring the description.
Reference throughout this disclosure to “one embodiment” or “an embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. Thus, the appearances of the phrases “in one embodiment” or “in an embodiment” in various places throughout this specification are not necessarily all referring to the same embodiment. Furthermore, the particular features, structures, or characteristics may be combined in any suitable manner in one or more embodiments.
Some communication systems allow the user of a device, such as a personal computer, to communicate across a packet-based computer network such as the Internet. Such communication systems include voice over internet protocol (“VoIP”) communication systems. These systems are beneficial to the user as they are often of significantly lower cost than conventional fixed line or mobile networks. This may particularly be the case for long-distance communication. To use a VoIP system, the user installs and executes client software on her/his device. The client software provides the VoIP connections as well as other functions such as registration and authentication. In addition to voice communication, the client may also provide further features such as video calling, instant messaging (“IM”), SMS messaging, file transfer and voicemail.
One type of communication system for packet-based communication uses a peer-to-peer (“P2P”) topology. In one embodiment, to enable access to a peer-to-peer system, a user executes client software provided by a communication system software provider (or a third party vendor) on their computer (which includes any supported computing device, including smart phones), and register with the P2P system. When the user registers with the P2P system, the client software is provided with a digital certificate from a server. Once the client software has been provided with the certificate, then calls or other communications can subsequently be set up and routed between users of the P2P system without the further use of a server in the set-up. Instead, the client looks up the required IP addresses from information distributed amongst the client software on other end users' computers within the P2P system. Once the IP address of a callee's terminal has thus been determined, the caller's client software then exchanges certificates with the callee's client software. The exchange of the digital certificates (or user identity certificates, “UIC”) between users provides proof of the users' identities and that they are suitably authorized and authenticated in the P2P system. Therefore, the presentation of digital certificates provides trust in the identity of the users. It is therefore a characteristic of peer-to-peer communication that, once registered, the users can set up their own communication routes through the P2P system in a decentralized manner based on distributed address look-up and the exchange of one or more digital certificates, without using a server for those purposes. Further details on such a P2P system are disclosed in WO 2005/008524 and WO 2005/009019. VoIP or other packet-based communications can also be implemented using non-P2P systems that do use centralized call set-up.
A plurality of user terminals 102 are arranged to communicate over one or more of the networks 101,103,112. For merely illustration purposes only,
An example mobile device 102a is shown schematically in
The one or more transceivers 207 enable the mobile device 102a to access the one or more networks 101,103,112. For example, mobile device 102a may comprise a cellular wireless transceiver for accessing the mobile cellular network 103 via the base stations 104, and/or a wired or wireless modem for accessing the Internet 101. In the case of a wireless modem, this typically comprises a short-range wireless transceiver (e.g. Wi-Fi) for accessing the Internet 101 via the wireless access points 106.
Access to the Internet 101 may also be achieved by other means such as GPRS (General Packet Radio Service) or HSPA (High Speed Packet Access). At a higher level of the cellular hierarchy, the cellular network 103 comprises a plurality of cellular controller stations 105 each coupled to a plurality of the base stations 104. The controller stations 105 are coupled to a traditional circuit-switched portion of the mobile cellular network 103 but also to the Internet 101. The controller stations 105 are thus arranged to allow access to packet-based communications via the base stations 104, including access to the Internet 101. The controller stations 105 may be referred to for example as Base Station Controllers (BSCs) in GSM/EDGE terminology or Radio Network Controllers (RNCs) in USTM or HSPA terminology.
The memory 213 may comprise a non-volatile memory such as an electronic erasable and programmable memory (EEPROM, or “flash” memory) coupled to the processor 211. The memory stores communications code arranged to be executed on the processor, and configured so as when executed to engage in communications over one or more networks 101,103,112. The communications code preferably comprises a communication client application 110a provided by a software provider associated with the communication system. The communication client application 110a may be executed for performing communications such as voice or video calls with other user terminals 102 over the Internet 101, via a short-range wireless transceiver 207 and wireless access points 106, and/or via a cellular wireless transceiver 207, base stations 104 and controller stations 105 of the cellular network 103 as discussed above. However, one or more of the user terminals 102 involved could alternatively communicate via a wired modem, e.g. in the case of a call between a mobile terminal and a desktop PC.
As shown in
The communication system 100 also includes a server 120. In one embodiment, the server 120 is a Peer-to-Peer (P2P) communication server. Further, in one embodiment, the server 120 provides one or more of the following functions: call setup, call management, routing calls among the user terminals connected to the Internet 101 and routing calls among the user terminals connected to the Internet 101 and telephones connected to the PSTN network 112, etc. In one embodiment, the server 120 works cooperatively with the user stations with the help of a client software that runs on the user stations and/or the gateway 114.
Image recognition software 204 may be stored in memory 213 or in a separate memory not shown in
Images and shapes may encapsulate data that can be encoded by a reader in conjunction with selected configurations. For example, a triangle shape may be configured to convey a particular meaning between two entities. Other types of shapes may be used for conveying different types of information. Further, barcodes are commonly known in the art to comprise encoded data such that they may be optically read, and the encoded information decoded in order to read information about an item that the barcode is attached to.
There are two types of barcodes, linear barcodes and two dimensional (2D) barcodes, sometimes referred to as ‘matrix’ barcodes. One type of 2D barcode is the Quick Response (QR) barcode.
Suppose the user of the user terminal 102a wants to pair the LCD TV 122 with the user terminal 102a in order to use the LCD TV 122 for Audio/Video. Alternatively, the user may want to transfer the communication session between the user terminal 102a and the user terminal 102b-1 from the user terminal 102a to the LCD TV 122 without interrupting the ongoing call session and transparently to the user of the user terminal 102b-1.
In one embodiment, the user of the user terminal 102a would invoke a user interface (e.g., the UI 222 in
It should be noted that the examples in this disclosure are provided merely to impart a better understanding of the described embodiments. A person skilled in the art would realize that the systems and methods disclosed herein are directed to pairing of devices. Once paired, the devices may participate in many other activities, such as data transfer, one device controlling the other, etc. Unlike traditional pairing of devices using other technologies, such as Bluetooth™, the two devices, once paired, communicate via a server. Therefore, there is no location restriction on the paired devices. However, in another embodiment, the pairing include both server supported pairing, as described herein, and a device-to-device pairing, so that devices may exchange information directly as well as via the server. Among other, one advantage of the pairing methods described herein is that no special “pairing specific” hardware is required to effectuate the pairing of the two devices. Therefore, existing devices without any pairing specific hardware (e.g., Bluetooth™ hardware) may be configured to be paired using the methods described herein.
The LCD TV 122 is connected to the Internet 122 and can be located by the server 120 using a distinct identification. In one example, suppose the user selects (via the user interface) to transfer the video stream of the communication session from the user terminal 102a to the LCD TV 122. Upon the selection of an option, the user interface activates the camera 107 and the image recognition software 204 of the user terminal 102a. A second user interface is invoked on the LCD TV 122 either via the client software in the LCD TV 122 (or in the user terminal 102b-2) or via a separate software, hardware, or combination thereof. The user interface on the LCD TV 122 displays a coded pattern 124. In another embodiment, a number is displayed on the LCD TV 122 instead of a graphical pattern. The coded pattern may be a QR code or a barcode. The coded pattern, in one example, includes the IP address of the LCD TV 122 (or the user terminal 102b-2). Additionally, the coded pattern may also include another number or code. Alternatively, the coded pattern may include just one number. In another example, the LCD TV 122 may simply display one or more words.
If the coded pattern is configured to include the IP address and a security code, the camera 107 of the user terminal 102a, when brought close to the displayed coded pattern, deciphers the coded pattern and extracts the IP address and the security code. In an alternative embodiment, the user terminal 102a send the encoded pattern to either the server 120 or another external device that is connected to the Internet 101 for deciphering the encoded pattern. The client software of the user terminal 102a sends the IP address and the security code to the server 120 with the instructions that the video portion of the ongoing communication session be transferred to the LCD TV 122. Upon receiving said instructions, the server asks the LCD TV for the security code. In one embodiment, other data related to the devices and/or the user may also be sent to the server 120. However, in yet another embodiment, only the IP address and the security code are sent to the server 120. The handshake is complete when the security code provided to the server 120 by the user terminal 102a matches with the security code received directly from the LCD TV 122. In one embodiment, the security code included in the coded pattern may be transient and may be valid only for a selected period of time. In another embodiment, the security check may be optional and the coded pattern may include only the IP address.
In another embodiment, any other identification of the LCD TV 122 may be included instead of the IP address so long as the server 120 can locate the LCD TV 122 on the Internet 101 by that identification.
In another example, instead of a coded pattern, the LCD TV 122 may simply display the IP address and/or any other type of temporary or permanent identification code (such as the MAC number, machine network name, etc.) in plain text and the image recognition software 204 may be configured to recognize plain text characters.
In one example, the server 120 is configured to maintain separate data streams for audio, video and chat. Hence, if the user terminal 102a requests the server 120 to transfer the video to the LCD TV 122, the server 120 redirects the video stream to the LCD TV 122. In a partial transfer of the communication session, the part that is transferred to LCD TV 122 runs concurrently and synchronously with the remaining part of the communication session on the user terminal 102a. Similarly, if the complete session transfer was desired, the server 120 redirects all data streams to the IP address of the LCD TV 122.
In one embodiment, if the P2P communication system requires P2P users to log in using unique user identifications, the server 120 is configured to automatically send a login certificate (corresponding to the user of the user terminal 102a) to the LCD TV 122 (or the user terminal 102b-2) so that the user of the user terminal 102a is automatically logged into the LCD TV 122 (or the user terminal 102b-2).
It should be noted that the code pattern itself may also include the operation to be performed by the other user terminal after a successful pairing, thus diminishing a need for displaying multiple options on the user interface of the user terminal 102a. Instead, the user interface on the LCD TV 122 may generated different codes for different operations. Further, it should also be noted that in the above example, a session may be transferred from the LCD TV 122 to the user terminal 102a using the same method as described above. Similarly, a particular user station may be paired with a plurality of other user stations or devices, each performing either a distinct or duplicate function of a selected session, as for example, two devices may be setup to display the video part of the communication session. However, as stated above, the above examples are being provided for the easy understanding of the invention. The above embodiments may also be used for performing other operations that require a pairing of two or more devices.
The method of establishing a communication relationship between mobile device 102a and user device 102b will now be discussed with reference to
When the communication client 110a is executed on user device 102a the client 110a presents an option to the user 108a to enter a pattern recognition mode. At step 302, the user 108a may enter this recognition mode by making an appropriate input selection for example pressing a button on mobile device 102a, touching the appropriate section of display 212 or making a voice command or the like.
At step 304, the user 108a points the camera 205 of the mobile device 102a at the coded pattern. The user 108a then makes an appropriate selection on mobile device 102a to capture image data of the coded pattern at step 306. It will be appreciated that the captured image data comprises encoded information including IP address of the user terminal 102 (or any other identification to enable the server 120 to locate the user terminal 102b in the Internet 101) and also optionally a security code and/or encoded data defining a communication event related to the entity that generated the coded pattern.
As a result of the communication client 110a being in the pattern recognition mode, at step 308 the CPU 211 executes the image recognition software 204 to decode the image data and supplies decoded information to the communication client application 110a, the decoded information including decoded contact information and decoded data defining a communication event related to the entity that generated the barcode.
It will be appreciated that when the client 110a is not in the barcode recognition mode, and the user 108a uses camera 205 to capture image data the captured image data is stored in memory (whether internal or external) and no further action is taken.
At step 310, in response to receiving the decoded contact information, the communication client 110a establishes a communication relationship using the decoded information. That is the communication client establishes a communication relationship, for example transferring an ongoing communication session or a part thereof, from the mobile device 102a to the user device 102b. As described above, the server 120 takes part in the process of establishing the communication relationship between the mobile device 102a and the user device 102b.
Using Image Identifiers from Remote Sources
In the discussion that follows, it is to be appreciated and understood that any of the techniques described above can be utilized in connection with the approaches described below.
In one or more embodiments, security codes in the form of image information that is utilized to establish device pairing can be obtained from one or more remote sources, such as a server or backend server. Any suitable type of image information can be utilized. For example, a device can obtain, from the remote source, image information in the form of a raw image and use that as a security code as described below. Alternately or additionally, the device can obtain, from the remote source, image information in the form of a string of bits, and modulate the string of bits into an image identifier which can be used to establish the device pairing.
In the illustrated and described approach, the remote source or backend server generates image information that can be used as a security code to authenticate that the user of a particular device is physically located in the same place as another of the devices to which pairing is desired.
As an example consider
Next, and responsive to recognizing the pairing initiation, the second device requests image information, at 502, from the remote source which, in this case, is a server. The image information can comprise any suitable type of image information. For example, in at least some embodiments, the image information can be a string of bits which is returned to the second device and subsequently encoded into an image identifier. Alternately or additionally, the image information can comprise a visual image identifier that is returned to the second device. Any suitable type of image identifier can be utilized. In at least some embodiments, the image identifier comprises a coded pattern such as a QR code. Other decoded patterns can be utilized without departing from the spirit and scope of the claimed subject matter.
Responsive to receiving the request for the image information, the server returns the image information to the second device at 504. The second device now uses the image information to display, at 506, an image identifier for the first device. This can be done in any suitable way. For example, in the event the image information returned to the second device from the server is in the form of a string of bits, the second device can encode the string of bits into an image identifier and display the image identifier for the first device. Alternately or additionally, in the event that the image information returned to the second device from the server is in the form of a visual image identifier, the second device can display the visual image identifier for the first device.
The first device can now process the image identifier by capturing an image of the image identifier and sending data associated with the image identifier, at 508, back to the server. In the event that the image identifier is the visual image identifier received from the server, the first device can simply send data back to the server in the form of the visual image identifier. Alternately or additionally, in the event that the image identifier is associated with an encoded identifier, such as an encoded string of bits, the first device can decode the image identifier to arrive at, e.g., the string of bits, and send that data back to the server.
When the server receives the data, it can confirm that the data matches and that the device pairing is authentic. Accordingly, the server can make note of the device pairing and return an authentication, at 510, to the second device.
Once the device pairing has been completed, the server can optionally store details of this device pairing so that in the future if other device is reset, for example in a meeting room, the pairing could be automatically completed upon the first stage of initiation from either end.
In the above example, the first device, e.g., the handheld device or phone, initiated the device pairing with the television. It is to be appreciated and understood, however, that device pairing can be initiated by the television or, in the above example, the second device.
In one or more other embodiments, devices can reside in a not paired/available state and advertise themselves to the backend server. For example, multiple devices such as televisions and other electronic devices can advertise themselves as available. Then, different devices, such as a handheld device, can query the backend servers for a list of devices with which it can connect based on network proximity identification. Once the user selects a particular available device from the list, the process can continue as described above. In these embodiments, the devices for pairing may reside on the same network.
While the forgoing is directed to specific embodiments, other and further embodiments may be devised without departing from the basic scope thereof. For example, aspects of the described embodiments may be implemented in hardware or software or in a combination of hardware and software. At least some embodiments may be implemented as a program product for use with a computer system. The program(s) of the program product define functions of the embodiments (including the methods described herein) and can be contained on a variety of computer-readable storage media. Illustrative computer-readable storage media include, but are not limited to: (i) non-writable storage media (e.g., read-only memory devices within a computer such as CD-ROM disks readable by a CD-ROM drive, flash memory, ROM chips or any type of solid-state non-volatile semiconductor memory) on which information is permanently stored; and (ii) writable storage media (e.g., floppy disks within a diskette drive or hard-disk drive or any type of solid-state random-access semiconductor memory) on which alterable information is stored. Such computer-readable storage media, when carrying computer-readable instructions that direct the functions described above constitute embodiments of the present invention.
Number | Date | Country | Kind |
---|---|---|---|
1119407.3 | Nov 2011 | GB | national |
This application is a continuation-in-part of and claims priority under 35 U.S.C. §120 to U.S. patent application Ser. No. 13/293,242, filed on Nov. 10, 2011 which claims priority under 35 U.S.C. §119 or 365 to UK Patent Application No. 1119407.3 entitled “Establishing a Communication Event,” filed concurrently on the same day, the disclosures of which are incorporated by reference herein in their entirety. Traditionally, electronic devices are paired using Bluetooth™ technology. The term “pairing” means that two devices exchange some data to agree to work together to provide a predefined function. For example, a Bluetooth™ enabled mobile phone may be paired with a Bluetooth™ headset and upon a successful pairing, the headset provides speakers and microphone to the mobile phone.
Number | Name | Date | Kind |
---|---|---|---|
7254708 | Silvester | Aug 2007 | B2 |
7729489 | Lee et al. | Jun 2010 | B2 |
8224354 | De Vries et al. | Jul 2012 | B2 |
9288229 | Kaufman | Mar 2016 | B2 |
20030095521 | Haller et al. | May 2003 | A1 |
20040253923 | Braley et al. | Dec 2004 | A1 |
20050159132 | Wright et al. | Jul 2005 | A1 |
20060046719 | Holtschneider | Mar 2006 | A1 |
20060143455 | Gitzinger | Jun 2006 | A1 |
20060282649 | Malamud et al. | Dec 2006 | A1 |
20070094490 | Lohr | Apr 2007 | A1 |
20070173212 | Mergler | Jul 2007 | A1 |
20080049704 | Witteman et al. | Feb 2008 | A1 |
20080244721 | Barrus et al. | Oct 2008 | A1 |
20090199279 | Lange et al. | Aug 2009 | A1 |
20090240814 | Brubacher et al. | Sep 2009 | A1 |
20090247152 | Manne | Oct 2009 | A1 |
20090287922 | Herwono et al. | Nov 2009 | A1 |
20100043056 | Ganapathy | Feb 2010 | A1 |
20100115591 | Kane-Esrig | May 2010 | A1 |
20100197322 | Preston et al. | Aug 2010 | A1 |
20100227549 | Kozlay | Sep 2010 | A1 |
20100262696 | Oshiba | Oct 2010 | A1 |
20100278345 | Alsina et al. | Nov 2010 | A1 |
20110047607 | Chen et al. | Feb 2011 | A1 |
20110072263 | Bishop | Mar 2011 | A1 |
20110086593 | Hardacker | Apr 2011 | A1 |
20110092155 | Piemonte et al. | Apr 2011 | A1 |
20110093266 | Tham | Apr 2011 | A1 |
20110096174 | King | Apr 2011 | A1 |
20110179182 | Vadla Ravnas | Jul 2011 | A1 |
20110183614 | Tamura | Jul 2011 | A1 |
20110208659 | Easterly et al. | Aug 2011 | A1 |
20110219105 | Kryze et al. | Sep 2011 | A1 |
20110281523 | Oshiba | Nov 2011 | A1 |
20110295502 | Faenger | Dec 2011 | A1 |
20110296506 | Caspi | Dec 2011 | A1 |
20120011575 | Cheswick et al. | Jan 2012 | A1 |
20120017081 | Courtney et al. | Jan 2012 | A1 |
20120044057 | Kang et al. | Feb 2012 | A1 |
20120045994 | Koh et al. | Feb 2012 | A1 |
20120054046 | Albisu | Mar 2012 | A1 |
20120131186 | Klos et al. | May 2012 | A1 |
20120140925 | Bekiares et al. | Jun 2012 | A1 |
20120158581 | Cooley et al. | Jun 2012 | A1 |
20120158898 | van Deventer et al. | Jun 2012 | A1 |
20120184372 | Laarakkers et al. | Jul 2012 | A1 |
20120188147 | Hosein et al. | Jul 2012 | A1 |
20120189140 | Hughes | Jul 2012 | A1 |
20120198531 | Ort et al. | Aug 2012 | A1 |
20120214416 | Kent et al. | Aug 2012 | A1 |
20120278727 | Ananthakrishnan | Nov 2012 | A1 |
20120322376 | Couse | Dec 2012 | A1 |
20120324076 | Zerr et al. | Dec 2012 | A1 |
20130031275 | Hanes | Jan 2013 | A1 |
20130036461 | Lowry | Feb 2013 | A1 |
20130088649 | Yum | Apr 2013 | A1 |
20130110723 | Huang et al. | May 2013 | A1 |
20130115880 | Dal Bello et al. | May 2013 | A1 |
20130122810 | Kaufman | May 2013 | A1 |
20130124292 | Juthani | May 2013 | A1 |
20130125224 | Kaufman | May 2013 | A1 |
20130265857 | Foulds et al. | Oct 2013 | A1 |
20140256260 | Ueda | Sep 2014 | A1 |
20140305828 | Salvo | Oct 2014 | A1 |
Number | Date | Country |
---|---|---|
1638383 | Jul 2005 | CN |
101350723 | Jan 2009 | CN |
101872448 | Oct 2010 | CN |
1551140 | Jul 2005 | EP |
2005122651 | May 2005 | JP |
WO-0158080 | Aug 2001 | WO |
WO-2011010925 | Jan 2011 | WO |
Entry |
---|
“Non-Final Office Action”, U.S. Appl. No. 13/293,242, (Jun. 3, 2013),12 pages. |
“PCT Search Report and Written Opinion”, Application No. PCT/US2012/064576, (May 7, 2013),10 pages. |
Saxena, Nitesh et al., “Secure Device Pairing Based on a Visual Channel”, IEEE Symposium on Security and Privacy, 2006, Available at <http://eprint.iacr.org/2006/050.pdf>,(May 2006), pp. 1-17. |
“Final Office Action”, U.S. Appl. No. 13/293,242, (Sep. 20, 2013),17 pages. |
“Final Office Action”, U.S. Appl. No. 13/293,245, Mar. 25, 2014, 14 pages. |
“International Search Report and Written Opinion”, Application No. PCT/US2012/064577, Feb. 21, 2013, 22 pages. |
“Non-Final Office Action”, U.S. Appl. No. 13/293,245, Nov. 26, 2013, 13 pages. |
“Non-Final Office Action”, U.S. Appl. No. 13/293,245, Jul. 3, 2014, 17 pages. |
Goodrich, et al., “Loud and Clear—Human-Verifiable Authentication Based on Audio”, Proceedings of the 26th IEEE International Conference on Distributed Computing Systems, Available at <https://www.cs.duke.edu/˜msirivia/publications/icdcs.pdf>,Jul. 4, 2006, 15 pages. |
Prasad, et al., “Efficient Device Pairing using Human-Comparable Synchronized Audiovisual Patterns”, Proceedings of the 6th International Conference on Applied Cryptography and Network Security, Available at <http://www.cis.uab.edu/saxena/docs/sr07.pdf>,Jun. 3, 2008, 19 pages. |
Soriente, et al., “HAPADEP—Human-Assisted Pure Audio Device Pairing”, Proceedings of the 11th International Conference on Information Security, Available at <http://sproutics.uci.edu/papers/hapadep.pdf>,Sep. 15, 2008, 11 pages. |
“Foreign Office Action”, CN Application No. 201210585999.1, Jul. 27, 2015, 16 pages. |
“Final Office Action”, U.S. Appl. No. 13/293,242, Apr. 17, 2015, 23 pages. |
“Final Office Action”, U.S. Appl. No. 13/828,717, May 20, 2015, 20 pages. |
“Non-Final Office Action”, U.S. Appl. No. 13/293,245, Apr. 17, 2015, 18 pages. |
“Final Office Action”, U.S. Appl. No. 13/293,245, Nov. 5, 2014, 17 pages. |
“Foreign Office Action”, CN Application No. 201210597199.1, Jan. 29, 2015, 14 pages. |
“Non-Final Office Action”, U.S. Appl. No. 13/293,242, Dec. 4, 2014, 19 pages. |
“Non-Final Office Action”, U.S. Appl. No. 13/828,717, Nov. 28, 2014, 20 pages. |
“Final Office Action”, U.S. Appl. No. 13/293,245, Dec. 9, 2015, 17 pages. |
“Foreign Office Action”, CN Application No. 201210597199.1, Oct. 19, 2015, 13 Pages. |
“Non-Final Office Action”, U.S. Appl. No. 13/828,717, Dec. 17, 2015, 25 pages. |
“Notice of Allowance”, U.S. Appl. No. 13/293,242, Nov. 12, 2015, 13 pages. |
“Notice of Allowance”, U.S. Appl. No. 13/293,242, Dec. 2, 2015, 11 pages. |
Final Office Action, U.S. Apl. No. 13/828,717, Jun. 13, 2016, 25 pages. |
Number | Date | Country | |
---|---|---|---|
20130276079 A1 | Oct 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13293242 | Nov 2011 | US |
Child | 13828343 | US |