Implementations of various methods to couple a computerized ball device which acts as a mobile computing device to project towards waveguide eyeglasses or contacts which will then allow a user to view their world in a mixed reality, augmented reality, and virtual reality world and to allow for group collaboration on a shared coupled device. The device additionally has the ability to be docked in a drone cradle which creates a database map of the user's environment while not being utilized by the user for an immediate task. The device may also be attached to a wrist band for mobility. The device decouples the traditional design of head mounted virtual and mixed reality wherein the traditional design places the camera or central processing unit (“CPU”) addition to the standard eyeglass which then violates human social contracts or concern the user or nearby bystanders are being recorded. The mobile computer ball device is decoupled from the waveguide lens so that a third person view of the user can be obtained in addition to a first person view. The computerized ball device allows standard shape eyeglasses to be coupled with the device in both a private user to ball setting or a device (ball) to private group setting. The device is superior in design to existing implementations of mixed and virtual reality technology with bulky head mounted apparatus which add weight that disturbs social contracts and bio-mechanical movement and weight distribution and does not allow for both first person and third person view points in the waveguide display. Family use of the technology allows for natural sharing of a coupled group technology sharing of experience. Commercial and Industrial work team use of the technology allows for natural sharing of a coupled group sharing of technology experience. The technology is a superior implementation of the way humans collaborate and work in groups. The technology is a superior implementation to extend natural human interaction without drawing attention to the technology instead of the shared human experience. Current technology deployment for computerized devices has focused on individual experience at the expense of group interaction. The implementation of the computerized ball device with head mounted eyeglass experience allows for a superior group experience. The computerized ball device may be placed in a mutual group setting or alternatively docked on a drone that has the ability to map a database of the user's environment. The computerized ball device has both an internal server and CPU as well as network connection capability to interface with a remote application server.
The following descriptions and examples are not admitted to be prior art by virtue of their inclusion within this section.
Current implementations of mixed reality, augmented reality and virtual reality computing devices generally have bulky head mounted variations which are tailored to the individual experience and that experience generally isolates human beings from interacting face to face with other human beings. Such examples of these technologies are those implemented by Microsoft and their HaloLens product (A Trademark of Microsoft, Inc.), Oculus Rift VR (A Trademark of Facebook, Inc. by assignment), Sony's PlayStation VR (A Trademark of Sony, Inc.), HTC Vive (A Trademark of HTC, Inc.), Google Glass (A Trademark of Google, Inc.), Samsung's Gear VR (A Trademark of Samsung, Inc.), Razer's OSVR (A Trademark of Razer, Inc.) or Magic Leap's proposed systems (A Trademark of Magic Leap, Inc.). While these technologies are very impressive and useful, they fall short of allowing humans to interact in groups, face to face while using the technology without violating traditional human social contracts. Clearly these technologies stand the promise of greatly improved human productivity, but it has also come at a great expense of isolating humans from natural shared face to face interaction. Humans have rejected the model of a camera on the eyeglasses because others are concerned they are being recorded or watched. Some proponents of the camera on the eye glass model argue that humans will gain such benefit that the cost of objection for being recorded will be overwhelmed by the tremendous benefits of the technology. Further objections have become apparent because others in a group are not able to share the same information which creates a barrier of unfair information advantage. The implementation of the method allows users to “group” or “pair” multiple sets of eyeglasses with a single ball mobile computing device. The “group” or “paired” user experience allows all users in a group comfort in the social setting because everyone in the group is able to share the same information thus no social contract has been violated such as the objection to others recording a user without permission. While customization of user experience is useful for an individual experience, the implementations of technology have decreased the skills of humans to interact with eye to eye contact which has limited the ability to maintain or improve non-technology dependent human communication skills. Prior art generally has limited the ability to user both first person and third person viewpoints simultaneously.
Implementations of methods have been made in systems that provide augmented or virtual reality experiences, but do so in a manner that violates traditional human social contracts:
Current systems of technology to integrate computing into everyday life have largely been accepted when human users can trust the other users of the technology. The isolated nature of CPU, desktop, laptop or mobile computing is accepted because it remains obvious if one user is recording another user and typically this is forbidden in standard social settings because it violates a human social contract of trust between each other. Innovations such as Google Glass or Microsoft HaloLens or other augmented or virtual reality devices have struggled with adoption because they violate social contracts between humans and breach trust between humans. Information is power and if humans know they have different levels of power, they will typically not interact freely. This problem has caused a rift to form between humans and humans with technology. Further, augmented and virtual reality requires mapping a user's environment so that the virtual objects can interact with real objects that have been mapped into a database. Implementations of methods thus far have introduced non-sharing technologies which then pit one user against another user in turn causing low levels of adoption and violate human social contracts of trust. Lastly, implementations of methods and systems thus far have dealt with the first person perspective rather than the proposed invention which allows the first person and third person omniscient perspective, which, alongside methods and systems of artificial intelligence iterative feedback, the ability for a user to not only listen to music, but be part of the band, it allows not only the ability to watch a movie, but to be in the movie alongside existing actors and actresses, it allows not only the ability to watch a lesson in cooking or music or athletics, but to be alongside the teacher or instructing professional in an augmented or mixed reality environment.
The invention of systems and methods to accelerate the adoption of mixed reality, augmented reality and virtual reality is directed towards a mobile computing ball device that pairs with standard shaped eye glasses or contacts. The mobile ball device has the ability to project light and record/map environments while pairing to other users in a private group. The shared device has the function of building social contract trust so that users can equally build the groups information advantage rather than destroying group trust because the technology is potentially being used against a group member creating biased or skewed information advantage. The shared device also has the functional ability to display a first person or third person omniscient perspective with artificial intelligence iterative feedback, for a user to not only listen to music, but be part of the band, it allows not only the ability to watch a movie, but to be in the movie alongside existing actors and actresses, it allows not only the ability to watch a lesson in cooking or music or athletics, but to be alongside the teacher or instructing professional in an augmented environment, it allows not only the ability to invent something, but to be alongside the inventor or mentor in an augmented or mixed reality environment.
A plurality of users is communicatively paired with the device for group interaction settings. The plurality of users can command the device independently for the benefit of the private group. Accordingly, a plurality of users is then connected to the local or remote application server through the ball device. The ball CPU and application host or network server then connects users to a plurality of application functions. The ball CPU device is not only capable of pairing users to transmit data and electromagnetic light to users, but it also maps the user's environment for interaction with the application server. Users may use the device privately as one would use a mobile smart phone or they may pair with other users for group purposes to ease social contract stress. Shared or private applications may include but are not limited to calendar, photos, camera, videos, maps, weather, credit cards, crypto currency, notes, clocks, music, application hosting servers, settings, physical fitness, news, video conferencing, hologram conferencing, home security, home lighting, home watering systems, home energy or temperature settings, home cooking, phone, texting services, mail, internet, social networking, blogs, investments, books, television, movies, device location, flashlights, music tuners, airlines, transportation, identification, translation, gaming, real estate, shopping, food, commodities, technology, memberships, applications, web applications, audio media, visual media, touch media, general communication, internet, etc. . . . .
In one embodiment of the invention, the application server uses artificial intelligence algorithms to more efficiently provide the user with services or applications that are needed after the device has scanned the common area for the group or private user who has been paired with the device. The application server will display through the ball device most frequently used applications as well as recommending more efficient application for data patterns of the user or group of users.
In another embodiment of the invention, the application server uses artificial intelligence to analyze the user's biomechanical movement for such exercises as athletics, music, performing arts, cooking, teaching, conferencing etc. The visualization allows both the first person and third person omniscient perspective because of the positioning of the camera and projection relative to the waveguide lens. The ball CPU device will then be able to provide the user of the CPU ball device with an augmented reality companion to teach them. The augmented reality companion may be a professional at the service such as but not limited to tennis with John McEnroe or football with Nick Sabin or cooking with Julia Child or piano with Beethoven or acting alongside Harrison Ford in Star Wars or singing alongside Carrie Underwood on stage or in a private performance or many other examples. The ball CPU device will be able to analyze the user's movements and provide customized feedback and interaction with other projected human holograms and objects based on artificial intelligence inherent in the augmented reality system.
In another embodiment of the invention, the ball CPU device is docked on a drone that will then transport the device within a fixed set radius of the user to scan the associated area of the user into a database. The drone docked CPU ball has the ability to scan the user's area and navigate hallways, stairways, rooms, and outside places within a fixed radius of the user as well as the user's movements. The application server that the ball CPU device accesses locally or remotely will then be able to draw upon the database to speed processing times and reduce memory leak.
In another embodiment of the invention, multiple users who each have the CPU ball or cube device may lock the other devices to a common group share over a network if a group of users is working together. These bounded group lock features are in place to build technology trust amongst users while not violating human social contracts. This implementation allows for all users to experience the common recording in their viewing through the planar waveguide lens to disallow asymmetric information which violates human social contracts of trust.
In another embodiment of the invention, reflective light and electromagnetic waves are projected from the CPU ball device onto the eyeglasses then from the eyeglasses through the aqueous humor lens and vitreous humor which then project onto the photo receptors of the retina. The projected images to provide the augmented reality experience reside within the structures at reactive variable depths of the eyeglasses or contact lenses. The impulses then synapse with the optic nerve to transmit images to the brain. Adjustments to the pupil and lense allow for the adjustment of light transmission to the retina. These features augment the users visual experience from not only the natural world but also the CPU ball generated objects to make a mixed reality or augmented reality experience.
People use eye glasses or contacts because the cornea and the length of the eye are often mismatched. The name for the mismatch is correlation error in optometry. When one is near sighted the cornea does not project all the way back to the retina and when one is far sighted the image is projected too far past the retina. Glasses to correct near sightedness are thinner in the middle and thicker on the edges which allows light to diffuse and diverge and project images further back toward the retina. To correct far sightedness, the glasses are thicker in the middle and thinner on the edges which allows for light to converge more quickly and the images are brought in further to project onto the retina accurately which allows for focus. More or less glasses “trick” the eye by moving the light toward different focal points.
Coherent laser light and electromagnetic waves are projected from the mobile CPU ball. Refracted images from light transmission into the structures in the glasses or contacts make the holographic images inside the contacts or eyeglasses. Refraction allows for the laser image that is projected to the head mounted eye glasses or contacts to be properly transmitted to the optic nerves for interpretation.
Implementations of various technologies will hereafter be described with reference to the accompanying drawings. It should be understood, however, that the accompanying drawings illustrate only the various implementations described herein and are not meant to limit the scope of various technologies described herein.
The discussion below is directed to certain specific implementations. It is to be understood that the discussion below is only for the purpose of enabling a person with ordinary skill in the art to make and use any subject matter defined now or later by the patent “claims” found in any issued patent herein.
The following paragraphs provide a brief summary of various techniques described herein such as illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The embodiment illustrated in
The CPUs 1104 run or execute various instructions compiled by software and applications which are stored in the memory 1102 that perform various functions on the device 1125. The RF circuitry 1106 receives and sends RF signals. The RF circuitry 1106 converts electrical signals to/from electromagnetic signals and communicates with communications networks 150 and 130 and other communication devices via the electromagnetic signals. The RF circuitry may be comprised of but not limited to an antenna system, a tuner, a digital signal processor, an analogue signal processor, various CODECs, a SIM card, memory, amplifiers, an oscillator and a transceiver. The wireless communication components may use a plurality of standard industry protocols such as Global System for Mobile Communication (“GSM”), Voice over internet protocol (“VOIP”), long-term evolution (“LTE”), code division multiple access (“CDMA”), Wireless Fidelity (“WiFi”), Bluetooth, Post office Protocol (“POP”), instant messaging, Enhanced Data GSM Environment (“EDGE”), short message service (“SMS”), or other communication protocol invented or not yet invented as of the filing or publish date of this document.
The input/output subsystem 1110 couples with input/output peripherals 1105 and other control devices 1111 and other laser projection systems 1114 to control the device 1125. The laser projection system 1114 and camera 1113 take infrared tracking information feedback from the user 120 into the peripheral interface 1105 and CPU 1104 to combine the data with instructions in the CPU 1104 and memory 1102 that provide an iterative instruction for the graphical user interface which is displayed in the waveguide lens 240 or 210 after comparison with information in the memory from the database server 260. The input control devices 1111 may be controlled by user 120 movements that are recorded by the laser projection system 1114 and camera 1113. The input control devices 1111 may include instructions from the user 120 movements based on interactions with the graphical user interface module that is a hologram 460 in the waveguide lens 240 or 210. Holograms 460 may take the form of representations of such things as graphical user interface modules which represent virtual keyboards, voice recognition, translation services, physical buttons, dials, sliders, joysticks, video game controllers, physical sporting equipment 1091, user 1050, comparisons of the user 1050 to a professional athlete 1080 or inventor 980 or author 880 or chef 780 or actress 680 or actor 690 or a musician 590, fashion apparel designer 980, weapons, cooking utensils 780, musical instruments 580, microphones, tools, books 880, movies 660, music 580, ordering food 1360 or drink 1380 with geofence location services, or ordering clothing 1460 or 1490, ordering retail goods in a virtual store 1460 or 1490, comparing the user 1591 and 1590 to a professional athlete 1560 using artificial intelligence on the database server 160 through interaction with the device 140 through the wireless network 150 or 130, virtual shopping 1225, virtual restaurant drive thru 1224 or other equipment for completing tasks.
The audio circuitry 1108, one or more speakers 1107 and 1115 and the microphone 1119 provide an audio interface between the user and the device 1125. The audio circuitry 1108 receives audio data from the peripherals interface 1105, converting the data to an electrical signal, and transmits the electrical signal to the speakers 1107 and 1115. The speakers 1107 and 1115 convert the electrical signals to human audible sound waves which are mechanotransducted into electrical impulses along auditory nerve fibers and further processed into the brain as neural signals. The audio circuitry 1108 also receives electrical signals converted by the microphone 1109 from sound waves. The audio circuitry 1108 converts the electrical signal to audio data and transmits the audio data to the peripherals interface 1105 for processing. Audio data may be retrieved and/or transmitted to memory 1102 and/or the RF circuitry 1106 by the peripherals interface 1105. In some embodiments the RF circuitry may produce ultra-high frequency waves that transmit to wireless headphones which then convert the electrical signals to human audible sound waves which are mechanotransducted into electrical impulses along auditory nerve fibers and further processed into the brain as neural signals.
In some embodiments, the graphical user interface hologram objects 460, 592, 660, 760, 860, 960, 1060, 1360, 1492, 1592 and other objects and representations of humans or objects interact with the user 120 through the laser projection system 1114 and camera 1113 take infrared tracking information feedback from the user 120 into the peripheral interface 1105 and CPU 1104 to combine the data with instructions in the CPU 1104 and memory 1102 that provide an iterative instruction for the graphical user interface which is displayed in the waveguide lens 240 or 210 after comparison with information in the memory from the database server 260. Both a first person and third person omniscient perspective mode are available due to the remote separation of the camera 1113 and laser projection system from the waveguide lens 490. The user 1550 may compare their movement 1590 from a third person omniscient perspective to the movements of a professional athlete 1560. The artificial intelligence algorithms on the network 250 and database server 260 provide real time analytical feedback to the user 230 on the biomechanical motion differences between the user 1550 and professional athlete 1590.
The device 1125 also includes a power supply 1117 and battery 1118 for powering the various components. The USB port 1125 may be used for providing power to the battery 1118 for storage of power.
Optical sensors 1116 are used in conjunction with the camera 1113 and laser projection system 1114 to capture images and movements by the user 120 and its environment 110 to capture images or video. If other users 220 are near the user 230, both users may couple to the device 270 utilizing the laser projection system 1114, RF circuitry 1106 and optical sensors 1116 to allow both users 220 and 230 or a plurality of users to view the same digital projection which then reduces the technological problem of asymmetric information or the “are you recording me” syndrome and assimilates the technology into more common social conventions and unwritten trust contracts.
The location sensor 1123 couples with the peripherals interface 1105 or input/output subsystem 1110 to disable the device if the device 1125 is placed in a pocket, purse or other dark area to prevent unnecessary power loss when the device 1125 is not being used.
The device 1125 may also utilize data from an accelerometer 1121 to assist in the detection of user motion in addition to the infrared laser light projection system to more efficiently process the location of the user 120 in relation to the device 120 and other objects 660 projected by the laser light projection system 1114 to the waveguide lens 670.
The software instructions stored in the memory 1102 may include an operating system (LINUX, OS X, WINDOWS, UNIX, or a proprietary operating system) of instructions of various graphical user interfaces 592, 660, 760, 860, 960, 1060 or other variations which include instructions for object hologram embodiments of a calendar 1201, photos 1212, camera 1212, videos 1209, maps 1211, weather 1202, credit cards 1215, banking 1215, crypto currency 1215, notes, clocks 1213, music 1206, application hosting servers 1220, settings 1220, physical fitness 1203, news 1216, video conferencing 1209, home security 1208, home lighting 1208, home watering systems 1208, home energy 1208 or temperature settings 1208, home cooking 1207, phone 1214, texting services, mail 1218, internet 1217, social networking 1219, blogs 1219, investments 1210, books, television 1209, movies 1209, device location, flashlights, music tuners 1206, airlines 1205, transportation 1205, identification 1219, translation, gaming 1221, real estate 1208, shopping, food 1207, commodities 1215, technology 1217, memberships, applications 1220, web applications 1217, audio media 1206, visual media 1209, mapping or GPS 1211, touch media 1217, general communication 1214, internet 1217, mail 1218, contacts 1219, cloud services 1220, games 1221, translation services 1223, virtual drive through with geofence location services for nearby restaurants to allow advance ordering of food and payment 1224, virtual shopping with custom measurements through infrared scans 1225, etc. . . . and facilitates communication between various hardware and software components. Artificial Intelligence algorithms on the network 150 and database server 160 provide iterative analytical feedback to the user 120. Software applications are not limited to the aforementioned embodiments. The operating system includes various software components and/or drivers for controlling and managing general system tasks such as but not limited to memory management, storage device control, power management, etc. . . . )
Examples of other applications in
In conjunction with RF circuitry 1106, audio circuitry 1108, speakers 1107, microphone 1109, laser projection and infrared detection laser systems 1114 the graphical user interface may be used to communicate with other users through the hologram connection communication system 1214. Hologram communication may be recorded for delayed delivery or conducted live. The remote recording capability of the recording CPU computing device may allow the user 120 to transmit text like communication as holograms or in script text. The artificial intelligence of the CPU 140 and database 160 may be used to communicate in multiple languages to any user 120 on the network 150. Hologram communication may be used to speak a language such as English and have the receiving user hear the hologram speak in Chinese. The translation module 1223 compiles network neurons to provide phrase mapping across various literary styles of writing in various languages. Phrase mapping is used along with neural networks for translation services 1223. Cooking assisted lessons from famous chefs 190 are on the cooking application 1207 with the ability to provide feedback in cooking technique and recipe variation from artificial intelligence on the network 150 and database server 160.
In conjunction with laser projection and infrared detection laser systems 1114, optical sensors 1116, camera 1113, microphone 1109, the graphical user interface may be used to capture still images or video (including a video stream) or hologram representation and store them into memory 1102. The user 120 may modify characteristics of a still image or video or hologram or delete them. The user 120 may arrange, modify, manipulate, present, store of delete still images or videos or holograms. The user 120 may play back videos, hologram representations stored in memory 1102 or accessed through the network 150 and database server 160.
In conjunction with laser projection and infrared detection laser systems 1114, optical sensors 1116, camera 1113, microphone 1109, the graphical user interface may be used to browse the internet module 1217 with traditional still representations, video representation or holographic representation.
In conjunction with laser projection and infrared detection laser systems 1114, optical sensors 1116, camera 1113, microphone 1109, calendar module 1201, contact module 1219, email module 1218, the graphical user interface may be used to search for a plurality of widgets or modules. In some embodiments a widget or module may include an HTML5 (recent or future version of a Hypertext Markup Language), CSS (cascading style sheets) or JavaScript or any future set of instructions the CPU can process for the holographic representation of the module.
In conjunction with RF circuitry 1106, audio circuitry 1108, speakers 1107, microphone 1109, GPS receiver 1120, laser projection and infrared detection laser systems 1114 the graphical user interface may utilize the transportation module 1205 which provides mathematical algorithms to minimize cost and transportation time across a variety of transportation systems. One embodiment of the transportation module 1205 utilizes the hub and spoke model which is more efficient than the point to point model. The point to point model consists of 45 routes for a network of N nodes where N=10 destination points. The hub and spoke model in the transportation module 1205 utilizes N−1 routes to optimize on miles driven subject to the constraint of maximizing persons in each vehicle. The hub and spoke model in the transportation module 1205 utilizes only 9 routes to service all 45 possible point combinations in a system of N=10 destination nodes subject to the constraint of maximizing vehicle capacity.
In conjunction with RF circuitry 1106, audio circuitry 1108, speakers 1107, microphone 1109, GPS receiver 1120, laser projection and infrared detection laser systems 1114, the graphical user interface may utilize the Shopping module 1225 which provides consumer good design creators the ability to display merchandise with three dimensional holograms. The three dimensional holograms have the ability to show the user 120 in both the first person and third person omniscient perspective where the graphical user interface shows the clothing on the user 120 by utilizing the laser projection and infrared detection laser systems 1114 to detect the current size and shape of the user to appropriately match and size consumer goods in the shopping module with the user 120.
In conjunction with RF circuitry 1106, audio circuitry 1108, speakers 1107, microphone 1109, GPS receiver 1120, laser projection and infrared detection laser systems 1114, the graphical user interface may utilize the real estate module 1208 to buy or sell houses or rent properties on an exchange-based system. The user 120 may buy or sell capacity in a house on a temporary or permanent basis through the real estate module 1208. The graphical user interface utilizes the laser projection and infrared projection system 1114 to display three dimensional hologram renderings and property tours for rent or purchase.
In conjunction with RF circuitry 1106, audio circuitry 1108, speakers 1107, microphone 1109, GPS receiver 1120, laser projection and infrared detection laser systems 1114, the graphical user interface may utilize the stock trading and commodity, currency and bond module 1210 to buy or sell securities on an exchange based system. The user 120 may tour the factory facilities of an investment with a hologram experience provided by the company whom is seeking investment or is publically or privately traded. The user 120 may participate in three dimensional hologram analyst calls or presentation with the management in real time. The infrared and laser detection system 1114 may record the eye movements and body language of company management to assess risky behavior or lying from the artificial intelligence database server 160 and network 150 to alert the user 120 of potential risk analysis.
In conjunction with RF circuitry 1106, audio circuitry 1108, speakers 1107, microphone 1109, GPS receiver 1120, laser projection and infrared detection laser systems 1114, the graphical user interface may the weather module 1202 to see hologram renderings of rain, cloud formations, sunrise or sunsets, tornado warnings, hurricane warnings, flood watches, and other weather data relevant to the user 120. The weather module 1202 may synchronize with the user's 120 calendar 1201 to alert the user for activities that will correspond to the users 120 calendar 1201 to alert the user 120 of proper clothing to wear or to reschedule appointments or outside activities that are at odds with the then current forecast in the weather module 1202.
In conjunction with RF circuitry 1106, audio circuitry 1108, speakers 1107, microphone 1109, GPS receiver 1120, laser projection and infrared detection laser systems 1114, the graphical user interface may be used for gaming and social media module 1221 to link everyday activity and transactions with helping people in developing areas of the world. Additional description of the module to link everyday transactions with helping people in developing areas of the world can be found in U.S. patent application Ser. No. 15/266,326, “Implementations of various methods to create economic incentives to directly link users of a social network or social network reality game to actual projects and people within a charity or developing world area,” filed Sep. 15, 2016, the content of which is hereby incorporated by reference in its entirety.
In conjunction with RF circuitry 1106, audio circuitry 1108, speakers 1107, microphone 1109, GPS receiver 1120, laser projection and infrared detection laser systems 1114, the graphical user interface may be used for listening to music with the music module 1206 with a live band in the room. The holograms of band members and instruments may be interactive with the laser projection and infrared detection laser system 1114 to include the user 120 in the band as a singer or musician. Artificial intelligence feedback from the database server 160 and network 150 may provide the user with interactive feedback and dialogue with band members. The camera 1113 may also record renderings of the user 591 performing with the band for posting on social media or local or cloud based network platforms.
In conjunction with RF circuitry 1106, audio circuitry 1108, speakers 1107, microphone 1109, GPS receiver 1120, laser projection and infrared detection laser systems 1114, the graphical user interface may be used to order food with the cooking and food module 1207 and 1225. The GPS receiver locates prepared and raw ingredient alternative shopping locations within a geofence near the user 1350 where the user 1350 sees three dimensional holograms rendering of the prepared or raw product. The artificial intelligence algorithms based on analysis from the user's 1350 blood markers and saliva markers provide mathematically optimized food intake based on the science of the user's 1350 specific body chemistry. Based on the user's 1350 selection of the three dimensional rendering of the food product 1390, the food product 1390 may be delivered to the user's 1350 place of work or residence or another alternative meeting point including but not limited to, the place at which the food was produced or prepared. The transaction for the food product 1390 is connected to the user's 1350 financial account information where the user 1350 stored the information in the memory 1102 or database server 160 to provide instant payment to the vendor or service provider.
In conjunction with RF circuitry 1106, audio circuitry 1108, speakers 1107, microphone 1109, GPS receiver 1120, laser projection and infrared detection laser systems 1114, the graphical user interface may be used to order custom fit clothing 1460 or consumer goods 1490 based upon the measurements provided by the laser projection and infrared detection laser systems 1114. The three dimensional rendering of the clothing or consumer goods has the option to view with a model rendered by the merchant or from a third person omniscient perspective placing the clothing or consumer good on the user 1491 and 1490. Based on the user's 1450 selection of the three dimensional rendering of the clothing 1490 or consumer good 1460, the product may be delivered to the user's 1450 place of work or residence or another alternative meeting point including but not limited to the place at which the product was produced or prepared. The transaction for the product 1490 is connected to the user's 1450 financial account information where the user 1450 stored the information in the memory 1102 or database server 160 to provide instant payment to the vendor or service provider.
In conjunction with RF circuitry 1106, audio circuitry 1108, speakers 1107, microphone 1109, GPS receiver 1120, laser projection and infrared detection laser systems 1114, the graphical user interface may be used to perform analysis on the biomechanical movements of the user 1550 or 1590 or 1591 compared to the movements of a professional athlete 1560 or trained professional 1560 based upon the measurements of the user's 1590 movements provided by the laser projection and infrared detection laser systems 1114 and in contrast to the movements of the trained professional 1560. The change analysis between the two three dimensional holographic renderings are geared towards competitive training and instruction in any given field of athletics, music, work or other trained skill. The three dimensional rendering of the trained professional has the option to view stand alone or from a third person omniscient perspective placing the trained professional hologram on the user 1591 and 1590. Based on the user's 1550 movement of the three dimensional rendering of the trained professional 1560, an instructive video may be rendered for the user 1550 or for the user 1550 to post to social media or another cloud based network subject to copyright laws.
This application is a continuation of U.S. patent application Ser. No. 15/406,374, filed Jan. 13, 2017, and issued as U.S. Ser. No. 10/460,520 on Oct. 29, 2019; and a continuation of U.S. patent application Ser. No. 16/666,264, filed on Dec. 13, 2019. All of these applications are incorporated by reference herein in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
4476954 | Johnson et al. | Oct 1984 | A |
5412560 | Dennison | May 1995 | A |
5604676 | Penzias | Feb 1997 | A |
5726885 | Klein et al. | Mar 1998 | A |
5751245 | Janky et al. | May 1998 | A |
5973619 | Paredes | Oct 1999 | A |
6175831 | Weinreich et al. | Jan 2001 | B1 |
6240396 | Walker et al. | May 2001 | B1 |
6285999 | Page | Sep 2001 | B1 |
6356838 | Paul | Mar 2002 | B1 |
6421606 | Asai et al. | Jul 2002 | B1 |
6434530 | Sloane et al. | Aug 2002 | B1 |
6609103 | Kolls | Aug 2003 | B1 |
6618062 | Brown et al. | Sep 2003 | B1 |
6646659 | Brown et al. | Nov 2003 | B1 |
6663564 | Miller-Kovach et al. | Dec 2003 | B2 |
6708879 | Hunt | Mar 2004 | B2 |
6850907 | Lutnick et al. | Feb 2005 | B2 |
7090638 | Vidgen | Aug 2006 | B2 |
7373320 | Mcdonough | May 2008 | B1 |
7584123 | Karonis et al. | Sep 2009 | B1 |
7634442 | Alvarado et al. | Dec 2009 | B2 |
7680690 | Catalano | Mar 2010 | B1 |
7680770 | Buyukkokten et al. | Mar 2010 | B1 |
7711629 | Laurent et al. | May 2010 | B2 |
7747739 | Bridges et al. | Jun 2010 | B2 |
7756633 | Huang et al. | Jul 2010 | B2 |
7788207 | Alcorn et al. | Aug 2010 | B2 |
7886166 | Shnekendorf et al. | Feb 2011 | B2 |
7987110 | Cases et al. | Jul 2011 | B2 |
8024234 | Thomas et al. | Sep 2011 | B1 |
8065191 | Senior | Nov 2011 | B2 |
8121780 | Gerdes et al. | Feb 2012 | B2 |
8249946 | Froseth et al. | Aug 2012 | B2 |
8296335 | Bouve et al. | Oct 2012 | B2 |
8388451 | Auterio et al. | Mar 2013 | B2 |
8570244 | Mukawa | Oct 2013 | B2 |
8762035 | Levine et al. | Jun 2014 | B2 |
8798593 | Haney | Aug 2014 | B2 |
8918411 | Latif et al. | Dec 2014 | B1 |
8920175 | Black et al. | Dec 2014 | B2 |
8930490 | Brown et al. | Jan 2015 | B2 |
8968099 | Hanke et al. | Mar 2015 | B1 |
9011153 | Bennett et al. | Apr 2015 | B2 |
9020763 | Faaborg et al. | Apr 2015 | B2 |
9077204 | More et al. | Jul 2015 | B2 |
9092826 | Deng et al. | Jul 2015 | B2 |
9159088 | Dillahunt et al. | Oct 2015 | B2 |
9213957 | Stefik et al. | Dec 2015 | B2 |
9274540 | Anglin et al. | Jan 2016 | B2 |
9292764 | Yun et al. | Mar 2016 | B2 |
9387928 | Gentry et al. | Jul 2016 | B1 |
9389090 | Levine et al. | Jul 2016 | B1 |
9389094 | Brenner et al. | Jul 2016 | B2 |
9410963 | Martin et al. | Aug 2016 | B2 |
9436923 | Sriram et al. | Sep 2016 | B1 |
9528972 | Minvielle | Dec 2016 | B2 |
9558515 | Babu et al. | Jan 2017 | B2 |
9665983 | Spivack | May 2017 | B2 |
9880577 | Dyess et al. | Jan 2018 | B2 |
9952042 | Abovitz et al. | Apr 2018 | B2 |
9960637 | Sanders et al. | May 2018 | B2 |
9978282 | Lambert et al. | May 2018 | B2 |
10262289 | Vaananen | Apr 2019 | B2 |
10403050 | Konrardy et al. | Aug 2019 | B1 |
10460520 | Simpson et al. | Oct 2019 | B2 |
10533850 | Abovitz et al. | Jan 2020 | B2 |
10586084 | Burch et al. | Mar 2020 | B2 |
10685503 | Ricci | Jun 2020 | B2 |
10737585 | Chaudhary et al. | Aug 2020 | B2 |
10872381 | Leise et al. | Oct 2020 | B1 |
10832337 | Floyd et al. | Nov 2020 | B1 |
11138827 | Simpson | Oct 2021 | B2 |
11288563 | Lee et al. | Mar 2022 | B2 |
11296897 | Endress et al. | Apr 2022 | B2 |
11298017 | Tran | Apr 2022 | B2 |
20020004788 | Gros et al. | Jan 2002 | A1 |
20020013718 | Cornwell | Jan 2002 | A1 |
20020013761 | Bundy | Jan 2002 | A1 |
20020017997 | Wall | Feb 2002 | A1 |
20020065738 | Riggs et al. | May 2002 | A1 |
20020065766 | Brown et al. | May 2002 | A1 |
20020133456 | Lancaster et al. | Sep 2002 | A1 |
20020161689 | Segal | Oct 2002 | A1 |
20030055776 | Samuelson | Mar 2003 | A1 |
20030191725 | Ratliff et al. | Oct 2003 | A1 |
20030233311 | Bramnick et al. | Dec 2003 | A1 |
20040019552 | Tobin | Jan 2004 | A1 |
20040115596 | Snyder et al. | Jun 2004 | A1 |
20040249742 | Laurent et al. | Dec 2004 | A1 |
20040260581 | Baranowski et al. | Dec 2004 | A1 |
20050021346 | Nadan et al. | Jan 2005 | A1 |
20050027637 | Kohler | Feb 2005 | A1 |
20050132070 | Redlich et al. | Jun 2005 | A1 |
20050288974 | Baranowski et al. | Dec 2005 | A1 |
20050288986 | Barts et al. | Dec 2005 | A1 |
20070005224 | Sutardja | Jan 2007 | A1 |
20080033833 | Senior | Feb 2008 | A1 |
20080040232 | Perchthaler | Feb 2008 | A1 |
20080077309 | Cobbold | Mar 2008 | A1 |
20080129490 | Linville et al. | Jun 2008 | A1 |
20080140557 | Bowlby et al. | Jun 2008 | A1 |
20080157990 | Belzer et al. | Jul 2008 | A1 |
20080195432 | Fell et al. | Aug 2008 | A1 |
20080262892 | Prager et al. | Oct 2008 | A1 |
20090231687 | Yamamoto | Sep 2009 | A1 |
20090271236 | Ye et al. | Oct 2009 | A1 |
20090275002 | Hoggle | Nov 2009 | A1 |
20090276154 | Subramanian et al. | Nov 2009 | A1 |
20090287401 | Levine et al. | Nov 2009 | A1 |
20100042421 | Bai et al. | Feb 2010 | A1 |
20100114790 | Strimling et al. | May 2010 | A1 |
20100191834 | Zampiello | Jul 2010 | A1 |
20100211441 | Sprigg et al. | Aug 2010 | A1 |
20100217680 | Fusz et al. | Aug 2010 | A1 |
20100228574 | Mundinger et al. | Sep 2010 | A1 |
20100280748 | Mundinger et al. | Nov 2010 | A1 |
20100280884 | Levine et al. | Nov 2010 | A1 |
20100306078 | Hwang | Dec 2010 | A1 |
20110025267 | Kamen et al. | Feb 2011 | A1 |
20110098056 | Rhoads | Apr 2011 | A1 |
20110106660 | Ajjarapu et al. | May 2011 | A1 |
20110202418 | Kempton et al. | Aug 2011 | A1 |
20120023032 | Visdomini | Jan 2012 | A1 |
20120078743 | Betancourt | Mar 2012 | A1 |
20120101629 | Olsen et al. | Apr 2012 | A1 |
20120158762 | IwuchukWu | Jun 2012 | A1 |
20120303259 | Prosser | Nov 2012 | A1 |
20120323645 | Spiegel et al. | Dec 2012 | A1 |
20130024041 | Golden et al. | Jan 2013 | A1 |
20130147820 | Kalai et al. | Jun 2013 | A1 |
20130173326 | Anglin et al. | Jul 2013 | A1 |
20130179205 | Slinin | Jul 2013 | A1 |
20130191237 | Tenorio | Jul 2013 | A1 |
20130211863 | White | Aug 2013 | A1 |
20130265174 | Scofield et al. | Oct 2013 | A1 |
20130268325 | Dembo | Oct 2013 | A1 |
20130275156 | Kinkaid et al. | Oct 2013 | A1 |
20130304522 | Cundle | Nov 2013 | A1 |
20130311264 | Solomon et al. | Nov 2013 | A1 |
20140098009 | Prest et al. | Apr 2014 | A1 |
20140122190 | Wolfson et al. | May 2014 | A1 |
20140129302 | Amin et al. | May 2014 | A1 |
20140149157 | Shaam et al. | May 2014 | A1 |
20140162598 | Villa-Real | Jun 2014 | A1 |
20140220516 | Marshall et al. | Aug 2014 | A1 |
20140229258 | Seriani | Aug 2014 | A1 |
20140236641 | Dawkins | Aug 2014 | A1 |
20140244413 | Senior | Aug 2014 | A1 |
20140310019 | Blander et al. | Oct 2014 | A1 |
20140324633 | Pollak et al. | Oct 2014 | A1 |
20140349672 | Kern et al. | Nov 2014 | A1 |
20150006428 | Miller et al. | Jan 2015 | A1 |
20150016777 | Abovitz et al. | Jan 2015 | A1 |
20150161564 | Sweeney et al. | Jun 2015 | A1 |
20150178642 | Abboud | Jun 2015 | A1 |
20150198459 | MacNeille et al. | Jul 2015 | A1 |
20150206443 | Aylesworth et al. | Jul 2015 | A1 |
20150220916 | Prakash et al. | Aug 2015 | A1 |
20150241236 | Slusar et al. | Aug 2015 | A1 |
20150248689 | Paul et al. | Sep 2015 | A1 |
20150260474 | Rublowsky et al. | Sep 2015 | A1 |
20150269865 | Volach et al. | Sep 2015 | A1 |
20150324831 | Barua et al. | Nov 2015 | A1 |
20150371186 | Podgurny et al. | Dec 2015 | A1 |
20160041628 | Verma | Feb 2016 | A1 |
20160117657 | Forbes, Jr. et al. | Apr 2016 | A1 |
20160117756 | Carr et al. | Apr 2016 | A1 |
20160162989 | Cole et al. | Jun 2016 | A1 |
20160171891 | Banatwala et al. | Jun 2016 | A1 |
20160203422 | Demarchi et al. | Jul 2016 | A1 |
20160224935 | Burnett | Aug 2016 | A1 |
20160225115 | Levy et al. | Aug 2016 | A1 |
20160307276 | Young | Sep 2016 | A1 |
20160297316 | Penilla et al. | Oct 2016 | A1 |
20160300296 | Alonso Cembrano | Oct 2016 | A1 |
20160307288 | Yehuda et al. | Oct 2016 | A1 |
20160307373 | Dean et al. | Oct 2016 | A1 |
20160321609 | Dube et al. | Nov 2016 | A1 |
20160349835 | Shapira | Dec 2016 | A1 |
20160364679 | Cao | Dec 2016 | A1 |
20170019496 | Orbach | Jan 2017 | A1 |
20170039770 | Lanier | Feb 2017 | A1 |
20170046658 | Jones et al. | Feb 2017 | A1 |
20170046664 | Haldenby et al. | Feb 2017 | A1 |
20170046799 | Chan et al. | Feb 2017 | A1 |
20170046806 | Haldenby et al. | Feb 2017 | A1 |
20170048216 | Chow et al. | Feb 2017 | A1 |
20170061509 | Rosenberg et al. | Mar 2017 | A1 |
20170089710 | Slusar | Mar 2017 | A1 |
20170122746 | Howard et al. | May 2017 | A1 |
20170146360 | Averbuch | May 2017 | A1 |
20170232300 | Tran et al. | Aug 2017 | A1 |
20170243286 | Castinado et al. | Aug 2017 | A1 |
20170243310 | Dawkins | Aug 2017 | A1 |
20170249626 | Marlatt | Aug 2017 | A1 |
20170276500 | Margalit et al. | Sep 2017 | A1 |
20170293881 | Narkulla | Oct 2017 | A1 |
20170293950 | Rathod | Oct 2017 | A1 |
20180013211 | Ricci | Jan 2018 | A1 |
20180025417 | Brathwaite et al. | Jan 2018 | A1 |
20180046431 | Thagadur Shivappa et al. | Feb 2018 | A1 |
20180053226 | Hutton et al. | Feb 2018 | A1 |
20180053237 | Hayes et al. | Feb 2018 | A1 |
20180075695 | Simpson | Mar 2018 | A1 |
20180095471 | Allan et al. | Apr 2018 | A1 |
20180102053 | Hillman et al. | Apr 2018 | A1 |
20180111494 | Penilla et al. | Apr 2018 | A1 |
20180117447 | Bao et al. | May 2018 | A1 |
20180121958 | Aist et al. | May 2018 | A1 |
20180129276 | Nguyen | May 2018 | A1 |
20180143029 | Nikulin et al. | May 2018 | A1 |
20180173742 | Liu et al. | Jun 2018 | A1 |
20180173800 | Chang et al. | Jun 2018 | A1 |
20180278984 | Aimone et al. | Sep 2018 | A1 |
20180293638 | Simpson | Oct 2018 | A1 |
20180313798 | Chokshi et al. | Nov 2018 | A1 |
20180342106 | Rosado | Nov 2018 | A1 |
20180348863 | Aimone et al. | Dec 2018 | A1 |
20180357899 | Krivacic et al. | Dec 2018 | A1 |
20180365598 | Jamail | Dec 2018 | A1 |
20180365904 | Holmes | Dec 2018 | A1 |
20180374268 | Niles | Dec 2018 | A1 |
20190047427 | Pogorelik | Feb 2019 | A1 |
20190050634 | Nerayoff et al. | Feb 2019 | A1 |
20190066528 | Hwang et al. | Feb 2019 | A1 |
20190102946 | Spivack et al. | Apr 2019 | A1 |
20190108686 | Spivack et al. | Apr 2019 | A1 |
20190139448 | Marshall et al. | May 2019 | A1 |
20190143828 | Sawada et al. | May 2019 | A1 |
20190146974 | Chung et al. | May 2019 | A1 |
20190158603 | Nelson et al. | May 2019 | A1 |
20190160958 | Chaudhary et al. | May 2019 | A1 |
20190178654 | Hare | Jun 2019 | A1 |
20190188450 | Spivack et al. | Jun 2019 | A1 |
20190205798 | Rosas-Maxemin et al. | Jul 2019 | A1 |
20190228269 | Brent et al. | Jul 2019 | A1 |
20190236741 | Bowman et al. | Aug 2019 | A1 |
20190236742 | Tomskii et al. | Aug 2019 | A1 |
20190271553 | Simpson | Sep 2019 | A1 |
20190318286 | Simpson | Oct 2019 | A1 |
20190333166 | Simpson | Oct 2019 | A1 |
20190333181 | Simpson | Oct 2019 | A1 |
20190353499 | Stenneth | Nov 2019 | A1 |
20200027096 | Cooner | Jan 2020 | A1 |
20200098071 | Jackson | Mar 2020 | A1 |
20200125999 | Simpson | Apr 2020 | A1 |
20200151816 | Simpson | May 2020 | A1 |
20200156495 | Lindup | May 2020 | A1 |
20200160461 | Kaniki | May 2020 | A1 |
20200173808 | Beaurepaire et al. | Jun 2020 | A1 |
20200317074 | Miller et al. | Oct 2020 | A1 |
20200317075 | Yokoyama et al. | Oct 2020 | A1 |
20200389301 | Detres et al. | Dec 2020 | A1 |
20210012278 | Alon et al. | Jan 2021 | A1 |
20210318132 | Simpson | Oct 2021 | A1 |
20220100731 | Tirapu Azpiroz et al. | Mar 2022 | A1 |
20220122026 | Okabe et al. | Apr 2022 | A1 |
Number | Date | Country |
---|---|---|
107341968 | Nov 2017 | CN |
2539556 | Dec 2016 | GB |
2003177034 | Dec 2001 | JP |
2001041084 | Jun 2001 | WO |
2015059691 | Apr 2015 | WO |
2015161307 | Apr 2015 | WO |
2018024844 | Feb 2018 | WO |
2019134005 | Jul 2019 | WO |
2019183468 | Sep 2019 | WO |
2021163675 | Aug 2021 | WO |
Entry |
---|
Papa, U., & Del Core, G. (Jun. 2015). Design of sonar sensor model for safe landing of an UAV. In 2015 IEEE Metrology for Aerospace (MetroAeroSpace) (pp. 346-350). IEEE. |
PCT International Search Report and Written Opinion; PCT/US2020/027543; dated Jul. 1, 2020. |
PCT International Search Report and Written Opinion; PCT/US2020/023223; dated Jun. 19, 2020. |
PCT International Search Report and Written Opinion; PCT/US2020/023729; dated Jun. 18, 2020. |
PCT International Search Report and Written Opinion; PCT/US2020/021546; dated Jun. 8, 2020. |
PCT International Search Report and Written Opinion; PCT/US2020/018012; dated Apr. 21, 2020. |
PCT International Search Report and Written Opinion; PCT/US2020/012208; dated Mar. 24, 2020. |
Westerman; Longitudinal Analysis of Biomarker Data from a Personalized Nutrition Platform in Healthy Subjects; Nature, Scientific Reports; vol. 8; Oct. 2, 2018 (retrieved Jun. 10, 2020). |
Ahmed, et al.; Energy Trading with Electric Vehicles in Smart Campus Parking Lots; Applied Sciences; Sep. 7, 2018. |
Fitzsimmons; Uber Hit with Cap as New York City Takes Lead in Crackdown; New York Times; Aug. 8, 2018 (retrieved Feb. 29, 2020). https://www.wral.com/uber-hit-with-cap-as-new-york-city-takes-lead-in-crackdow/17755819/?version=amp?. |
Peters, et al.; Student Support Services for Online Learning Re-Imagined and Re-Invigorated: Then, Now and What's To Come; Contact North | Contact Nord; Sep. 2017. |
Soccer ball-shaped drone might be the safest flying robot yet https://mashable.com/2015/12/21/soccer-ball-drone/ ; Dec. 21, 2015. |
Fleishman; Use Parking Apps to Find Lots, Garages, Valet, and Meters; Macworld; Jul. 19, 2015. |
Borras, et al. Intelligent Tourism Reminder Systems: A Survey; Expert Systems with Applications 41; Elsevier; Jun. 9, 2014. |
Pentland; After Decades of Doubt, Deregulation Delivers Lower Electricity Rates; Forbes; Oct. 13, 2013 (retrieved Feb. 29, 2020). https://www.forbes.com/sites/williampentland/2013/10/13/after-decades-of-doubt-deregulation-delivers-lower-electricity-prices/#201d4a9c1d13. |
Sun, et al.; Real-Time MUAV Video Augmentation with Geo-Information for Remote Monitoring; 2013 Fifth International Conference on Geo-Information Technologies for Natural Disaster Management; pp. 114-118; IEEE; 2013. |
U.S. Appl. No. 60/035,205; filed Jan. 10, 1997; Page. |
The Wayback Machine, Interest Rate Swaps, https://web.archive.org/web/20171006212154/https://global.pimco.com/en/gbl/resources/education/understanding-interest-rate-swaps, 2016, pp. 1-7. |
Freight Derivatives—a Vital Tool For Your Business, https://www.reedsmith.com/-/media/files/perspectives/2007/02/freight-derivatives--a-vital-tool-for-your-business/files/freight-derivatives--a vital-tool-for-your-business/fileattachment/etcfreightderivativesavitaltoolforyourbusiness.pdf (Year: 2007), Energy, Trade & Commodities, pp. 1-3. |
Barry, Kieth, App lets drivers auction public parking spaces, Wired, Aug. 11, 2011, pp. 1-4. |
Jiang, Landu, et al., Sun Chase: Energy-Efficient Route Planning for solar-powered Evs, IEEE 37th international conference on distrubuted computing systems, 2017, pp. 1-11. |
Netlingo, https://web.archive.org/web/20170122184857/https://www.netlingo.com/word/electronic-exchange.php,dated Oct. 22, 2017. |
Laseter, Tim, “B2B benchmark: The State of Electronic Exchanges”, Tech & Innovation, dated Oct. 1, 2001. |
Directed Graph, https://en.wikipedia.org/wiki/Directed_graph, pp. 1-6, 2022. |
About IBM Food Trust, https://www.ibm.com/downloads/cas/E9DBNDJG, pp. 1-17, 2019. |
IBM Blockchain Transparent Supply, https://www.ibm.com/downloads/cas/BKQDKOM2, pp. 1-14, Aug. 2020. |
Radocchia, Samantha, 3 Innovative Ways Blockchain Will Build Trust In The Food Industry, https://www.forbes.com/sites/samantharadocchia/2018/04/26/3-innovative-ways-blockchain-will-build-trust-in-the-food-industry/? sh=65bc79f42afc, Forbes, pp. 1-5, Apr. 26, 2018. |
Change the World, https://fortune.com/change-the-world/2019/ibm/, Fortune Media IP Limited, pp. 1-5, 2022. |
IBM Food Trust, https://www.constellationr.com/node/17601/vote/application/view/588, Constellation Research Inc., pp. 1-4, 2010-2022. |
Dey, Somdip, et al., FoodSQRBlock: Digitizing Food Production and the Supply Chain with Blockchain and QR Code in the Cloud, https://www.mdpi.com/2071-1050/13/6/3486/htm, MDPI, pp. 1-27, Mar. 22, 2021. |
Ramasubramanian, Vasant, “Quadrasense: Immersive UAV-based cross-reality environmental sensor networks,” phD diss., Massachusetts Institute of Technology, pp. 1-75, 2015. |
Wyzant, https://web.archive.org/web/20190327185429/https://www.wyzant.com/hotitworks/students,Wyzant tutoring, pp. 1-13 , Mar. 27, 2019. |
PCT International Search Report and Written Opinion; PCT/US2021/065855; dated Mar. 29, 2022. |
PCT International Search Report and Written Opinion; PCT/US2022/012717; dated Mar. 30, 2022. |
Zhao, et al., Incentives in Ridesharing with Deficit Control, Proceedings of the 13th International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2014), May 5-9, 2014, pp. 1021-1028. |
PCT International Search Report and Written Opinion; PCT/US2022/027077; dated Nov. 1, 2022. |
Wei, et al. “impact of aircraft size and seat availability on airlines demand and market share in duopoly markets” Published by Elsevier, 2005, pp. 315-327. |
PCT International Search Report and Written Opinion; PCT/US2022/052969; dated Mar. 21, 2023. |
Little, T.D., et al., On the Joys of Missing Data, Journal of pediatric psychology, 2014, pp. 151-162. |
Honaker, J., et al., What to do About Missing Values in Time-Series Cross-Section Data, American Journal of Political Science, Sep. 6, 2008, pp. 561-581. |
Westerhoff, Market Depth and Price Dynamics: A Note, University of Osnabrueck, Department of Economics Rolandstrasse 8, D-49069 Osnabrueck, German, Mar. 30, 2004, pp. 1-8. |
PCT International Search Report and Written Opinion; PCTUS2022/051998; dated Mar. 8, 2023. |
EP23153137.7 European Search Report, dated May 24, 2023, pp. 1-10. |
EP20787830.7 European Search Report, dated May 12, 2023, pp. 1-10. |
Zheyong, Bian, et al., “Planning the Ridesharing Route for the First-Mile Service Linking to Railway Passenger Transportation,” Joint Rail Conference, Apr. 2017, pp. 1-11. |
EP23168879.7 European Search Report, dated Jul. 5, 2023, pp. 1-13. |
Number | Date | Country | |
---|---|---|---|
20230043986 A1 | Feb 2023 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 16666264 | Oct 2019 | US |
Child | 17971535 | US | |
Parent | 15406374 | Jan 2017 | US |
Child | 16666264 | US |