PERSONALIZED All MEDIA SEARCH

Information

  • Patent Application
  • 20210365509
  • Publication Number
    20210365509
  • Date Filed
    May 19, 2020
    4 years ago
  • Date Published
    November 25, 2021
    2 years ago
  • CPC
    • G06F16/9535
    • G06F16/337
    • G06F16/173
  • International Classifications
    • G06F16/9535
    • G06F16/17
    • G06F16/335
Abstract
A system and method for receiving user information from a user, wherein the user information includes location information, preferences, biographic data, previous purchase information, and previous search information. Next, creating a micro-moment profile of the user in real time from the user information, wherein the user information is updated in real time. Then, determining a plurality of micro-moment attributes from the micro-moment profile and searching all content available to the user using the plurality of micro-moment attributes to determine a unique set of search results for the user. Lastly, sending the unique set of search results to an electronic device associated with the user.
Description
TECHNICAL FIELD

The present disclosure relates generally to a system and method for searching, separately across all media for a user and more specifically as an individual to gain content, information, data and the like using SmartData, and all media to deliver and receive individualized content, information, data and the like in the search of all content.





BRIEF DESCRIPTION OF THE DRAWINGS

The embodiments herein may be better understood by referring to the following description in conjunction with the accompanying drawings in which like reference numerals indicate identically or functionally similar elements, of which:



FIG. 1 illustrates a representation of the Smart Platform Architecture;



FIG. 2 illustrates a representation of all content components;



FIG. 3 illustrates the Hyperspatial Device Specific Content & Communication;



FIG. 4 illustrates the Smart Platform Content Delivery Services and Networks;



FIG. 5 illustrates the Human-Centric Digital Data & Information Content Curation Communications (C3) Dimensionality; and



FIG. 6 illustrates the Smart Platform Server.





DETAILED DESCRIPTION

The embodiments herein will redefine search as it has come to be understood as the process of interactively searching for and retrieving requested information via a computer from databases that are online. Additionally, current search engines are a software structure that searches for information on the World Wide Web. The search outcomes are commonly presented in a line of results often referred to as search engine results pages. The current search engine information may be a mix of web pages, images, and other types of files. Some search engines also mine data available in databases or open directories. Search engines also maintain real-time information by running an algorithm on a web crawler.


The uniqueness of the present invention is that it is personalized in every aspect based on the specific user's individual micro-moment profile. Utilizing every form of media will enable a wider list of search parameters, extending the information that is today available from an amalgam of web pages, images and other types of files. The search results can be from any available content the user has access to. Content can includes film, video, audio, print, social media, public documents, and includes all content available through the networks of things. The present invention considers All Media communications including, but not limited to, verbal and non verbal, sighted and non sighted. Advertising media, broadcast media, digital media, electronic media, hypermedia, mass media, multimedia and social media are other forms of all media.


Personalizing each form of media is a process that takes into consideration the individual and their personal micro-moment attributes. An individual is commensurate to a snowflake in that no two are exactly alike. Individuals that commence to gain information from the present invention are defined by their uniqueness and are therefore linked to the specific information that they are searching. There may be a myriad of people in the world searching at the same time for information on understood by referring to the following description in conjunction with the accompanying drawings.


SmartData, as disclosed in U.S. application No. 62/367,772 on Jul. 28, 2016 entitled SmartData, is hereby incorporated by reference, is a system and method for determining useful (valued) and contextually relevant data, components of data, statistics, facts, figures, numbers, documented information, observations from real time and or videoed and reviewed after the fact that can be derived from singular or in combination from any form of live or recorded video, audio, audio-video, sensors developed for a myriad of uses, coupled with one or more of GPS, Compass, accelerometer, including but not limited to partial and entire findings, conclusions, of networked and or individual devices, instruments, devices, databases, analytics, visualization and processing architecture. SmartData determines which data, captures all data, and analyzing all varied data sources to determine appropriate information from the data. Reducing data and model complexity, and simplifying analyses provides unique and meaningful insights into many of the Big Data ecosystems problems faced by individual users, corporate, academic, vocational and avocational and all other organizations, groups and all other group including governmental bodies, agencies, groups etc. today. SmartData aims to resolve non-linear real world problems from received and captured data.



FIG. 1
380 shows a representative architecture for the Smart Platform that integrates the HCDDI 350 and an Affective Platform 100. Users are distributed throughout a hybrid network and appear/disappear based upon their associated activities, and can process, share, cache, store, and forward personally- or group-secured content with digital key security encryption, enabled by a Unified Security Management 273 process and Ribbon Encoding/Decoding 306, 307. User A 130, with a smart device 135, may contain all-media content (e.g., video, audio, images, print etc.) that can be partial or complete in nature and securely concealed or embedded using an individual or shared embedded code. Users who do not possess embedded code security access or the proper digital signatures will have limited access to content. Users can be any arbitrary process, requiring data, information, content or connectivity supporting a goal or objective within the Smart Platform. User A 130, User B 140, and User C 120, with process, components, and things ranging from smart devices 135, 145, 125 and IoT 415-417 devices to the Networks of Things 418. Each user can be equally represented as individuals, family, groups, organizations, enterprises, and governments system 380. Users communicate through various server applications, as represented by Server 200, and with the Smart Platform Content Delivery System (not shown). The SmartData Processing Unit (disclosed in U.S. No. 62/367,772), in concert with contextual and predictive activity modeling, data sources 260, 261, and the Affective Sense-Making & Micro-Moment Digital Information (ASMDI) Filter 349, render applicably filtered content 262-269 to each user, or from each user, or any combination there within, as determined through an asynchronous or synchronous network and communication architecture.


To further exemplify the Smart Platform, imagine User A 130 is a celebrity or artist. User A 130 embeds a Ribbon or an embedded code into their content 267-269 for distribution for their fan club members (i.e., group) can consume and unlock embedded content using their personal and/or group Ribbons (security keys). User A forms a participatory commerce (PC) chain (Participatory Commerce, U.S. Ser. 62/411,666, filed on Oct. 23, 2016 and is hereby incorporated by reference) with Brands, Advertisers, Merchants, and Media, resulting in embedded content from PC partners, provided the user's HCDDI and affective dimensional space are within the hyperspatial user dimensions. This content may be in the form of personal messages, video-audio clips of a new songs, discount and merchandise offers that can be redeemed nationally (AdPlexing, U.S. Ser. No. 62/268,003, which is hereby incorporated by reference) or locally (LocalPlexing, U.S. Ser. No. 62/358,538, which is hereby incorporated by reference), including behind-the-scenes information, content or access etc. User B 140 may be any individual or a group with a shared ribbon or shared embedded code, enabling the decoding and utility of hierarchical embedded content whereby the hierarchy is determined by the Ribbon or embedded code access security and authorization. Although User B has access to the concealed and embedded content, the ASMDI Filter 349 determines the level and type of content decoding. User B 130 may need to be Inspire 351 more than the other affective HCDDI components [Convince 352, Support 353, Educate 354, Inform 355, and Entertain 356], resulting in a “message” of encouragement from User A 130. The level of inspiration is determined from several affective sources including IoT 415-417 and the Network of Things 418 disclosed in Networks of Things, U.S. Ser. No. 62/358,546, which is hereby incorporated by reference, components, things, and processing. Real-time analyses from the SmartData Processing Unit 201, including AI, Machine and Deep Learning, determine the appropriate content 262-269 for each user during all activities, events, and situations. Contextual and hyperspatial conditions influence the nature and level of information (all-media) encoding/decoding and user content rendering. This results in a truly self-aware and cognitive content curation and communication 312 using small devices, hybrid networks, and the Small Platform. Note that an electronic smart device 130, 140, 120 may be a smart-phone, tablet, laptop, wearable technology, television, electronic glasses, watch, embedded device, or other portable electronic device that incorporates sensors such as at least one of camera, microphone, accelerometer, GPS, or transmission capability via wireless telephone, Wi-Fi, Bluetooth, NFC, etc.


The system and method connects all devices. The devices may include anything within the internet of things, but also anything within the Network of Things. The devices may be smart phones, smart devices, laptops, computers, televisions, television boxes, smart boxes, wearable technology, embedded devices, electronic devises, tablet, electronic glasses, watch, embedded device, or other portable electronic device that incorporates sensors such as at least one of camera, microphone, accelerometer, GPS, or transmission capability via wireless telephone, Wi-Fi, Bluetooth, NFC, etc. The network of things deals with the interconnectivity of all hardware, but focuses on the self-forming networks of knowledge and sense-making, that are particular to an individual, group, organization, or institution. Since each network entity on the IoT possesses a processor and a communication mechanism, from RFIDs products to computers, each device can communicate with a Ribbon or unique identifier with different levels of complexity that defines the networks entity in relationship to the formed network. The formation of each network depends on the network structure and query that are dependent on Ribbon identifier or unique identifies, node structure of the network, and utility.


Also the Network of Things allows the use of embedded codes to link all kinds of things to the internet, to the individual, group, organization, and institution, to their needs, wants and desires, either now or in the future. The use of an embedded code within a printed publication, an embedded code within an advertisement, an embedded code within a radio broadcast, etc. may all be connected to the internet and tracked as the embedded code is read by different devices, which may or may not be autonomous in nature. The Network of Things is communication agnostic and operates on top of any communication network/system such as RFID, Zigbee, WiFi, and TCP/IP Internet based protocol, creating personalized networks, component of networks, and things.


The system and method includes content from multiple sources. FIG. 1 demonstrates some of the foundational elements comprising the Smart Platform, which offers interactive, personalized, and affective (emotional) content and networks that are ubiquitous, systematic, individualized that form dynamically (automatic or self-forming) or manually 400, that gather, analyze, subscribe, delivery, and share arbitrary content. The content type is determined based upon platform user's profiles, Ribbons (e.g., disclosed in, US Serial 2014/0303991, 2004/0117255), and SmartData (disclosed in U.S. No. 62/367,772), whereby the ubiquitous network communications utilizes People 302, Processes 303, Things 304 and Data 305, in conjunction with Server 200, to cognify content across User A 401, User B 402, Group A 403, Group B 404, Object A 405, Object B 406, Process A 407, Process B 408, AI Agent A 409, AI Agent B 410, Brand A 411, Brand B 412, Advertiser A 413, Media A 414 and more.


Cognitive curation and contextual delivery of consumable all media content requires not only the proper security and personalization, but a mechanism that can determine and predict the dynamic evolution of a user's personalization requirements. FIG. 4360 describes the content selection process that is dependent on a user's need, as a function of context, activity, time, location and more. The multidimensional Human-Centric Digital Data & Information 350 user requirements are hyperspatial influenced and determined in part from previous experiences (preferences, opinions, moods, past content consumption), as well as real-time experiences and behaviors. As shown in FIG. 5, User A 130 has a unique HCDDI profile 361 that includes a subset of all content, across the HCDDI profile dimensions, including Inspire 351, Convince 352, Support 353, Educate 354, Inform 355, and Entertain 356. A bounding graphical surface shown in FIG. 4360 displays the relative importance of each HCDDI 350 category for User A 130 and User B 140, as 362 and 365, respectively. Each user's HCDDI subset is unique and depends on an individual's goals, objectives, and experiences. The hyperspatial dynamics can be seen by comparing User A 130 and User B 140 HCDDI subsets, as 362 and 365 that evolve to 372 and 375. The HCDDI subsets change ascribing a real-time cognitive user sensitivity for consumable content. This optimizes each user's experiences, and thereby empowers the user to achieve deterministic goals and objectives. The HCDDI user subsets represent a dynamical multidimensional parameter space that is solved for each user in real-time, employing parametric and non-parametric solutions, contained within the SmartData Processing Unit 201.


User A's and User B's content, as exemplified by equations 369 in FIG. 3360, are functions of Brand A 411, Advertiser A 413, and Media A 414 content, given HCDDI user subset. Although equations 369 explicitly show a limited set of content providers, the architecture and methodology is extensible and supports an arbitrary group or set of content providers across all media, and may include but is not limited to OTT 262, Media Content 263, Advertising Content 264, Brand Content 265, Merchant Content 266, Celebrity/Artist Content 267, Social Media Content 268, Individual Content 269, and more. The Smart Platform transforms existing content engagement mechanism, systems and networks into Cognitive Content and Communications (C3) Networks & Processes.


An example of networked components, things, and devices includes users, groups, objects, processes, AI agents, brands, advertisers, and media channels. Each entity has goals, plans, strategies, and actions and may cooperatively (or non-cooperatively) engage any entity to accomplish a single or collective goal and objective. FIG. 1380 shows 3 arbitrary users (e.g., individuals, family, groups, organizations, enterprises, and governments) participating on the Smart Platform in either a synchronous or asynchronous manner, which may cooperative or not cooperative. FIG. 3360 demonstrates how cognitive content changes as a function of each user, their context, environment and personalized preferences. And in particular, how cooperative engagement leads to an optimization for all users as highlighted in the commerce chain of users, brands, advertisers, and media 369. Embedded content (i.e., content inside content), that is engineered specifically for each user, changes the landscape of content and communications 312, by providing specialized data, and information to each user, or collection of users, that have the proper authorization to decode and view this content. This secure and non-invasive data, information, and content allows Brand A 411, Advertiser A 413, and Media A 414 to engage users with intelligent content tailored to user's needs, including discounts, offers, videos, experiential opportunities, enhancing communication and engagement.


User A 130, with smart device 135, has either no content, a partial representation, or a complete copy of content on their smart device, representing Individual Content (not shown). Similarly, User B 140 and User C 120 have smart devices and a representation of personalized content. Each user on the Smart Platform is provided a dynamic and personalized Ribbon. User A and User B are assigned Ribbons RUA and RUB, respectively. Alternative the Ribbons may be embedded codes. User C is a contributor to the network (e.g., anonymous sign-in) and shares content but has limited access without a Ribbon. Users communicate with Server 200 and/or between peers in order to manage their data, information, applications, and content during their daily activities and life events. Advertiser A, Brand A, and Media A distribute content whereby each one has content that is delivered separately or collectively to the Smart Platform, and then to users. Content is processed in real-time by the SmartData Processing Unit 201 and is stored, cached, forwarded, distributed and delivered to User A and User B with Ribbons RUA and RUB, respectively, and to User C without any Ribbon Encoding 306. Server 200 uses individualized HCDDI 350 subset specifications to determine the appropriate content filtering with recommendations determined by numerical methods employing AI, Machine Learning, and Deep Learning Neural Nets. Content 262-269, when offered, selected or served to users is encoded (content inside content), allowing for a Media Content overlay or hierarchy 308. Each user, although receiving the same content has access to a varied level of diverse embedded content. In this scenario, User C 120 receives generic content 309; however, User B 140 decodes more content 310, while User A 130, due to its user attributes, receives more content 311. The content ranges from videos, images, audio, and print to discounts from brands on selected purchases and more. Without a Ribbon or embedded code, User C 120 is not able to share and engage fully in the cognitive hybrid network.


The Ribbon or embedded code is the user's network securitized content and communications key, which unlocks specialized data and information. The Ribbon Encoding 306 process supports arbitrary encoding of embedded or conceal information inside ail media without increasing the payload of the content. Examples include hierarchical content embedding with arbitrary levels of encoding and encryption such as video inside video, with audio encoded with images such as discounts. The level of Decoding 307 is based upon the user's Ribbon and profile, including hyperspatial context. The Smart Platform Ribbon Encoding 306 and Decoding 307 are non-invasive and seamless using multi-level encrypted transform methods.


User A 130, with smart device 135, creates a mobile content delivery networks with its peers. User B 140 and User C 120. The nature of the shared information is defined by the personalized Ribbon and SmartData, as a shared key-value Ribbon pair that allows for either the unlocking or locking of embedded content inside content. Ribbon RUA (User A 130) and RUB (User B 140) share an encrypted content key, which enables peer-to-peer communication between both User A and User B, content sharing, and the unlocking of embedded content. Each Ribbon also provided for secure server content sharing. Both users share curated content based upon their individual and shared Ribbon, including a mesh topology between mobile, Super-Nodes, and static (Server 201-206) CDNs. If the desired content is not found among its peers, or Super-Nodes, then the CDN servers provide the necessary information. Depending upon the distribution of nodes, partial or complete information or content replication occurs within the mobile or static smart, devices and servers. Human-Centric Digital Data & Information 350 is utilized with architecture described in FIG. 1380, to determine the data, information, curated content and media content distribution. This information also allows for CDN Services to predict CDN services, pre-caching, and improving network quality of service (QoS). In all cases, Server 200 mediates each Smart Platform users and Cognitive Content Communications Networks & Processes.



FIG. 6 shows a schematic block diagram of an example Server 200 that is used with one or more embodiments of this invention and described herein. Server 200 initiates the Smart Platform Analyzer 381, Cognitive Content Curation Analyzer 313, HCDDI Analyzer 314 and Content Distribution and Delivery Network Analyzer 257 as part of the Smart Platform's Cognitive Content and Communications Networks & Processes (this included manual, automatic and arbitrated services with synchronous and asynchronous information and dynamics) and manages the main processes and curates, delivers, distributes Ribbonized (security keys) content across hybrid networks. Several key components, as part of the Server 200 processes, facilitated cognitive content curation and communication, such as Activity Analyzer 247, Location Analyzer 244 and Spatiotemporal Analyzer 249 that work together to help determine the activity, context, location, time, including the behavioral and historical significance of the user's state as given by the SmartData Analyzer 256. Coupled with the Preference Analyzer 246 and Decision Module 254, the Recommender Module in concert with the HCDDI 314, Cognitive Content Curation 313, and the Content Distribution and Delivery Network 257 Analyzers, determines the most salient content including how to deliver the content through unique user Interface and Artificial Intelligence Agents 232. Participatory Commerce Analyzer 208 is used to determine the nature of a commerce chain whereby discounts, offers, and opportunities are offered to users. All communications are securitized by the Uniform Security Management 273 module and the Ribbon Analyzer 234, which provides dynamic cognitive curated content communications and distribution.


The server 200 may comprise one or more network interfaces 210 (e.g., wired, wireless, etc.), at least one processor 220, and a memory 240 interconnected by a system bus 250, as well as a power supply (e.g., battery, plug-in, etc.). Additionally, or in combination server 200 may be implemented in a distributed cloud system. The network interface(s) 210 contain the mechanical, electrical, and signaling circuitry for communicating with mobile/digital service provider 135, 145, 125 (FIG. 1) and/or any communication method or device that enables and supports (synchronous or asynchronous) Smart Platform users (e.g., smart device which can be a smart-phone, tablet, laptop, smart television, wearable technology, electronic glasses, watch, or other portable electronic device that incorporates sensors such as at least one of camera, microphone, GPS, or transmission capability via wireless telephone, Wi-Fi, Bluetooth, NFC, etc.). The network interfaces may be configured to transmit and/or receive data using a variety of different communication protocols. Note, further, that server 200 may have two different types of network connections 210, e.g., wireless and wired/physical connections, and that the view herein is merely for illustration.


The memory 240 comprises a plurality of storage locations that are addressable by the processor 220 and the network interfaces 210 for storing software programs and data structures associated with the embodiments described herein. The processor 220 may comprise hardware elements or hardware logic adapted to execute the software programs and manipulate data structures. An operating system 242, portions of which are typically resident in memory 240 and executed by the processor, functionally organizes the server 200 by, inter alia, invoking operations in support of software processes and/or services executing on the device. These software processes and/or services may comprise a Signal Analyzer 252, Preference Analyzer 246, Location Analyzer 244, Activity Analyzer 247, Spatiotemporal Analyzer 249, Artificial Intelligence Agents 248, Ribbon Analyzer 234, Interface Agents 232, Network Analyzer 233, Brand Connector 231, and Digital Concierge 230, Network of Things Analyzer 253, Decision Module 254, Recommender Module 355, SmartData Analyzer 256, Content Distribution and Delivery Network 257, Market Basket Module 356, On-Demand Opt-on Service 251, Participatory Commerce Module 208, Smart Platform Analyzer 381, Cognitive Curation Analyzer 313, Media Search Analyzer 810, Human-Centric-Digital-Data & Information (HCDDI) Analyzer 314 that all play critical roles in interpreting and supporting manual, automatic and or arbitration processes as shown in FIG. 1380, FIG. 3360, and as described herein. Note a centralized memory 240 is shown in FIG. 4, alternative embodiments provide for the process to be specifically operated within the network interfaces 210. Another alternative uses a plurality of stand-alone servers, with each server performing steps of a single or multiple processes. Signal Analyzer 252, Preference Analyzer 246, Location Analyzer 244, Activity Analyzer 247, Spatiotemporal Analyzer 249, Artificial Intelligence Agents 248, Ribbon Analyzer 234, Interface Agents 232, Network Analyzer 233, Brand Connector 231, and Digital Concierge 230, Network of Things 253, Decision Module 254, Recommender Module 355, SmartData Analyzer 256, Content Distribution and Delivery Network 257, Market Basket Module 356, On-Demand Opt-on Service 251, Participatory Commerce Module 208, Smart Platform Analyzer 381, Cognitive Curation Analyzer 313, Human-Centric-Digital-Data & Information (HCDDI) Analyzer 314, Media Search Analyzer 810, and Affective Sense-Making Micro-Moment Analyzer 315, all perform multiple analyses utilizing various techniques (AI, ML, Deep Learning, EM, GA, NN and others), however each analysis may be performed by a separate process. Each separate process may be performed by a single server or a combination of servers that may or may not be distributed in the cloud.


Network Analyzer 233 plays a critical role in the Smart Platform architecture and manages the hybrid network communications. Combined with the Smart Platform Analyzer 381, Cognitive Curation Analyzer 313, HCDDI Analyzer 314, and the Affective Sense-Making Micro-Moment Analyzer 315, the Network Analyzer 233 facilitates the communication and delivery, distribution, caching of cognitive content with Ribbonized security, from the Ribbon Analyzer 234 and the Unified Security Management 273 system.


Media Search Analyzer 810 assists in searching all media, associated with each user throughout their life events and activities. Although Search engines have a robust heritage from the early 90's often times the information a user request or query is not always found within the first several search page results. Starting from a text-based-ranking-system, search engines have evolved to include more sophisticated algorithms including one of the more influential algorithms referred to as Page Rank, which determines the importance of a web page from the number of pages that link to the page. With billions of websites and even more webpages, Page Ranks are constantly recalculated, and that is for only the known webpages excluding the Dark Net. Depending on how these websites and webpages are created, and searched by users, affects the Page Rank of each webpage. The idea of PageRank is to regard web surfing as a Markov chain. Imagine a collection of webpages indexed as S={1, 2, . . . , N}, and suppose we have a personalization vector u∈RN×1 which signifies a generic surfer's preference for each page in S. The Page Rank algorithm provides a nice interpretation whereby if the i-th page has outlinks, the surfer will move to one of them with an equal probability; if no outlinks i exist, the surfer will move to any page in S at probabilities according to preference.







Q


(

i
,
j

)


=

{







G


(

i
,
j

)






l
=
1

N







G


(

i
,
l

)









u


(
j
)










if






G


(

i
,
m

)



=



1





for





some





m




S
i






otherwise






G


(

i
,
j

)




=

{





1




0








if





there





is





an





outlink





from





I





to





j

;

or





0





otherwise










This gives rise to a homogeneous discrete-time Markov chain with a simply the transition probability matrix Q for this Markov chain. The idea of PageRank is that the “importance” conferred by a webpage can be defined as the limiting probability associated with it as each time step increase. The limiting distribution can be interpreted as the proportion of time the surfer spends on each webpage. Personalization is accomplished by associated search history that determines the strength behind search entities such as query, responsive documents, search session, time, advertisement, anchor text and associated domains with localization and device specific information. If the user is not logged into Google, cookies are used to determine search histories. Combining this information, personalized and neutral search results are presented through SERPs (Search Engine Result Pages).


SERPs, including sematic web and social media, only represents a limited view and understanding of an individual user's preferences, emotions, decisions, likes, dislikes and more. SmartData combines affective micro-moment data from smart devices, which determines a unique affective value or utility each action (e.g., media consumption, advertising, search, activities, commerce etc.). SmartData incorporates data from any and all media (e.g., TV, OTT, radio, audio, images, social, documents, print etc.) that has been viewed or created by individuals or enterprises including hyperspatial context. These unique data sources are stored and used to determine and predict the individual needs using probabilistic methods incorporating a multidimensional personalization matrix, that is devices and hyperspatial sensitive, as well as incorporating native and explicit user all-media consumptions. Temporal and spatial features of all-media sources and individual queries are merged with a user's SmartData data and profile, and with Machine Learning (e.g., Bayesian, genetic, classification, and regression analysis and) and MCMC (e.g., Markov Chain Monte Carlo) methods, a next generation of personalized all media search is invented, which is called SmartSearch.


SmartSearch assists and performs an expansive knowledge discover and anticipation based upon hyperspatial affective SmartData, which is indexed and ranked similar to webpages but whose indexing is in a higher-dimensional space. This new Knowledge Map enables SmartSearch to perform a search through an affective experience space, merged with neutral subspace options and selections as represented by a universal user profile. SmartSearch finds content and information based upon a truly personalized user experience that may or may not be part of the main-stream search or trend. This removes search results that are limited or irrelevant to the user that require a constant restructuring or creation of new query key words and or topics. Using natural language, touch, gestures or any form of Human-Machine-Interface, a user can initiate a personalized content search using SmartSearch and receive an All-Media Personalized Search (AMPS) result spanning all content and media types, independent of time, location, or experience.


Content has a broad definition and simply is what the end-user derives value from and can refer to the data and information provided through a particular medium, the way in which the information is presented, as well as the added features included in the medium in which that information is delivered. The medium is the content channel through which content is delivered and affects how the end user perceives the content. With increases in connectivity, smart devices, and high-speed, high-bandwidth communications, content delivery and diversity is increasing exponentially. FIG. 4350 shows the complexity and dimensionality of content delivery and consumption. Individuals are presently overloaded with digital noise from emails, social media, news as well as advertisers which produce over 5000 ad impressions per person per day. This overload is expected to increase as the IoT (Internet of Things) develops. Making sense of the information, from any arbitrary network infrastructure, is needed to facilitated the next generation of digital experiences that benefit the user personally thought out their life experiences, and may include applications such as Health and Wellness, Education, Marketing & Advertising, Financial Technologies (FinTech), Entertainment and more.



FIG. 2 illustrates a schematic block diagram of an all media search components that may be used with one or more embodiments of this invention and described herein. User A performs a search. User A enters an adaptive query 211 and that is communicated with the Content Search Selection/Delivery 212. A key word optimization 213 is applied to the search terms. Hyperspatial attributes 214 are applied to the keyword optimization 213. User command content 216 and user queried content 217 are stored in respective containers.


The user adaptive query input is applied on the content delivery network 222, which works in connection with server 200. An all media search function 221 is applied on server 200 and works with SmartData processing unit 201 to obtain data from database(s) 260 and SmartData Database(s) 261.


The all media space 219 includes print, audio, images, videos, text, email, social media, wearable data, billboards, internet history, purchase history, saved data, etc. The hyperspatial device content and communication 360 and affective cognitive capacity management 380 connect the content with SmartData Processing unit 201.


Illustratively, the techniques described herein may be performed by hardware, software, and/or firmware. It will be apparent to those skilled in the art that other processor and memory types, including various computer-readable media, may be used to store and execute program instructions pertaining to the techniques described herein. Also, while the description illustrates various processes, it is expressly contemplated that various processes may be embodied as modules configured to operate in accordance with the techniques herein (e.g., according to the functionality of a similar process). Further, while the processes have been shown separately, those skilled in the art will appreciate that processes may be routines or modules within other processes.


While there have been shown and described illustrative embodiments that provide for enhancing advertisements sent to users based on location, it is to be understood that various other adaptations and modifications may be made within the spirit and scope of the embodiments herein. For example, the embodiments have been shown and described herein with relation to user's personal device. However, the embodiments in their broader sense are not as limited.


The foregoing description has been directed to specific embodiments. It will be apparent; however, that other variations and modifications may be made to the described embodiments, with the attainment of some or all of their advantages. For instance, it is expressly contemplated that the components and/or elements described herein can be implemented as software being stored on a tangible (non-transitory) computer-readable medium (e.g., disks/CDs/RAM/EEPROM/etc.) having program instructions executing on a computer, hardware, firmware, or a combination thereof. Accordingly, this description is to be taken only by way of example and not to otherwise limit the scope of the embodiments herein. Therefore, it is the object of the appended claims to cover all such variations and modifications as come within the true spirit and scope of the embodiments herein.

Claims
  • 1. A method, comprising: receiving user information from a user, wherein the user information includes location information, preferences, biographic data, previous purchase information, and previous search information;creating a micro-moment profile of the user in real time from the user information, wherein the user information is updated in real time;determining a plurality of micro-moment attributes from the micro-moment profile;searching all content available to the user using the plurality of micro-moment attributes to determine a unique set of search results for the user; andsending the unique set of search results to an electronic device associated with the user.
  • 2. The method of claim 1, wherein the all content include web pages, images, film, video, audio, print, advertising media, broadcast media, digital media, electronic media, hypermedia, mass media, multimedia, and social media.
  • 3. The method of claim 1, further comprising: linking the user to specific information that the user is searching to increase information stored with user information to further define the user's micro moment profile.
  • 4. The method of claim 1, further comprising: linking user information learned from a network of things to the user's micro moment profile.