The following relates generally to identifying complementary objects from images and, in particular, to using images containing objects determined based on an input, to identify such complementary objects.
Product recommendations may be used by merchants to notify users of related products or similar products that have been sold with selected products by other users. These product recommendations may be used to drive additional sales, e.g., when a user is interacting with an e-commerce platform.
Embodiments will now be described with reference to the appended drawings wherein:
Current methods used to provide product recommendations may rely on techniques such as collaborative filtering, which recommends items by matching the current user to a group of similar users for whom there is data on products purchased or otherwise engaged with. Other techniques may involve presenting a buyer with items that are commonly purchased together or items that are related. These solutions may be adequate for finding ‘similar’ or ‘related’ products, but for a buyer looking to make purchases of products that complement each other, these techniques may be inadequate for, or even incapable of, providing useful and personalized results. For example, for a user looking to purchase a complete outfit such as a blazer, dress shirt, belt, dress pants and shoes, these techniques may find similar or related products for one piece of the outfit, but not necessarily the other complementary products.
Additionally, the aforementioned solutions may be biased towards products for which there is a large dataset of buyer engagement signals, e.g., products that have been part of a catalog for a long time and have been browsed many times. As a result, discoverability may be low for products that have recently been introduced to the catalog.
To determine and provide collections of complementary products to those determined from an input, the proposed solution uses products that are determined to be similar to the input-related product as a proxy, such that images of the similar products are used to find such complementary products and may result in a collection of complementary products based on the images. Matching the input to the similar products may use embeddings of the input and of the object(s) in an appropriate embedding space. The collection of complementary products may then be clustered, and the clusters used in various user experience flows, such as to pre-emptively provide product recommendations or to refine subsequent search results related to the input product, e.g., by linking and refining subsequent search queries when the size of the matching cluster satisfies a size threshold.
While certain examples provided herein relate to products and product queries, the principles equally apply to any object. The methods described herein may therefore be applied to not only e-commerce applications but also to various services (e.g., travel, media), design (e.g., technical, fashion, architectural) and other user experiences that involve determining objects that are complementary to one another.
In general, the method determines at least one object from an input, e.g., a search query, an object selection, or other user input that identifies the at least one object. A similarity engine or equivalent tool may be used to find objects that are similar to the input object. For example, matching the input to object(s) may use embeddings of the input and of the at least one object in an appropriate embedding space. The object(s) determined from the input can be located from an available source, such as in a local database (e.g., catalogue) or an external source. For each of the at least one object, there exists one or more associated images and this may include a set or collection of images. For example, a similar object may have a set of photos stored in a catalogue entry, which can be used as a proxy to find related objects. That is, each object related to the input may itself be associated with more than one image having other products, e.g., complementary products.
The related objects may be identified by analyzing the contents of the image(s) using embeddings of the input and of the object(s). This can be done in various ways based on the tools and associated data that is available. For example, in a simple case, the image may already be tagged with metadata associated with various objects shown in the image and this metadata may be used to infer that an object is complementary to the input object. In other examples, a large language model (LLM) may be used to look for certain complementary objects based on a candidate list of complementary objects. The candidate list may be taken from a taxonomy of objects that may be related, a rule set, a question set, or some other guide or instruction as to how to query the LLM to determine a refined set of complementary objects that are in the image. The LLM may take the form of an artificial intelligence chatbot that the user or the system may interact with. In other examples, image processing can be applied directly to the image to perform object detection using machine learning models trained on how to identify such objects from images. For example, bounding boxes may be used to identify objects for further analysis, e.g., by referencing a taxonomy of potentially related objects.
In this way, the associated images of the similar products may be used as a proxy to determine, from a candidate list of potential complementary objects, those that are able to be identified in the associated images, to generate and output a collection of complementary objects.
The collection of complementary objects may then be used for subsequent steps in a user experience workflow, such as in providing recommendations or refining a search for an object that is associated with the collection of complementary objects. The collection of complementary objects, particularly when they become large, may be clustered based on an object type or object characteristic. Using the above example, a search for a white shirt may then generate clusters of complementary pants, belts, blazers, and shoes. The clusters may be used to provide the subsequent recommendations and/or to refine search results for items that may have a cluster identified. Moreover, for large clusters, such clusters may be divided into new, more refined, clusters or sub-clusters. For example, a large cluster of complementary belts may be sub-clustered into black belts and brown belts.
In one aspect, there is provided a computer-implemented method, comprising: determining one or more objects based on an input, using an input embedding associated with the input, and object embeddings associated with the one or more objects; determining a plurality of images containing the one or more objects; identifying complementary objects in the images; and providing an output collection of identified complementary objects.
In certain example embodiments, the one or more objects are determined based on a similarity to an object associated with the input.
In certain example embodiments, the method may further include clustering the output collection of identified complementary objects into object clusters based on similarities between the objects.
In certain example embodiments, the method may further include using the object clusters to provide at least one recommended object associated with the input.
In certain example embodiments, the object clusters may be used if satisfying a size threshold.
In certain example embodiments, the at least one recommended object may be provided in response to an input related to another object that is associated with at least one of the object clusters.
In certain example embodiments, the another object may be identified from a search query associated with the another object.
In certain example embodiments, the method may further include providing the at least one recommended object if a distance metric is satisfied.
In certain example embodiments, the distance metric may include a time since identifying the input or a distance between the input and the another object in a taxonomy of objects.
In certain example embodiments, the method may further include displaying a query comprising a characteristic of one of the object clusters; and displaying a reordered result list of objects based on a response to the query.
In certain example embodiments, identifying the complementary objects in the images may include instructing a large language model (LLM) to analyze the images for presence of at least one of the complementary objects.
In certain example embodiments, the LLM may include an artificial intelligence chatbot.
In certain example embodiments, identifying the complementary objects in the images may include applying an image processing technique to the images to detect objects in the images and identify detected objects using a machine learning model trained to detect objects in images.
In certain example embodiments, identifying the complementary objects in the images may include using metadata accompanying a corresponding one of the images, the metadata comprising at least one similarity to one of the complementary objects.
In certain example embodiments, at least one of the objects may include a plurality of associated images used to identify the complementary objects.
In certain example embodiments, an object associated with the input may be determined from a search query related to the object associated with the input.
In certain example embodiments, the input may be determined from a selection of an object associated with the input.
In certain example embodiments, the object associated with the input may be selected from a search query related to the object.
In certain example embodiments, the input and the output collection of identified complementary objects may correspond to products viewable via an e-commerce platform.
In certain example embodiments, associated images comprising the complementary objects may be accessed from a database available to the e-commerce platform.
In certain example embodiments, a candidate list of complementary objects may be used to identify the complementary objects in the images.
In certain example embodiments, the candidate list of complementary objects may be selected from one of a plurality of candidate lists based on an object type.
In another aspect, there is provided a system comprising a processor; an output device coupled to the processor; at least one input device coupled to the processor; and at least one memory. The at least one memory includes processor executable instructions that, when executed by the at least one processor, causes the system to: determine one or more objects based on an input, using an input embedding associated with the input, and object embeddings associated with the one or more objects; determine a plurality of images containing the one or more objects; identify complementary objects in the images; and provide an output collection of identified complementary objects.
In another aspect, there is provided a computer-readable medium comprising processor executable instructions that, when executed by a processor, cause the processor to: determine one or more objects based on an input, using an input embedding associated with the input, and object embeddings associated with the one or more objects; determine a plurality of images containing the one or more objects; identify complementary objects in the images; and provide an output collection of identified complementary objects.
To assist in understanding the present disclosure, some concepts relevant to neural networks and machine learning (ML) are first discussed.
Generally, a neural network comprises a number of computation units (sometimes referred to as “neurons”). Each neuron receives an input value and applies a function to the input to generate an output value. The function typically includes a parameter (also referred to as a “weight”) whose value is learned through the process of training. A plurality of neurons may be organized into a neural network layer (or simply “layer”) and there may be multiple such layers in a neural network. The output of one layer may be provided as input to a subsequent layer. Thus, input to a neural network may be processed through a succession of layers until an output of the neural network is generated by a final layer. This is a simplistic discussion of neural networks and there may be more complex neural network designs that include feedback connections, skip connections, and/or other such possible connections between neurons and/or layers, which need not be discussed in detail here.
A deep neural network (DNN) is a type of neural network having multiple layers and/or a large number of neurons. The term DNN may encompass any neural network having multiple layers, including convolutional neural networks (CNNs), recurrent neural networks (RNNs), and multilayer perceptrons (MLPs), among others.
DNNs are often used as ML-based models for modeling complex behaviors (e.g., human language, image recognition, object classification, etc.) in order to improve accuracy of outputs (e.g., more accurate predictions) such as, for example, as compared with models with fewer layers. In the present disclosure, the term “ML-based model” or more simply “ML model” may be understood to refer to a DNN. Training a ML model refers to a process of learning the values of the parameters (or weights) of the neurons in the layers such that the ML model is able to model the target behavior to a desired degree of accuracy. Training typically requires the use of a training dataset, which is a set of data that is relevant to the target behavior of the ML model. For example, to train a ML model that is intended to model human language (also referred to as a language model), the training dataset may be a collection of text documents, referred to as a text corpus (or simply referred to as a corpus). The corpus may represent a language domain (e.g., a single language), a subject domain (e.g., scientific papers), and/or may encompass another domain or domains, be they larger or smaller than a single language or subject domain. For example, a relatively large, multilingual and non-subject-specific corpus may be created by extracting text from online webpages and/or publicly available social media posts. In another example, to train a ML model that is intended to classify images, the training dataset may be a collection of images. Training data may be annotated with ground truth labels (e.g. each data entry in the training dataset may be paired with a label), or may be unlabeled.
Training a ML model generally involves inputting into an ML model (e.g. an untrained ML model) training data to be processed by the ML model, processing the training data using the ML model, collecting the output generated by the ML model (e.g. based on the inputted training data), and comparing the output to a desired set of target values. If the training data is labeled, the desired target values may be, e.g., the ground truth labels of the training data. If the training data is unlabeled, the desired target value may be a reconstructed (or otherwise processed) version of the corresponding ML model input (e.g., in the case of an autoencoder), or may be a measure of some target observable effect on the environment (e.g., in the case of a reinforcement learning agent). The parameters of the ML model are updated based on a difference between the generated output value and the desired target value. For example, if the value outputted by the ML model is excessively high, the parameters may be adjusted so as to lower the output value in future training iterations. An objective function is a way to quantitatively represent how close the output value is to the target value. An objective function represents a quantity (or one or more quantities) to be optimized (e.g., minimize a loss or maximize a reward) in order to bring the output value as close to the target value as possible. The goal of training the ML model typically is to minimize a loss function or maximize a reward function.
The training data may be a subset of a larger data set. For example, a data set may be split into three mutually exclusive subsets: a training set, a validation (or cross-validation) set, and a testing set. The three subsets of data may be used sequentially during ML model training. For example, the training set may be first used to train one or more ML models, each ML model, e.g., having a particular architecture, having a particular training procedure, being describable by a set of model hyperparameters, and/or otherwise being varied from the other of the one or more ML models. The validation (or cross-validation) set may then be used as input data into the trained ML models to, e.g., measure the performance of the trained ML models and/or compare performance between them. Where hyperparameters are used, a new set of hyperparameters may be determined based on the measured performance of one or more of the trained ML models, and the first step of training (i.e., with the training set) may begin again on a different ML model described by the new set of determined hyperparameters. In this way, these steps may be repeated to produce a more performant trained ML model. Once such a trained ML model is obtained (e.g., after the hyperparameters have been adjusted to achieve a desired level of performance), a third step of collecting the output generated by the trained ML model applied to the third subset (the testing set) may begin. The output generated from the testing set may be compared with the corresponding desired target values to give a final assessment of the trained ML model's accuracy. Other segmentations of the larger data set and/or schemes for using the segments for training one or more ML models are possible.
Backpropagation is an algorithm for training a ML model. Backpropagation is used to adjust (also referred to as update) the value of the parameters in the ML model, with the goal of optimizing the objective function. For example, a defined loss function is calculated by forward propagation of an input to obtain an output of the ML model and comparison of the output value with the target value. Backpropagation calculates a gradient of the loss function with respect to the parameters of the ML model, and a gradient algorithm (e.g., gradient descent) is used to update (i.e., “learn”) the parameters to reduce the loss function. Backpropagation is performed iteratively, so that the loss function is converged or minimized. Other techniques for learning the parameters of the ML model may be used. The process of updating (or learning) the parameters over many iterations is referred to as training. Training may be carried out iteratively until a convergence condition is met (e.g., a predefined maximum number of iterations has been performed, or the value outputted by the ML model is sufficiently converged with the desired target value), after which the ML model is considered to be sufficiently trained. The values of the learned parameters may then be fixed and the ML model may be deployed to generate output in real-world applications (also referred to as “inference”).
In some examples, a trained ML model may be fine-tuned, meaning that the values of the learned parameters may be adjusted slightly in order for the ML model to better model a specific task. Fine-tuning of a ML model typically involves further training the ML model on a number of data samples (which may be smaller in number/cardinality than those used to train the model initially) that closely target the specific task. For example, a ML model for generating natural language that has been trained generically on publically-available text corpuses may be, e.g., fine-tuned by further training using the complete works of Shakespeare as training data samples (e.g., where the intended use of the ML model is generating a scene of a play or other textual content in the style of Shakespeare).
The CNN 10 includes a plurality of layers that process the image 12 in order to generate an output, such as a predicted classification or predicted label for the image 12. For simplicity, only a few layers of the CNN 10 are illustrated including at least one convolutional layer 14. The convolutional layer 14 performs convolution processing, which may involve computing a dot product between the input to the convolutional layer 14 and a convolution kernel. A convolutional kernel is typically a 2D matrix of learned parameters that is applied to the input in order to extract image features. Different convolutional kernels may be applied to extract different image information, such as shape information, color information, etc.
The output of the convolution layer 14 is a set of feature maps 16 (sometimes referred to as activation maps). Each feature map 16 generally has smaller width and height than the image 12. The set of feature maps 16 encode image features that may be processed by subsequent layers of the CNN 10, depending on the design and intended task for the CNN 10. In this example, a fully connected layer 18 processes the set of feature maps 16 in order to perform a classification of the image, based on the features encoded in the set of feature maps 16. The fully connected layer 18 contains learned parameters that, when applied to the set of feature maps 16, outputs a set of probabilities representing the likelihood that the image 12 belongs to each of a defined set of possible classes. The class having the highest probability may then be outputted as the predicted classification for the image 12.
In general, a CNN may have different numbers and different types of layers, such as multiple convolution layers, max-pooling layers and/or a fully connected layer, among others. The parameters of the CNN may be learned through training, using data having ground truth labels specific to the desired task (e.g., class labels if the CNN is being trained for a classification task, pixel masks if the CNN is being trained for a segmentation task, text annotations if the CNN is being trained for a captioning task, etc.), as discussed above.
Some concepts in ML-based language models are now discussed. It may be noted that, while the term “language model” has been commonly used to refer to a ML-based language model, there could exist non-ML language models. In the present disclosure, the term “language model” may be used as shorthand for ML-based language model (i.e., a language model that is implemented using a neural network or other ML architecture), unless stated otherwise. For example, unless stated otherwise, “language model” encompasses LLMs.
A language model may use a neural network (typically a DNN) to perform natural language processing (NLP) tasks such as language translation, image captioning, grammatical error correction, and language generation, among others. A language model may be trained to model how words relate to each other in a textual sequence, based on probabilities. A language model may contain hundreds of thousands of learned parameters or in the case of a large language model (LLM) may contain millions or billions of learned parameters or more.
In recent years, there has been interest in a type of neural network architecture, referred to as a transformer, for use as language models. For example, the Bidirectional Encoder Representations from Transformers (BERT) model, the Transformer-XL model and the Generative Pre-trained Transformer (GPT) models are types of transformers. A transformer is a type of neural network architecture that uses self-attention mechanisms in order to generate predicted output based on input data that has some sequential meaning (i.e., the order of the input data is meaningful, which is the case for most text input). Although transformer-based language models are described herein, it should be understood that the present disclosure may be applicable to any ML-based language model, including language models based on other neural network architectures such as recurrent neural network (RNN)-based language models.
The transformer 50 may be trained on a text corpus that is labelled (e.g., annotated to indicate verbs, nouns, etc.) or unlabelled. LLMs may be trained on a large unlabelled corpus. Some LLMs may be trained on a large multi-language, multi-domain corpus, to enable the model to be versatile at a variety of language-based tasks such as generative tasks (e.g., generating human-like natural language responses to natural language input).
An example of how the transformer 50 may process textual input data is now described. Input to a language model (whether transformer-based or otherwise) typically is in the form of natural language as may be parsed into tokens. It should be appreciated that the term “token” in the context of language models and NLP has a different meaning from the use of the same term in other contexts such as data security. Tokenization, in the context of language models and NLP, refers to the process of parsing textual input (e.g., a character, a word, a phrase, a sentence, a paragraph, etc.) into a sequence of shorter segments that are converted to numerical representations referred to as tokens (or “compute tokens”). Typically, a token may be an integer that corresponds to the index of a text segment (e.g., a word) in a vocabulary dataset. Often, the vocabulary dataset is arranged by frequency of use. Commonly occurring text, such as punctuation, may have a lower vocabulary index in the dataset and thus be represented by a token having a smaller integer value than less commonly occurring text. Tokens frequently correspond to words, with or without whitespace appended. In some examples, a token may correspond to a portion of a word. For example, the word “lower” may be represented by a token for [low] and a second token for [er]. In another example, the text sequence “Come here, look!” may be parsed into the segments [Come], [here], [,], [look] and [!], each of which may be represented by a respective numerical token. In addition to tokens that are parsed from the textual sequence (e.g., tokens that correspond to words and punctuation), there may also be special tokens to encode non-textual information. For example, a [CLASS] token may be a special token that corresponds to a classification of the textual sequence (e.g., may classify the textual sequence as a poem, a list, a paragraph, etc.), a [EOT] token may be another special token that indicates the end of the textual sequence, other tokens may provide formatting information, etc.
In
The generated embeddings 60 are input into the encoder 52. The encoder 52 serves to encode the embeddings 60 into feature vectors 62 that represent the latent features of the embeddings 60. The encoder 52 may encode positional information (i.e., information about the sequence of the input) in the feature vectors 62. The feature vectors 62 may have very high dimensionality (e.g., on the order of thousands or tens of thousands), with each element in a feature vector 62 corresponding to a respective feature. The numerical weight of each element in a feature vector 62 represents the importance of the corresponding feature. The space of all possible feature vectors 62 that can be generated by the encoder 52 may be referred to as the latent space or feature space.
Conceptually, the decoder 54 is designed to map the features represented by the feature vectors 62 into meaningful output, which may depend on the task that was assigned to the transformer 50. For example, if the transformer 50 is used for a translation task, the decoder 54 may map the feature vectors 62 into text output in a target language different from the language of the original tokens 56. Generally, in a generative language model, the decoder 54 serves to decode the feature vectors 62 into a sequence of tokens. The decoder 54 may generate output tokens 64 one by one. Each output token 64 may be fed back as input to the decoder 54 in order to generate the next output token 64. By feeding back the generated output and applying self-attention, the decoder 54 is able to generate a sequence of output tokens 64 that has sequential meaning (e.g., the resulting output text sequence is understandable as a sentence and obeys grammatical rules). The decoder 54 may generate output tokens 64 until a special [EOT] token (indicating the end of the text) is generated. The resulting sequence of output tokens 64 may then be converted to a text sequence in post-processing. For example, each output token 64 may be an integer number that corresponds to a vocabulary index. By looking up the text segment using the vocabulary index, the text segment corresponding to each output token 64 can be retrieved, the text segments can be concatenated together and the final output text sequence (in this example, “Viens ici, regarde!”) can be obtained.
Although a general transformer architecture for a language model and its theory of operation have been described above, this is not intended to be limiting. Existing language models include language models that are based only on the encoder of the transformer or only on the decoder of the transformer. An encoder-only language model encodes the input text sequence into feature vectors that can then be further processed by a task-specific layer (e.g., a classification layer). BERT is an example of a language model that may be considered to be an encoder-only language model. A decoder-only language model accepts embeddings as input and may use auto-regression to generate an output text sequence. Transformer-XL and GPT-type models may be language models that are considered to be decoder-only language models.
Because GPT-type language models tend to have a large number of parameters, these language models may be considered LLMs. An example GPT-type LLM is GPT-3. GPT-3 is a type of GPT language model that has been trained (in an unsupervised manner) on a large corpus derived from documents available to the public online. GPT-3 has a very large number of learned parameters (on the order of hundreds of billions), is able to accept a large number of tokens as input (e.g., up to 2048 input tokens), and is able to generate a large number of tokens as output (e.g., up to 2048 tokens). GPT-3 has been trained as a generative model, meaning that it can process input text sequences to predictively generate a meaningful output text sequence. ChatGPT is built on top of a GPT-type LLM, and has been fine-tuned with training datasets based on text-based chats (e.g., chatbot conversations). ChatGPT is designed for processing natural language, receiving chat-like inputs and generating chat-like outputs.
A computing system may access a remote language model (e.g., a cloud-based language model), such as ChatGPT or GPT-3, via a software interface (e.g., an application programming interface (API)). Additionally or alternatively, such a remote language model may be accessed via a network such as, for example, the Internet. In some implementations such as, for example, potentially in the case of a cloud-based language model, a remote language model may be hosted by a computer system as may include a plurality of cooperating (e.g., cooperating via a network) computer systems such as may be in, for example, a distributed arrangement. Notably, a remote language model may employ a plurality of processors (e.g., hardware processors such as, for example, processors of cooperating computer systems). Indeed, processing of inputs by an LLM may be computationally expensive/may involve a large number of operations (e.g., many instructions may be executed/large data structures may be accessed from memory) and providing output in a required timeframe (e.g., real-time or near real-time) may require the use of a plurality of processors/cooperating computing devices as discussed above.
Inputs to an LLM may be referred to as a prompt, which is a natural language input that includes instructions to the LLM to generate a desired output. A computing system may generate a prompt that is provided as input to the LLM via its API. As described above, the prompt may optionally be processed or pre-processed into a token sequence prior to being provided as input to the LLM via its API. A prompt can include one or more examples of the desired output, which provides the LLM with additional information to enable the LLM to better generate output according to the desired output. Additionally or alternatively, the examples included in a prompt may provide inputs (e.g., example inputs) corresponding to/as may be expected to result in the desired outputs provided. A one-shot prompt refers to a prompt that includes one example, and a few-shot prompt refers to a prompt that includes multiple examples. A prompt that includes no examples may be referred to as a zero-shot prompt.
The example computing system 400 includes at least one processing unit, such as a processor 402, and at least one physical memory 404. The processor 402 may be, for example, a central processing unit, a microprocessor, a digital signal processor, an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA), a dedicated logic circuitry, a dedicated artificial intelligence processor unit, a graphics processing unit (GPU), a tensor processing unit (TPU), a neural processing unit (NPU), a hardware accelerator, or combinations thereof. The memory 404 may include a volatile or non-volatile memory (e.g., a flash memory, a random access memory (RAM), and/or a read-only memory (ROM)). The memory 404 may store instructions for execution by the processor 402, to the computing system 400 to carry out examples of the methods, functionalities, systems and modules disclosed herein.
The computing system 400 may also include at least one network interface 406 for wired and/or wireless communications with an external system and/or network (e.g., an intranet, the Internet, a P2P network, a WAN and/or a LAN). A network interface may enable the computing system 400 to carry out communications (e.g., wireless communications) with systems external to the computing system 400, such as a language model residing on a remote system.
The computing system 400 may optionally include at least one input/output (I/O) interface 408, which may interface with optional input device(s) 410 and/or optional output device(s) 412. Input device(s) 410 may include, for example, buttons, a microphone, a touchscreen, a keyboard, etc. Output device(s) 412 may include, for example, a display, a speaker, etc. In this example, optional input device(s) 410 and optional output device(s) 412 are shown external to the computing system 400. In other examples, one or more of the input device(s) 410 and/or output device(s) 412 may be an internal component of the computing system 400.
A computing system, such as the computing system 400 of
To determine and provide collections of complementary products to those determined from an input, the proposed solution uses products that are determined to be similar to the input-related product as a proxy, such that images of the similar products are used to find such complementary products and may result in a collection of complementary products based on the images. Matching the input to the similar products may use embeddings of the input and of the object(s) in an appropriate embedding space. The collection of complementary products may then be clustered, and the clusters used in various user experience flows, such as to pre-emptively provide product recommendations or to refine subsequent search results related to the input product, e.g., by linking and refining subsequent search queries when the size of the matching cluster satisfies a size threshold.
For example, the computing environment 200 shown in
The application 202 and search function 210 are coupled to a display 204 to render and present/display user interface (UI) elements, UI components, and UI controls utilized by a UI rendered by application 202 (e.g., via or including the search function 210), on the display 204. While examples referred to herein may refer to a single display 204 for ease of illustration, the principles discussed herein can also be applied to multiple displays 204, e.g., to view portions of UIs rendered by or with the application 202 on separate side-by-side screens. That is, any reference to a display 204 can include any one or more displays 204 or screens providing similar visual functions. The application 202 receives one or more inputs from one or more input devices 410, which can include or incorporate inputs made via the display 204 as illustrated in
The application 202 or a computing device 240 within the computing environment 200 can include or have access to application data 248 (see
The application 202 may include, or as shown in the example configuration, have access to, a complementary object engine 300. The complementary object engine 300 may be used to process an input associated with an object (e.g., a search query for that object) to determine one or more complementary objects to that object using images that contain similar objects as a proxy.
The complementary object engine 300 in this example is coupled to a communication network 216, e.g., to communicate with a computing device 240 or computing system 400 in the computing environment 200. Such communication network(s) 216 may include a telephone network, cellular, and/or data communication network to connect different types of client- and/or server-type devices. For example, the communication network 216 may include a private or public switched telephone network (PSTN), mobile network (e.g., code division multiple access (CDMA) network, global system for mobile communications (GSM) network, and/or any 3G, 4G, or 5G wireless carrier network, etc.), WiFi or other similar wireless network, and a private and/or public wide area network (e.g., the Internet).
The complementary object engine 300 may receive one or more inputs 218 via the communication network 216 and generate and provide an output such as a complementary object collection 234, e.g., to/from the application 202. The one or more inputs 218 are provided to a similarity engine 220. The similarity engine 220 determines an object associated with the one or more inputs 218, e.g., a product identified in a search query entered using the search function 210. The similarity engine 220 may use the identified object that is associated with the one or more inputs 218 to determine a set of one or more related objects from an object database 222. For example, if the object is a product that is the subject of a search query, the similarity engine 220 may use text from the search and/or an embedding for/with data associated with the input(s) 218 to determine related objects, such as other products that possess similar features or characteristics. The embedding may include feature representations for product identifiers, images associated with the product, or any other available embedding used with the data associated with the input(s) 218. For example, a user may perform a text-based search for a “shirt”, which in itself provides little insight. However, the user may then select a particular shirt from a list of search results, with the particular shirt being white in color and which includes buttons, a collar, and has long sleeves. These features may be determinable from a product description or image embeddings that is/are associated with the selected shirt.
That is, the input 218 may include one or more embeddings that may be used by the similarity engine 220 to find similar white shirts in the object database 222 (e.g., in a product catalogue). For example, the similarity engine 220 may use the embeddings related to color and at least one feature such as buttons to find similar products. By identifying the related object(s), the similarity engine 220 may locate one or more images 226 of the similar objects, e.g., from an object images database 224 that is associated with the object database 222. For example, a product catalogue of items may have an associated database of object images 224 that is used to provide product details in a product page UI. The object database 222 may therefore also include products that have embeddings that may be used by the similarity engine 220 to match, associate, correlate or otherwise find similarity with the embeddings associated with the input 218. Once matched by the similarity engine 220, the relate object images database 224 may be accessed to obtain a set of one or more images 226 for each similar object (e.g., for each similar white shirt in this example).
The images 226 may then be used as a proxy to find complementary objects to the object associated with the input(s) 218. In the configuration shown in
For example, a LLM 230 may be used by the image analyzer 228 to determine from the contents of the images 226 whether any complementary products are present. The LLM 230 may be queried by the image analyzer 228 to determine from the embeddings for the images 226 whether certain other objects are present. This may include using a product taxonomy to structure a set of queries to be fed to the LLM 230 based on the product identified by the similarity engine 220. In other words, a candidate list of complementary objects may be used to identify a more specific list of complementary objects determined from the image(s) 226.
In the above example, since the similarity engine 220 identifies one or more similar white shirts, the image analyzer 228 may use a clothing taxonomy to query the LLM 230 for related products such as pants, belts, shoes, etc. This may include a general query such as “identify other clothing in the image” or more specific queries such as “does the image include a pair of pants”. Other lines of inquiry may also be generated to target the complementary objects. For instance, the image analyzer 228 may use the LLM 230 to determine if the image(s) 226 contain a human. If so, the LLM 230 may be used to determine what else the human is wearing. If the image 226 does not contain a human, the LLM 230 may be queried to determine if anything else is shown in the image 226. In this way, the LLM 230 may be leveraged to locate other relevant items in the image 226 as well as to exclude irrelevant objects. For example, the LLM 230 may be queried to determine what objects are shown in the scene containing the object associated with the input 218. Background objects may be used to exclude objects that may be relevant to the scene but not necessarily complementary to the object associated with the input 218. In the above example, the LLM 230 may be used to search image embeddings to determine that several objects in the scene are part of a background shelf while a human standing in front of the shelf is wearing and/or holding certain items that may be more relevant.
The image analyzer 228 may, additionally or alternatively, use an image processor 232 to identify complementary objects in the images 226. For example, the image processor 232 may use an object recognition process to detect objects in the images 226 and use bounding boxes to target those objects for further analysis. The targeted objects may have embeddings associated with them, e.g., via image embeddings, tags or other metadata. The image processor 232 may also use a machine learning process to identify the objects using a model that is trained to identify objects in images. Such a machine learning model may be a generic model or may be specific to a type or genre of object/product. For example, if the similarity engine 220 is analyzing a set of images 226 for similar white shirts, the image processor 232 may use a clothing related model to identify other clothing pieces or accessories in the images 226 that may then be inferred to be complementary on the basis that they are shown with the white shirt in the images 226.
The image analyzer 228 may also use any available structured or unstructured data in or with the images 226 using any suitable processing technique. For example, the images 226 obtained from the object images database 224 may have already been tagged with products, features, and characteristics for some other purpose and can be leveraged with minimal processing to identify the complementary object collection 234.
The objects identified by the image analyzer 228 may be linked to the object database 222 and therefore themselves have related object images 224 which may be accessed directly, or may be used to find similar objects, e.g., using the similarity engine 220 as shown in
The complementary object collection 234 may include one or more complementary objects and the collection 234 may be clustered when the collection 234 becomes large. For example, a set of similar objects may each produce multiple images 226, each of which includes multiple potential complementary objects. When a number of similar objects are analyzed, the collection 234 may grow relatively large and require at least some filtering, classification, or organization, e.g., via clustering. Clustering allows the complementary object collection 234 to categorize the complementary objects. The size of each cluster may be used as a metric to determine the objects relative complementarity or as a threshold to determine whether or not to recommend that object. The complementary object collection 234 may therefore be of varying size and complexity with clustering or other organizational techniques such as classification or categorization applied to provide the complementary object cache 212 with manageable data for use in a subsequent operation such as in organizing subsequent search results, providing product recommendations, etc.
While the complementary object engine 300 is shown as being a separate tool accessed by the application 202 over communication network 216 this is only one possible configuration. That is, the complementary object engine 300 may, additionally or alternatively, be integral to the application 202 or may reside on the same computing device 240 or computing system 400 as the application 202 in other example configurations. Moreover, the communication network 216 shown in
While not delineated in
Referring now to
At block 262, the similarity engine 220 may determine images containing the objects, for example, the object images 224 that are associated with the object database 222. It can be appreciated that the object images 224 may be obtained from any available and suitable source, including internal or external sources, which may be private or public databases and/or other online content (e.g., product webpages).
At block 264, the image analyzer 228 may be used to identify complementary objects in the images 226 that were determined in block 262. For example, as discussed above, the image analyzer 228 may use a LLM 230 or image processor 232 or any available data processing tool that is capable of analyzing the contents or related data for the images 226, to identify such complementary objects. Depending on the number of similar objects that are analyzed and the number of images 226 that are processed, the complementary objects that are identified may be organized, grouped or clustered as provided by way of example below. Whether one or many complementary objects are identified, at block 266, a collection 234 of the identified complementary objects may be output, e.g., by returning the complementary object collection 234 to the application 202 as illustrated in
Referring now to
For each similar white shirt (shown as Object 1, Object 2, . . . . Object N in
For example, an object detection model may be trained to identify apparel in an image. The output of the model may include a bounding box and label for each identified item of clothing. The analyses can be conducted by using a rule set, taxonomy or list of questions (e.g., to query a LLM 230) that is associated with the white shirt. For example, by identifying the input object as a white shirt, a clothing taxonomy can be used to determine which objects may be complementary in any given associated image 226. In this way, irrelevant objects can be ignored. For example, an associated image 226 of a human wearing a white shirt and complementary clothing may be relevant, while background objects such as a bench, building, landscaping, vehicles, etc., may not. That is, the associated images 226 may be processed using a candidate list of complementary objects, which then permits an analysis technique to generate a refined list of complementary products. This permits a more personalized collection to be created by not only finding objects in the same image, but finding objects that are more likely to be complementary and thus more relevant to the user.
The output collection of complementary products (i.e., the complementary object collection 234) may then be used in a subsequent step in a design, purchase or other workflow as shown in
As indicated above, to organize the output collection 234, particularly with a relatively large number of results, clustering can be performed at block 270 as described below making reference to
While the collection 234 is held in the cache 212, a subsequent option 272 may utilize the collection 234. For example, referring back to the white shirt example, the buyer, after adding the white shirt to the cart, may search for grey pants. By having used the associated images 226 as a proxy to find potential complementary products, a trouser cluster (or grey pant sub-cluster) may be referenced to prompt the buyer with a targeted and personalized question, such as “Are you looking for pants to match your white shirt?”. Other features may be identified to provide additional queries such as a color or style or any other attribute of the product.
Referring now to
At block 282, the application 202 (or complementary object engine 300) may determine a matching cluster and prompt the user 208 regarding the second product. For example, if the buyer searches for “grey pants”, the application 202 may use the cache 212 to identify a matching cluster for grey pants and may prompt the buyer with the question indicated above, namely: “Are you looking for pants to match your white shirt?”. The user may reply to this prompt by indicating “yes”, which may have the application 202 (or complementary object engine 300) at block 284 determine one or more features associated with the second product (i.e., the grey pants in this example) that is/are complementary to the first product (i.e., the white shirt in this example). The feature(s) may be used at block 286 to apply an operation to displayed results based on the cluster and feature similarities. For example, using the matched grey pants cluster confirmed by the user and, optionally, using feature(s) associated with the grey pants that is/are complementary to the white shirt (e.g., dress pants versus track pants, high waisted versus low waisted, etc.), a search result list for “grey pants” can be reranked based on their similarity to the matching cluster and the common feature(s).
The prompts and questions illustrated above may be refined or controlled based on certain conditions. For example, a cluster may need to meet a certain size threshold to be considered complementary. Similarly, the type of clusters, number of clusters, and cluster sizes may be used to intelligently reorder search results in addition to, or instead of, posing questions to the user. For example, in the above example, by searching for “grey pants”, the search results may be reordered to push the complementary grey pants (or other complementary pants of another color) to the top of the search results. This process may be repeated for multiple subsequent searches. For example, a search for “brown shoes” may utilize the output collection 234 (clustered or not) or may refine the collection 234 based on the addition of particular grey pants that resulted from the above search. That is, the output collection of complementary objects and any clustering or organization of such a collection can be refined or replaced with a fresh output collection 234 as a session progresses and further objects may be selected and identified as new input objects. Similarly, sessions that include disparate searches for products added to the same cart may involve repeating the process and generating multiple output collections 234 that are based on different candidate lists of objects.
Referring now to
Referring to
Referring now to
An Example e-Commerce Platform
Although integration with a commerce platform is not required, in some embodiments, the methods disclosed herein may be performed on or in association with a commerce platform such as an e-commerce platform. Therefore, an example of a commerce platform will be described.
While the disclosure throughout contemplates that a ‘merchant’ and a ‘customer’ may be more than individuals, for simplicity the description herein may generally refer to merchants and customers as such. All references to merchants and customers throughout this disclosure should also be understood to be references to groups of individuals, companies, corporations, computing entities, and the like, and may represent for-profit or not-for-profit exchange of products. Further, while the disclosure throughout refers to ‘merchants’ and ‘customers’, and describes their roles as such, the e-commerce platform 100 should be understood to more generally support users in an e-commerce environment, and all references to merchants and customers throughout this disclosure should also be understood to be references to users, such as where a user is a merchant-user (e.g., a seller, retailer, wholesaler, or provider of products), a customer-user (e.g., a buyer, purchase agent, consumer, or user of products), a prospective user (e.g., a user browsing and not yet committed to a purchase, a user evaluating the e-commerce platform 100 for potential use in marketing and selling products, and the like), a service provider user (e.g., a shipping provider 112, a financial provider, and the like), a company or corporate user (e.g., a company representative for purchase, sales, or use of products; an enterprise user; a customer relations or customer management agent, and the like), an information technology user, a computing entity user (e.g., a computing bot for purchase, sales, or use of products), and the like. Furthermore, it may be recognized that while a given user may act in a given role (e.g., as a merchant) and their associated device may be referred to accordingly (e.g., as a merchant device) in one context, that same individual may act in a different role in another context (e.g., as a customer) and that same or another associated device may be referred to accordingly (e.g., as a customer device). For example, an individual may be a merchant for one type of product (e.g., shoes), and a customer/consumer of other types of products (e.g., groceries). In another example, an individual may be both a consumer and a merchant of the same type of product. In a particular example, a merchant that trades in a particular category of goods may act as a customer for that same category of goods when they order from a wholesaler (the wholesaler acting as merchant).
The e-commerce platform 100 provides merchants with online services/facilities to manage their business. The facilities described herein are shown implemented as part of the platform 100 but could also be configured separately from the platform 100, in whole or in part, as stand-alone services. Furthermore, such facilities may, in some embodiments, may, additionally or alternatively, be provided by one or more providers/entities.
In the example of
The online store 138 may represent a multi-tenant facility comprising a plurality of virtual storefronts. In embodiments, merchants may configure and/or manage one or more storefronts in the online store 138, such as, for example, through a merchant device 102 (e.g., computer, laptop computer, mobile computing device, and the like), and offer products to customers through a number of different channels 110A-B (e.g., an online store 138; an application 142A-B; a physical storefront through a POS device 152; an electronic marketplace, such, for example, through an electronic buy button integrated into a website or social media channel such as on a social network, social media page, social media messaging system; and/or the like). A merchant may sell across channels 110A-B and then manage their sales through the e-commerce platform 100, where channels 110A may be provided as a facility or service internal or external to the e-commerce platform 100. A merchant may, additionally or alternatively, sell in their physical retail store, at pop ups, through wholesale, over the phone, and the like, and then manage their sales through the e-commerce platform 100. A merchant may employ all or any combination of these operational modalities. Notably, it may be that by employing a variety of and/or a particular combination of modalities, a merchant may improve the probability and/or volume of sales. Throughout this disclosure the terms online store 138 and storefront may be used synonymously to refer to a merchant's online e-commerce service offering through the e-commerce platform 100, where an online store 138 may refer either to a collection of storefronts supported by the e-commerce platform 100 (e.g., for one or a plurality of merchants) or to an individual merchant's storefront (e.g., a merchant's online store).
In some embodiments, a customer may interact with the platform 100 through a customer device 150 (e.g., computer, laptop computer, mobile computing device, or the like), a POS device 152 (e.g., retail device, kiosk, automated (self-service) checkout system, or the like), and/or any other commerce interface device known in the art. The e-commerce platform 100 may enable merchants to reach customers through the online store 138, through applications 142A-B, through POS devices 152 in physical locations (e.g., a merchant's storefront or elsewhere), to communicate with customers via electronic communication facility 129, and/or the like so as to provide a system for reaching customers and facilitating merchant services for the real or virtual pathways available for reaching and interacting with customers.
In some embodiments, and as described further herein, the e-commerce platform 100 may be implemented through a processing facility. Such a processing facility may include a processor and a memory. The processor may be a hardware processor. The memory may be and/or may include a non-transitory computer-readable medium. The memory may be and/or may include random access memory (RAM) and/or persisted storage (e.g., magnetic storage). The processing facility may store a set of instructions (e.g., in the memory) that, when executed, cause the e-commerce platform 100 to perform the e-commerce and support functions as described herein. The processing facility may be or may be a part of one or more of a server, client, network infrastructure, mobile computing platform, cloud computing platform, stationary computing platform, and/or some other computing platform, and may provide electronic connectivity and communications between and amongst the components of the e-commerce platform 100, merchant devices 102, payment gateways 106, applications 142A-B, channels 110A-B, shipping providers 112, customer devices 150, point of sale devices 152, etc. In some implementations, the processing facility may be or may include one or more such computing devices acting in concert. For example, it may be that a plurality of co-operating computing devices serves as/to provide the processing facility. The e-commerce platform 100 may be implemented as or using one or more of a cloud computing service, software as a service (SaaS), infrastructure as a service (IaaS), platform as a service (PaaS), desktop as a service (DaaS), managed software as a service (MSaaS), mobile backend as a service (MBaaS), information technology management as a service (ITMaaS), and/or the like. For example, it may be that the underlying software implementing the facilities described herein (e.g., the online store 138) is provided as a service, and is centrally hosted (e.g., and then accessed by users via a web browser or other application, and/or through customer devices 150, POS devices 152, and/or the like). In some embodiments, elements of the e-commerce platform 100 may be implemented to operate and/or integrate with various other platforms and operating systems.
In some embodiments, the facilities of the e-commerce platform 100 (e.g., the online store 138) may serve content to a customer device 150 (using data 134) such as, for example, through a network connected to the e-commerce platform 100. For example, the online store 138 may serve or send content in response to requests for data 134 from the customer device 150, where a browser (or other application) connects to the online store 138 through a network using a network communication protocol (e.g., an internet protocol). The content may be written in machine readable language and may include Hypertext Markup Language (HTML), template language, JavaScript, and the like, and/or any combination thereof.
In some embodiments, online store 138 may be or may include service instances that serve content to customer devices and allow customers to browse and purchase the various products available (e.g., add them to a cart, purchase through a buy-button, and the like). Merchants may also customize the look and feel of their website through a theme system, such as, for example, a theme system where merchants can select and change the look and feel of their online store 138 by changing their theme while having the same underlying product and business data shown within the online store's product information. It may be that themes can be further customized through a theme editor, a design interface that enables users to customize their website's design with flexibility. Additionally or alternatively, it may be that themes can, additionally or alternatively, be customized using theme-specific settings such as, for example, settings as may change aspects of a given theme, such as, for example, specific colors, fonts, and pre-built layout schemes. In some implementations, the online store may implement a content management system for website content. Merchants may employ such a content management system in authoring blog posts or static pages and publish them to their online store 138, such as through blogs, articles, landing pages, and the like, as well as configure navigation menus. Merchants may upload images (e.g., for products), video, content, data, and the like to the e-commerce platform 100, such as for storage by the system (e.g., as data 134). In some embodiments, the e-commerce platform 100 may provide functions for manipulating such images and content such as, for example, functions for resizing images, associating an image with a product, adding and associating text with an image, adding an image for a new product variant, protecting images, and the like.
As described herein, the e-commerce platform 100 may provide merchants with sales and marketing services for products through a number of different channels 110A-B, including, for example, the online store 138, applications 142A-B, as well as through physical POS devices 152 as described herein. The e-commerce platform 100 may, additionally or alternatively, include business support services 116, an administrator 114, a warehouse management system, and the like associated with running an on-line business, such as, for example, one or more of providing a domain registration service 118 associated with their online store, payment services 120 for facilitating transactions with a customer, shipping services 122 for providing customer shipping options for purchased products, fulfillment services for managing inventory, risk and insurance services 124 associated with product protection and liability, merchant billing, and the like. Services 116 may be provided via the e-commerce platform 100 or in association with external facilities, such as through a payment gateway 106 for payment processing, shipping providers 112 for expediting the shipment of products, and the like.
In some embodiments, the e-commerce platform 100 may be configured with shipping services 122 (e.g., through an e-commerce platform shipping facility or through a third-party shipping carrier), to provide various shipping-related information to merchants and/or their customers such as, for example, shipping label or rate information, real-time delivery updates, tracking, and/or the like.
More detailed information about commerce and visitors to a merchant's online store 138 may be viewed through reports or metrics. Reports may include, for example, acquisition reports, behavior reports, customer reports, finance reports, marketing reports, sales reports, product reports, and custom reports. The merchant may be able to view sales data for different channels 110A-B from different periods of time (e.g., days, weeks, months, and the like), such as by using drop-down menus. An overview dashboard may also be provided for a merchant who wants a more detailed view of the store's sales and engagement data. An activity feed in the home metrics section may be provided to illustrate an overview of the activity on the merchant's account. For example, by clicking on a ‘view all recent activity’ dashboard button, the merchant may be able to see a longer feed of recent activity on their account. A home page may show notifications about the merchant's online store 138, such as based on account status, growth, recent customer activity, order updates, and the like. Notifications may be provided to assist a merchant with navigating through workflows configured for the online store 138, such as, for example, a payment workflow, an order fulfillment workflow, an order archiving workflow, a return workflow, and the like.
The e-commerce platform 100 may provide for a communications facility 129 and associated merchant interface for providing electronic communications and marketing, such as utilizing an electronic messaging facility for collecting and analyzing communication interactions between merchants, customers, merchant devices 102, customer devices 150, POS devices 152, and the like, to aggregate and analyze the communications, such as for increasing sale conversions, and the like. For instance, a customer may have a question related to a product, which may produce a dialog between the customer and the merchant (or an automated processor-based agent/chatbot representing the merchant), where the communications facility 129 is configured to provide automated responses to customer requests and/or provide recommendations to the merchant on how to respond such as, for example, to improve the probability of a sale.
The e-commerce platform 100 may provide a financial facility 120 for secure financial transactions with customers, such as through a secure card server environment. The e-commerce platform 100 may store credit card information, such as in payment card industry data (PCI) environments (e.g., a card server), to reconcile financials, bill merchants, perform automated clearing house (ACH) transfers between the e-commerce platform 100 and a merchant's bank account, and the like. The financial facility 120 may also provide merchants and buyers with financial support, such as through the lending of capital (e.g., lending funds, cash advances, and the like) and provision of insurance. In some embodiments, online store 138 may support a number of independently administered storefronts and process a large volume of transactional data on a daily basis for a variety of products and services. Transactional data may include any customer information indicative of a customer, a customer account or transactions carried out by a customer such as, for example, contact information, billing information, shipping information, returns/refund information, discount/offer information, payment information, or online store events or information such as page views, product search information (search keywords, click-through events), product reviews, abandoned carts, and/or other transactional information associated with business through the e-commerce platform 100. In some embodiments, the e-commerce platform 100 may store this data in a data facility 134. Referring again to
Implementing functions as applications 142A-B may enable the commerce management engine 136 to remain responsive and reduce or avoid service degradation or more serious infrastructure failures, and the like.
Although isolating online store data can be important to maintaining data privacy between online stores 138 and merchants, there may be reasons for collecting and using cross-store data, such as for example, with an order risk assessment system or a platform payment facility, both of which require information from multiple online stores 138 to perform well. In some embodiments, it may be preferable to move these components out of the commerce management engine 136 and into their own infrastructure within the e-commerce platform 100.
Platform payment facility 120 is an example of a component that utilizes data from the commerce management engine 136 but is implemented as a separate component or service. The platform payment facility 120 may allow customers interacting with online stores 138 to have their payment information stored safely by the commerce management engine 136 such that they only have to enter it once. When a customer visits a different online store 138, even if they have never been there before, the platform payment facility 120 may recall their information to enable a more rapid and/or potentially less-error prone (e.g., through avoidance of possible mis-keying of their information if they needed to instead re-enter it) checkout. This may provide a cross-platform network effect, where the e-commerce platform 100 becomes more useful to its merchants and buyers as more merchants and buyers join, such as because there are more customers who checkout more often because of the ease of use with respect to customer purchases. To maximize the effect of this network, payment information for a given customer may be retrievable and made available globally across multiple online stores 138.
For functions that are not included within the commerce management engine 136, applications 142A-B provide a way to add features to the e-commerce platform 100 or individual online stores 138. For example, applications 142A-B may be able to access and modify data on a merchant's online store 138, perform tasks through the administrator 114, implement new flows for a merchant through a user interface (e.g., that is surfaced through extensions/API), and the like. Merchants may be enabled to discover and install applications 142A-B through application search, recommendations, and support 128. In some embodiments, the commerce management engine 136, applications 142A-B, and the administrator 114 may be developed to work together. For instance, application extension points may be built inside the commerce management engine 136, accessed by applications 142A and 142B through the interfaces 140B and 140A to deliver additional functionality, and surfaced to the merchant in the user interface of the administrator 114.
In some embodiments, applications 142A-B may deliver functionality to a merchant through the interface 140A-B, such as where an application 142A-B is able to surface transaction data to a merchant (e.g., App: “Engine, surface my app data in the Mobile App or administrator 114”), and/or where the commerce management engine 136 is able to ask the application to perform work on demand (Engine: “App, give me a local tax calculation for this checkout”).
Applications 142A-B may be connected to the commerce management engine 136 through an interface 140A-B (e.g., through REST (REpresentational State Transfer) and/or GraphQL APIs) to expose the functionality and/or data available through and within the commerce management engine 136 to the functionality of applications. For instance, the e-commerce platform 100 may provide API interfaces 140A-B to applications 142A-B which may connect to products and services external to the platform 100. The flexibility offered through use of applications and APIs (e.g., as offered for application development) enable the e-commerce platform 100 to better accommodate new and unique needs of merchants or to address specific use cases without requiring constant change to the commerce management engine 136. For instance, shipping services 122 may be integrated with the commerce management engine 136 through a shipping or carrier service API, thus enabling the e-commerce platform 100 to provide shipping service functionality without directly impacting code running in the commerce management engine 136.
Depending on the implementation, applications 142A-B may utilize APIs to pull data on demand (e.g., customer creation events, product change events, or order cancelation events, etc.) or have the data pushed when updates occur. A subscription model may be used to provide applications 142A-B with events as they occur or to provide updates with respect to a changed state of the commerce management engine 136. In some embodiments, when a change related to an update event subscription occurs, the commerce management engine 136 may post a request, such as to a predefined callback URL. The body of this request may contain a new state of the object and a description of the action or event. Update event subscriptions may be created manually, in the administrator facility 114, or automatically (e.g., via the API 140A-B). In some embodiments, update events may be queued and processed asynchronously from a state change that triggered them, which may produce an update event notification that is not distributed in real-time or near-real time.
In some embodiments, the e-commerce platform 100 may provide one or more of application search, recommendation and support 128. Application search, recommendation and support 128 may include developer products and tools to aid in the development of applications, an application dashboard (e.g., to provide developers with a development interface, to administrators for management of applications, to merchants for customization of applications, and the like), facilities for installing and providing permissions with respect to providing access to an application 142A-B (e.g., for public access, such as where criteria must be met before being installed, or for private use by a merchant), application searching to make it easy for a merchant to search for applications 142A-B that satisfy a need for their online store 138, application recommendations to provide merchants with suggestions on how they can improve the user experience through their online store 138, and the like. In some embodiments, applications 142A-B may be assigned an application identifier (ID), such as for linking to an application (e.g., through an API), searching for an application, making application recommendations, and the like.
Applications 142A-B may be grouped roughly into three categories: customer-facing applications, merchant-facing applications, integration applications, and the like. Customer-facing applications 142A-B may include an online store 138 or channels 110A-B that are places where merchants can list products and have them purchased (e.g., the online store, applications for flash sales (e.g., merchant products or from opportunistic sales opportunities from third-party sources), a mobile store application, a social media channel, an application for providing wholesale purchasing, and the like). Merchant-facing applications 142A-B may include applications that allow the merchant to administer their online store 138 (e.g., through applications related to the web or website or to mobile devices), run their business (e.g., through applications related to POS devices), to grow their business (e.g., through applications related to shipping (e.g., drop shipping), use of automated agents, use of process flow development and improvements), and the like. Integration applications may include applications that provide useful integrations that participate in the running of a business, such as shipping providers 112 and payment gateways 106.
As such, the e-commerce platform 100 can be configured to provide an online shopping experience through a flexible system architecture that enables merchants to connect with customers in a flexible and transparent manner. A typical customer experience may be better understood through an embodiment example purchase workflow, where the customer browses the merchant's products on a channel 110A-B, adds what they intend to buy to their cart, proceeds to checkout, and pays for the content of their cart resulting in the creation of an order for the merchant. The merchant may then review and fulfill (or cancel) the order. The product is then delivered to the customer. If the customer is not satisfied, they might return the products to the merchant.
In an example embodiment, a customer may browse a merchant's products through a number of different channels 110A-B such as, for example, the merchant's online store 138, a physical storefront through a POS device 152; an electronic marketplace, through an electronic buy button integrated into a website or a social media channel). In some cases, channels 110A-B may be modeled as applications 142A-B. A merchandising component in the commerce management engine 136 may be configured for creating, and managing product listings (using product data objects or models for example) to allow merchants to describe what they want to sell and where they sell it. The association between a product listing and a channel may be modeled as a product publication and accessed by channel applications, such as via a product listing API. A product may have many attributes and/or characteristics, like size and color, and many variants that expand the available options into specific combinations of all the attributes, like a variant that is size extra small and green, or a variant that is size large and blue. Products may have at least one variant (e.g., a “default variant”) created for a product without any options. To facilitate browsing and management, products may be grouped into collections, provided product identifiers (e.g., stock keeping unit (SKU)) and the like. Collections of products may be built by either manually categorizing products into one (e.g., a custom collection), by building rulesets for automatic classification (e.g., a smart collection), and the like. Product listings may include 2D images, 3D images or models, which may be viewed through a virtual or augmented reality interface, and the like.
In some embodiments, a shopping cart object is used to store or keep track of the products that the customer intends to buy. The shopping cart object may be channel specific and can be composed of multiple cart line items, where each cart line item tracks the quantity for a particular product variant. Since adding a product to a cart does not imply any commitment from the customer or the merchant, and the expected lifespan of a cart may be in the order of minutes (not days), cart objects/data representing a cart may be persisted to an ephemeral data store.
The customer then proceeds to checkout. A checkout object or page generated by the commerce management engine 136 may be configured to receive customer information to complete the order such as the customer's contact information, billing information and/or shipping details. If the customer inputs their contact information but does not proceed to payment, the e-commerce platform 100 may (e.g., via an abandoned checkout component) to transmit a message to the customer device 150 to encourage the customer to complete the checkout. For those reasons, checkout objects can have much longer lifespans than cart objects (hours or even days) and may therefore be persisted. Customers then pay for the content of their cart resulting in the creation of an order for the merchant. In some embodiments, the commerce management engine 136 may be configured to communicate with various payment gateways and services 106 (e.g., online payment systems, mobile payment systems, digital wallets, credit card gateways) via a payment processing component. The actual interactions with the payment gateways 106 may be provided through a card server environment. At the end of the checkout process, an order is created. An order is a contract of sale between the merchant and the customer where the merchant agrees to provide the goods and services listed on the order (e.g., order line items, shipping line items, and the like) and the customer agrees to provide payment (including taxes). Once an order is created, an order confirmation notification may be sent to the customer and an order placed notification sent to the merchant via a notification component. Inventory may be reserved when a payment processing job starts to avoid over selling (e.g., merchants may control this behavior using an inventory policy or configuration for each variant). Inventory reservation may have a short time span (minutes) and may need to be fast and scalable to support flash sales or “drops”, which are events during which a discount, promotion or limited inventory of a product may be offered for sale for buyers in a particular location and/or for a particular (usually short) time. The reservation is released if the payment fails. When the payment succeeds, and an order is created, the reservation is converted into a permanent (long-term) inventory commitment allocated to a specific location. An inventory component of the commerce management engine 136 may record where variants are stocked, and may track quantities for variants that have inventory tracking enabled. It may decouple product variants (a customer-facing concept representing the template of a product listing) from inventory items (a merchant-facing concept that represents an item whose quantity and location is managed). An inventory level component may keep track of quantities that are available for sale, committed to an order or incoming from an inventory transfer component (e.g., from a vendor).
The merchant may then review and fulfill (or cancel) the order. A review component of the commerce management engine 136 may implement a business process merchant's use to ensure orders are suitable for fulfillment before actually fulfilling them. Orders may be fraudulent, require verification (e.g., ID checking), have a payment method which requires the merchant to wait to make sure they will receive their funds, and the like. Risks and recommendations may be persisted in an order risk model. Order risks may be generated from a fraud detection tool, submitted by a third-party through an order risk API, and the like. Before proceeding to fulfillment, the merchant may need to obtain or capture the payment information (e.g., credit card information) or wait to receive it (e.g., via a bank transfer, check, and the like) before it marks the order as paid. The merchant may now prepare the products for delivery. In some embodiments, this business process may be implemented by a fulfillment component of the commerce management engine 136. The fulfillment component may group the line items of the order into a logical fulfillment unit of work based on an inventory location and fulfillment service. The merchant may review, adjust the unit of work, and trigger the relevant fulfillment services, such as through a manual fulfillment service (e.g., at merchant managed locations) used when the merchant picks and packs the products in a box, purchase a shipping label and input its tracking number, or just mark the item as fulfilled. Alternatively, an API fulfillment service may trigger a third party application or service to create a fulfillment record for a third-party fulfillment service. Other possibilities exist for fulfilling an order. If the customer is not satisfied, they may be able to return the product(s) to the merchant. The business process merchants may go through to “un-sell” an item may be implemented by a return component. Returns may consist of a variety of different actions, such as a restock, where the product that was sold actually comes back into the business and is sellable again; a refund, where the money that was collected from the customer is partially or fully returned; an accounting adjustment noting how much money was refunded (e.g., including if there was any restocking fees or goods that weren't returned and remain in the customer's hands); and the like. A return may represent a change to the contract of sale (e.g., the order), and where the e-commerce platform 100 may make the merchant aware of compliance issues with respect to legal obligations (e.g., with respect to taxes). In some embodiments, the e-commerce platform 100 may enable merchants to keep track of changes to the contract of sales over time, such as implemented through a sales model component (e.g., an append-only date based ledger that records sale-related events that happened to an item).
Implementation in an e-Commerce Platform
The functionality described herein may be used in commerce to provide improved customer or buyer experiences. The e-commerce platform 100 could implement the functionality for any of a variety of different applications, examples of which are described elsewhere herein.
Although the engine 300 is illustrated as a distinct component of the e-commerce platform 100 in
The engine 300 could implement at least some of the functionality described herein, for example based on the examples shown in
For simplicity and clarity of illustration, where considered appropriate, reference numerals may be repeated among the figures to indicate corresponding or analogous elements. In addition, numerous specific details are set forth in order to provide a thorough understanding of the examples described herein. However, it will be understood by those of ordinary skill in the art that the examples described herein may be practiced without these specific details. In other instances, well-known methods, procedures and components have not been described in detail so as not to obscure the examples described herein. Also, the description is not to be considered as limiting the scope of the examples described herein.
It will be appreciated that the examples and corresponding diagrams used herein are for illustrative purposes only. Different configurations and terminology can be used without departing from the principles expressed herein. For instance, components and modules can be added, deleted, modified, or arranged with differing connections without departing from these principles.
It will also be appreciated that any module or component exemplified herein that executes instructions may include or otherwise have access to computer readable media such as transitory or non-transitory storage media, computer storage media, or data storage devices (removable and/or non-removable) such as, for example, magnetic disks, optical disks, or tape. Computer storage media may include volatile and non-volatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data. Examples of computer storage media include RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other non-transitory computer readable medium which can be used to store the desired information and which can be accessed by an application, module, or both. Any such computer storage media may be part of the computing environment 200, computing device 240, computing system 400, e-commerce platform 100, or engine 300, any component of or related thereto, etc., or accessible or connectable thereto. Any application or module herein described may be implemented using computer readable/executable instructions that may be stored or otherwise held by such computer readable media.
The steps or operations in the flow charts and diagrams described herein are provided by way of example. There may be many variations to these steps or operations without departing from the principles discussed above. For instance, the steps may be performed in a differing order, or steps may be added, deleted, or modified.
Although the above principles have been described with reference to certain specific examples, various modifications thereof will be apparent to those skilled in the art as having regard to the appended claims in view of the specification as a whole.
This application claims priority to U.S. Provisional Patent Application No. 63/497,233 filed on Apr. 20, 2023, the contents of which are incorporated herein by reference in their entirety.
Number | Date | Country | |
---|---|---|---|
63497233 | Apr 2023 | US |