The invention relates generally to computer vision techniques and, more particularly to, computer vision techniques for adaptive advertising and marketing for retail applications.
Due to increasing competition and shrinking margins in the retail environments, retailers are interested in understanding the behaviors and purchase decision processes of their customers. Further, it is desirable to use this information in determining the advertising and/or marketing strategy for products. Typically, such information is obtained through direct observation of shoppers or indirectly via focus groups or specialized experiments in controlled environments. In particular, data is gathered using video, audio and other sensors observing people reacting to products. To obtain the information regarding the behaviors of the customers, several inspection techniques have been used. For example, downward looking stereo cameras are employed to track location of the shoppers in the retail environment. However, this requires dedicated stereo sensors, which are expensive and are uncommon in retail environments.
The gathered information regarding the behaviors of the shoppers is analyzed to determine factors of importance to marketing analysis. However, such process is labor-intensive and has low reliability. Therefore, manufacturers of products in the retail environment have to rely upon manual assessments and product sales as a guiding factor to determine success or failure of the products. Additionally, the current store advertisements are static entities and cannot be adjusted to enhance the sales of the products.
It is therefore desirable to provide a real-time, efficient, reliable, and cost-effective technique for obtaining information regarding behaviors of the shoppers in a retail environment. It is also desirable to provide techniques that enable adjusting the advertising and marketing strategy of the products based upon the obtained information.
Briefly, in accordance with one aspect of the invention, a method of adaptive advertising is provided. The method provides for obtaining at least one of demographic and behavioral profiles of a plurality of individuals in an environment and adjusting an advertising strategy in the environment of one or more products based upon the demographic and behavioral profiles of the plurality of individuals. Systems that afford such functionality may be provided by the present technique.
In accordance with another aspect of the present technique, a method is provided for enhancing sales of one or more products in a retail environment. The method provides for obtaining information regarding behavioral profiles of a plurality of individuals visiting the retail environment, analyzing the obtained information regarding the behavioral profiles of the individuals and changing at least one of an advertising strategy or a product marketing strategy of the one or more products in response to the information regarding the behavioral profiles of the plurality of individuals. Here again, systems affording such functionality may be provided by the present technique.
In accordance with a further aspect of the present technique, an adaptive advertising and marketing system is provided. The system includes a plurality of imaging devices, each device being configured to capture an image of one or more individuals in an environment and a video analytics system configured to receive captured images from the plurality of imaging devices and to extract at least one of demographic and behavioral profiles of the one or more individuals to change at least one of an advertising or a product market strategy of one or more products.
These and other advantages and features will be more readily understood from the following detailed description of preferred embodiments of the invention that is provided in connection with the accompanying drawings.
Embodiments of the invention are generally directed to detection of behaviors of individuals in an environment. Such techniques may be useful in a variety of applications such as marketing, merchandising, store operations and data mining that require efficient, reliable, cost-effective, and rapid monitoring of movement and behaviors of individuals. Although examples are provided herein in the context of retail environments, one of ordinary skill in the art will readily comprehend that embodiments may be utilized in other contexts and remain within the scope of the invention.
Referring now to
The system 10 further includes a video analytics system 22 configured to receive captured images from the plurality of imaging devices 12 and to extract at least one of demographic and behavioral profiles of the one or more individuals 16, 18 and 20. Further, the demographic and behavioral profiles of the one or more individuals 16, 18 and 20 are utilized to change an advertising strategy of one or more products available in the environment 14. Alternately, the demographic and behavioral profiles of the one or more individuals 16, 18 and 20 are utilized to change a product market strategy of the one or more products available in the environment 14. As used herein, the term “demographic profiles” refers to information regarding a demographic grouping of the one or more individuals 16, 18 and 20 visiting the environment 14. For example, the demographic profiles may include information regarding age bands, social class bands and gender of the one or more individuals 16, 18 and 20.
The behavioral profiles of the one or more individuals 16, 18 and 20 include information related to interaction of the one or more individuals 16, 18 and 20 with the one or more products. Moreover, the behavioral profiles also includes information related to interaction of the one or more individuals 16, 18 and 20 with products displays such as represented by reference numerals 24, 26 and 28. Examples of such information include, but are not limited to, a gaze direction of the individuals 16, 18 and 20, time spent by the individuals 16, 18 and 20 in browsing the product displays 24, 26 and 28, time spent by the individuals 16, 18 and 20 while interacting with the one or more products, number of eye gazes towards the one or more products or the product displays 24, 26 and 28.
The system 10 also includes one or more communication modules 30 disposed in the facility 14, and optionally at a remote location, to transmit still images or video signals to the video analytics server 22. The communication modules 30 include wired or wireless networks, which communicatively link the imaging devices 12 to the video analytics server 22. For example, the communication modules 16 may operate via telephone lines, cable lines, Ethernet lines, optical lines, satellite communications, radio frequency (RF) communications, and so forth.
The video analytics server 22 includes a processor 32 configured to process the still images or video signals and to extract the demographic and behavioral profiles of the one or more individuals 16, 18 and 20. Further, the video analytics server 22 includes a variety of software and hardware for performing facial recognition of the one or more individuals 16, 18 and 20 entering and traveling about the facility 14. For example, the video analytics server 22 may include file servers, application servers, web servers, disk servers, database servers, transaction servers, telnet servers, proxy servers, mail servers, list servers, groupware servers, File Transfer Protocol (FTP) servers, fax servers, audio/video servers, LAN servers, DNS servers, firewalls, and so forth.
The video analytics server 22 also includes one or more databases 34 and memory 36. The memory 36 may include hard disk drives, optical drives, tape drives, random access memory (RAM), read-only memory (ROM), programmable read-only memory (PROM), Redundant Arrays of Independent Disks (RAID), flash memory, magneto-optical memory, holographic memory, bubble memory, magnetic drum, memory stick, Mylar® tape, smartdisk, thin film memory, zip drive, and so forth. The database 34 may utilize the memory 36 to store facial images of the one or more individuals 16, 18 and 20, information about location of the individuals 16, 18 and 20, and other data or code to obtain behavioral and demographic profiles of the individuals 16, 18 and 20. Moreover, the system 10 includes a display 38 configured to display the demographic and behavioral profiles of the one or more individuals 16, 18 and 20 to a user of the system 10.
In operation, each imaging device 12 may acquire a series of images including facial images of the individual 16, 18 and 20 as they visit different sections within the environment 14. It should be noted that the plurality of imaging devices 12 are configured to obtain information regarding number and location of the one or more individuals 16, 18 and 20 visiting the different sections of the environment 14. The captured images from the plurality of imaging devices 12 are transmitted to the video analytics system 22. Further, the processor 32 is configured to process the captured images and to extract the demographic and behavioral profiles of the one or more individuals 16, 18 and 20.
In particular, the movement of the one or more individuals 16, 18 and 20 is tracked within the environment 14 and information regarding the demographics and behaviors of the individuals 16, 18 and 20 is extracted using the captured images via the imaging devices 12. In certain embodiments, information regarding an articulated motion, or a facial expression of the one or more individuals 16, 18 and 20 is extracted using the captured images. In certain embodiments, a customer gaze is determined for the individuals 16, 18 and 20 using face models such as active appearance models (AAM) that will be described in detail below with reference to
The demographic and behavioral profiles of the one or more individuals 16, 18 and 20 are further utilized to change the advertising or a product market strategy of the one or more products available in the environment. In particular, the processor 32 is configured to analyze the demographic and behavioral profiles and other information related to the one or more individuals 16, 18 and 20 and to develop a modified advertising or a product market strategy of the one or more products. For example, the modified advertising strategy may include customizing the product displays 24, 26 and 28 based upon the extracted demographic and behavioral profiles of the one or more individuals 16, 18 and 20.
Further, the modified product market strategy may include changing a location of the one or more products in the environment 14. Alternatively, the modified product market strategy may include changing a design or a quality of the one or more products in the environment 14. The modified advertising or a product market strategy of the one or more products may be made available to a user through the display 38. In certain the modified advertising strategy may be communicated to a controller 40 for controlling content of the product displays 24, 26 and 28 based upon the modified advertising strategy.
The processor 32 (
X={X
j=(xj,yj), j=0, . . . ,Nt} (1)
Each of the targets is associated with size and height information. Additionally, the target is composed of several parts. For example, a part k of the target may be denoted by Ok. When the target configuration X is projected into the image, a label image denoted by Oi=ki may be generated where at each image location i part ki is visible. It should be noted that if no part is visible, then Oi may be assigned a background label denoted by BG.
The probability of the foreground image F at time is represented by the following equation:
where: Ft[i] represents discretized probability of seeing foreground at image location i. The above equation (2) may be simplified to the following equation where constant contributions from the background BG may be factored out during optimization:
where hk(p) represents a histogram of likelihood ratios for part k given foreground pixel probabilities p.
The goal of the shopper detection task is to find the most likely target configuration (X) that maximizes equation (3). As will be appreciated by one skilled in the art certain assumptions and approximations may be made to facilitate real time execution of the shopper detection task. For example, projected ellipsoids may be approximated by their bounding boxes. Further, the bounding boxes may be subdivided into one or more several parts and separate body part labels may be assigned to top, middle and bottom third of the bounding box. In certain embodiments, targets may only be located at discrete ground plane locations in the camera view that allows a user to pre-compute the bounding boxes.
Once a shopper is detected in the environment 52, his movement and location is tracked as the shopper moves within the environment 52. The tracking of the shopper is performed in a similar manner as described above. In particular, at every step, detections are projected into the ground plane and may be supplied to a centralized tracker (not shown) that sequentially processes the locations of these detections from all camera views. Thus, tracking of extended targets in the imagery is reduced to tracking of two-dimensional point locations in the ground plane. In certain embodiments, the central tracker may operate on a physically separate processing node, connected to individual processing units that perform detection using a network connection. Further, the detections may be time stamped according to a synchronous clock, buffered and re-ordered by the central tracker before processing. In certain embodiments, the tracking may be performed using a joint probabilistic data association filter (JPDAF) algorithm. Alternatively, the tracking may be performed using Bayesian multi-target trackers. However, other tracking algorithms may be employed.
As described above, the shopping path 50 of the shopper may be tracked using the method described above. The tracking of shopping path 50 of shoppers in the environment 52 provides information such as about frequently visited sections of the environment 52 by the shoppers, time spent by the shoppers within different sections of the environment and so forth. Such information may be utilized to adjust the advertising or a product market strategy for enhancing sales of the one or more products available in the environment 52. For example, the location of the one or more products may be adjusted based upon such information. Further, location of the product displays and content displayed on the product displays may be adjusted based upon such information.
The arrival and departure information 60 may be utilized for adjusting the advertising strategy for the one or more products in the retail environment. In certain embodiments, such information 60 may be utilized to determine the staffing requirements for the retail environment during the day. Further, in certain embodiments, the arrival and departure information along with the demographic profiles of one or more individuals visiting the retail environment may be utilized to customize the advertising strategy of the one or more products.
Additionally, the captured images from the imaging devices 12 are processed to extract the behavioral profiles of the shoppers visiting the retail environment. In certain embodiments, a plurality of in-shelf imaging devices may be employed for estimating the gaze direction of the shoppers.
An AAM 86 applied to faces of a shopper is a two-stage model including a facial shape and appearance designed to fit the faces of different persons at different orientations. The shape model describes a distribution of locations of a set of land-mark points. In certain embodiments, principal component analysis (PCA) may be used to reduce a dimensionality of a shape space while capturing major modes of variation across a training set population. PCA is a statistical method for analysis of factors that reduces the large dimensionality of the data space (observed variables) to a smaller intrinsic dimensionality of feature space (independent variables) that describes the features of the image. In other words, PCA can be utilized to predict the features, remove redundant variants, extract relevant features, compress data, and so forth.
A generic AAM is trained using the training set having a plurality of images. Typically, the images come from different subjects to ensure that the trained AAM covers shapes and appearance variation of a relative large population. Advantageously, the trained AAM can be used to fit to facial image from an unseen object. Furthermore, model enhancement may be applied on the AAM trained with the manual labels.
It should be noted that a completely trained AAM can synthesize face images that vary continuously over appearance and shape. In certain embodiments, AAM is fit to a new face as it appears in a video frame. This may be achieved by solving for the face shape such that model synthesized face matches the face in the video frame warped with the shape parameters. In certain embodiments, simultaneous inverse compositional (SIC) algorithm may be employed to solve the fitting problem. Further, shape parameters may be utilized for estimating the gaze of the shopper.
In certain embodiments, facial images with various head poses may be used in the AAM training. As illustrated in
The various aspects of the methods and systems described hereinabove have utility in a variety of retail applications. The methods and systems described above enable detection and tracking of shoppers in retail environments. In particular, the methods and systems discussed herein utilize an efficient, reliable, and cost-effective technique for obtaining information regarding behaviors of shoppers in retail environments. Further, the embodiments described above also provide techniques that enable real-time adjustment of the advertising and marketing strategy of the products based upon the obtained information.
While the invention has been described in detail in connection with only a limited number of embodiments, it should be readily understood that the invention is not limited to such disclosed embodiments. Rather, the invention can be modified to incorporate any number of variations, alterations, substitutions or equivalent arrangements not heretofore described, but which are commensurate with the spirit and scope of the invention. Additionally, while various embodiments of the invention have been described, it is to be understood that aspects of the invention may include only some of the described embodiments. Accordingly, the invention is not to be seen as limited by the foregoing description, but is only limited by the scope of the appended claims.
This application claims priority to U.S. Provisional Application No. 60/908,991, filed on Mar. 30, 2007.
Number | Date | Country | |
---|---|---|---|
60908991 | Mar 2007 | US |