The claimed subject matter relates generally to image modification and, more specifically, to the dynamic retouching of live video images.
Manipulation of digital pictures to make the images more appealing is well known in the advertising and entertainment fields. Human faces and bodies are commonly the targets of such “photo retouching” or simply “retouching.” While retouching was once done exclusively by specialized people, or “retouching artists,” the process has in recent years become automated and performed by computing devices. Such automated photo retouching may be a complement or replacement for traditional makeup, hair and lighting experts.
In computing systems, faces may he automatically detected, measured and modified according to specific “fitness” criteria such as, but not limited to rules on dimension and color. In the case of video images, multiple frames may need modification such that the changes remain consistent through successive frames in a particular sequence of frames.
Provided are techniques for processing a subset of a plurality of video frames to produce a 3D model of a particular element within each frame of the subset; processing the plurality of video frames to identify a particular feature within the particular element in each frame of the plurality of frames in which the particular feature appears to produce a plurality of instances of the element, each instance of the plurality of instances associated with a corresponding frame of the plurality of frames; modifying each instance of the plurality of instances in accordance with user-defined parameters and in conformity with the 3D model to generate a plurality of modified instances; and replacing, each instance of the plurality of instances with a corresponding modified instance in the corresponding frame of the plurality of frames to produce a modified video file.
This summary is not intended as a comprehensive description of the claimed subject matter but, rather, is intended to provide a brief overview of some of the functionality associated therewith. Other systems, methods, functionality, features and advantages of the claimed subject matter will he or will become apparent to one with skill in the art upon examination of the following figures and detailed description.
A better understanding of the claimed subject matter can be obtained when the following, detailed description of the disclosed embodiments is considered in conjunction with the following figures, in which:
As will be appreciated by one skilled in the art, aspects of the present invention may be embodied as a system, method or computer program product. Accordingly, aspects of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects of the present invention may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device.
A computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may he any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device.
Program code embodied on a computer readable medium may he transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable. RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C++ or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
Aspects of the present invention are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational actions to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
Turning now to the figures,
In this example, computing system 102 and CPU 104 are connected to the Internet 120, which is also connected to a server computer, or simply “server,” 122. Although in this example, computing system 102 and server 122 are communicatively coupled via the Internet 120, they could also be coupled through any number of communication mediums such as, but not limited to, a local area network (LAN) (not shown). Coupled to server 122 is a CRSM 124. Typically, server 122 would also include a CPU, display, keyboard and mouse, like 104, 106, 108 and 110, but for the sake of simplicity they are not shown.
Also included in VCEA 100 is a video camera 126, which is illustrated with a wireless connection 128 to the Internet 120. Wireless connection 128 is used as one example of a means to transmit video files captured by video camera 126 to other devices such as computing system 100 and server 122. One with skill in the relevant arts will appreciate that they are many transmission devices such as, but not limited to, memory sticks and cables. Video camera 126 is used as an example of a device other than computing system 102 and server 122 that may also incorporate the disclosed technology as described below in conjunction with APRS 116. Other devices not illustrated that may incorporate the claimed subject matter, perhaps in conjunction with a built-in video recording device, include, but are not limited to, tablet computers and smart phones. Further, it should be noted there are many possible video capture and editing architectures and configurations, of which VCEA 100 is only one simple example. Further, the claimed subject matter does not necessarily have to he implemented upon devices that capture video as the disclosed technology is also applicable to devices that merely process video files captured by other devices.
Further, the representation of APRS 116 in
Input/Output (I/O) module 140 handles any communication APRS 116 has with other components of APRS 116 and computing system 102. Included in module 140 is a graphical user interface (GUI) 141. GUI 141 enables users of APRS 116 to interact with and to define the desired functionality of APRS 116.
Data module 142 is a data repository for information, including model and parameters, that APRS 116 requires during normal operation. Examples of the types of information stored in data module 142 include model data 152, modeling rules 154 and operational parameters 156. Model data 152 stores both 3D model templates and 3D models generated by APRS 116, sometimes based upon the templates. For example there might he various templates for rendering a human head with a model for a specific head detected within a video file based upon a modification of a similar, selected template. Throughout the Specification, the primary example will be the modeling and rendering of a human head and facial features although it should he understood that the disclosed techniques are equally applicable to other video elements.
Modeling rules 154 stores various algorithms for extrapolating data with respect to models stored in model data 152. In addition to various geometric rules, addition rules might include, but are not limited to, the ratios of distances between facial features, rules for correction due to the angle of and distance to a facial feature such as eyes, nose and mouth. Addition rules may apply to a deformation of a 3D mesh corresponding to a model, e.g., plastic deformation, or to the color of individual pixels depending upon lighting, angle and other criteria, e.g. makeup and skin polishing. Operation parameters 156 stores user set parameters for controlling APRS 116. Examples include, but are not limited to, resolution of changes or input, e.g. pixels/inch, a number of frames to use for establishing as 3D mesh and an indication of specific algorithms for procedures for addressing missing pixels and feature-removal.
FDM 144 processes images to detect specific, defined features within the images. For example, FDM 144 may be directed to detect and store data relating to the presence of faces in the images of a video tile (see 310,
FTM 146 generates a frame by frame correlation of features detected by FDM 144. MM 148 uses the data from FDM 144 and FTM 146 to generate a 3D model of the identified and tracked feature(s) (see 330,
MRM 150 modifies features identified by FEW 144 and tracked by FTM 146 in accordance with modeling rules 154, operational parameters 156 and corresponding models 152. Components 142, 144, 146, 148, 150, 152, 154 and 156 are described in more detail below in conjunction with
Process 200 starts in a “Begin Modify Video” block 202 and proceeds immediately to a “Receive Video” block 204. During processing associated with block 204, a video file is received for processing. Although described with respect to batch processing of a video file, it should be understood that aspects of the disclosed technology may also be processed frame by frame, i.e. serially. During processing associated with a “Receive Instructions” block 206, the particular modifications and operational parameters 156 (
During processing associated with a “Model Frames” 208, a particular feature or features identified in the instructions received during processing associated with block 206 are identified (see FDM 144,
During processing associated with a “Get Next Frame” block 210, the next frame for processing is identified. Typically, during the first time through block 210, the first frame that includes the feature targeted for modification is the next frame. During processing associated with an “Identify Feature(s)” block 212, the particular feature or features targeted for modification are identified in the frame retrieved during processing associated with block 210. During processing associated with a “modify Feature(s)” block 214, the feature or features identified during processing associated with block 212 are modified in accordance with the instructions received during processing associated with block 206. Once modified, the frame is saved in either the original tile or a modified video file.
During processing associated with a “More Frames?” block 216, a determination is made as to whether or not there are more frames to process. If so, control returns to block 210, the next frame is retrieved and processing continues as described above. It should be noted that a particular feature may appear in non-successive frames, i.e. a multiple intervals in a video file. In that case, the next frame may not he the next frame in the video file but rather the next frame in the video file that includes the identified feature or features. If, during block 216, a determination is made that there are no more frames to process, control proceeds to an “End Modify Video” block 219 in which process 200 is complete.
Process 250 starts in a “Begin Model Frames” block 252 and proceeds immediately to a “Get First N Frames” block 254. During processing associated with block 254, a specified number of frames corresponding to a video file to be modified are retrieved, The specific number ‘N’ may be specified by means of a user defined parameter (see 156,
During processing associated with a “Get Frame” block 256, one of the frames selected during processing associated with block 254 is identified for processing. Typically, frames are processed in chronological order and, therefore during the first iteration through block 256, the first frame selected is the one that was captured earliest. During processing associated with an “Identify Elements” block 258, in this example, human thee presence and position is detected by running face detection algorithms, which are well known to those with sill in the relevant arts and are currently implemented even in consumer-level product such as photo cameras (see 144,
During processing associated with a “More Frames?” block 260, a determination is made as to whether or not all the frames identified during processing associated with block 254 have been processed. If not, processing returns to block 256, the next unprocessed frame is selected and processing continues as described above. If so, control proceeds to “Correlate Frame Elements” block 262. During processing associated with block 262, the instances of the elements identified during processing associated with block 258 are correlated from frame to frame (see 146,
During processing associated with a “Build and Save 3D Model” block 264. the elements identified during processing associated with block 258 and correlated during processing associated with block 262 are employed to generate a 3D model of the particular feature that is to be modified (see 148,
Once 3D modeling has been completed, the model is saved in models 152 (
Frame 300 is displaying a character 306 with a face 308. An element ID box 312 is isolating one or more elements, which in the example are face 308 and hair 310 for analysis and modification in accordance with the claimed subject matter. As explained above in conjunction with
An element ID box 328 has selected particular elements for modification. In this example the color of hair 310 (
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
The corresponding structures, materials, acts, and equivalents of all means or step plus function elements in the claims below are intended to include any structure, material, or act for performing the function in combination with other claimed elements as specifically claimed. The description of the present invention has been presented for purposes of illustration and description, but is not intended to he exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the invention. The embodiment was chosen and described in order to best explain the principles of the invention and the practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as arc suited to the particular use contemplated.
The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block, in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, he executed substantially concurrently, or the blocks may sometimes he executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.