The present invention relates to a method for garment capture from a photograph.
The present invention contrives to solve the disadvantages of the prior art.
An object of the invention is to provide a method for garment capture from a photograph.
The method for garment capturing from a photograph of a garment comprises steps for:
inputting a photograph of the garment;
extracting a silhouette of the garment from the photograph;
identifying a garment type and a plurality of primary body sizes (PBSs) and creating a plurality of sized drafts;
generating a plurality of panels using the garment type and the plurality of PBSs; and
draping the plurality of panels on a mannequin.
The method may further comprise, prior to the step for inputting, steps for: providing a camera and the mannequin, wherein the positions of the camera and the mannequin are fixed, so that photographs taken with and without the garment have pixel-to-pixel correspondence; and pre-processing the mannequin to obtain and store three-dimensional geometry of the mannequin and primary body sizes (PBSs).
The step for pre-processing the mannequin may comprise steps for: scanning the mannequin; modeling the scanned data graphically; and storing the graphically modeled data in a computer file.
Relationship between real world distance and pixel distance of a plurality points of the mannequin and an environment in which the camera and the mannequin are disposed is established a computer using the graphically modeled data.
The step for extracting a silhouette may comprise a step for providing a base mask by subtracting an exposed mask from a mannequin mask, and the mannequin mask is obtained from the input photograph of the mannequin and the exposed mask comprises a non-garment region of the input photograph.
The step for identifying a garment type may comprise a step for searching a closest match from choices in a garment type database using
where SI is an input garment silhouette image, SD the silhouette in the garment type database, and T a transformation comprising an arbitrary combination of rotation, translation, and scaling.
The garment type database may comprise a plurality of classes and subclasses.
The step for identifying a plurality of primary body sizes (PBSs) may comprise a step for identifying, labeling, and pre-registering of mannequin-silhouette landmark points (MSLPs) and garment-silhouette landmark points (GSLPs).
The plurality of primary body sizes (PBSs) may be identified by searching candidate points of the garment-silhouette according to
where MF is one of the filters shown in
The method may further comprise a step for extracting one-repeat texture from the input photograph.
The step for extracting one-repeat texture may comprise steps for eliminating distortion first and then extracting the one-repeat texture from an undistorted image.
The step for extracting one-repeat texture may comprise a step for extracting lines by applying the Sobel filter, then constructing a 2D triangle mesh based on the extracted lines.
A deformation transfer technique may be applied to straighten the 2D triangle mesh, using an affine transformation T as
T={tilde over (V)}V
−1 (3)
for each triangle, where V and V− represent undeformed and deformed triangle matrices, respectively, and using only a smoothness term ES and an identity term EI,
and formulating the optimization problem as
where wS and wI are the user controlled weights, Lh and Lv are horizontal and vertical lines, respectively, and yV-i is y coordinate of vertex i.
Although the present invention is briefly summarized, the fuller understanding of the invention can be obtained by the following drawings, detailed description and appended claims.
The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the Office upon request and payment of the necessary fee.
These and other features, aspects and advantages of the present invention will become better understood with reference to the accompanying drawings, wherein:
Referring to the figures, the embodiments of the invention are described in detail.
Creation of virtual garments is demanded from various applications. This paper notes that such demand arises also from the consumers at home who would like to graphically coordinate the clothes in her closet to her own avatar. For that purpose, the existing garments need to be converted to virtual garments.
For the consumer, using the CAD programs for digitizing (i.e., identifying and creating the comprising cloth panels, positioning the panels around the body, defining the seams, extracting and mapping the textures, then draping on the avatar) her clothing collection is practically out of question. That job is difficult and cumbersome even for the clothing experts. This paper proposes a new method to instantly create the virtual garment from a single photograph of the existing garment put on to the mannequin, the setup of which is shown in
Millimeter-scale accuracy in the sewing pattern is not the quality this method promises. From insufficient information (thus easy to use), the method aims to create practically usable clothes that are just sufficient for the graphical outfit coordination. For the above purpose, the proposed method is very successful. As
We attribute the above success to the following two novel approaches this paper takes: (1) silhouette-based and (2) pattern-based. The use of vision-based techniques is not new in the context of virtual garment creation. Instead of trying to analyze the interior of the foreground, however, this paper devises a garment creation algorithm that utilizes only the silhouette, which can be captured a lot more robustly. This robustness trades-off with the foreground details such as buttons or collars, but we give up them in this paper to obtain a practically usable technique.
Another bifurcation this paper makes is, instead of working directly in the 3D-shape space, it works in the 2D-pattern space. In fact, our method is based on the pattern drafting theory which is well established in the conventional pattern-making study [1]. The proposed method is different from sketch or photograph based shape-in-3D-then-flatten approaches in that it does not call for flattening of the 3D surfaces. Flattening of a triangular mesh cannot be done in the theoretical (differential-geometrical) sense thus inevitably introduces errors, which emerge as unnaturalness to keen human eyes. Our method's obviation of the flattening significantly contributes to producing more realistic results.
Since it is based on pattern drafting, our work is applicable only to the types of garments whose drafting is already acquired. In this work, the goal of which is to demonstrate the potential of the proposed approach, we limit the scope to simple casual designs (shirt, skirt, pants, and one-piece dress) shown in
To summarize the contribution, to our knowledge, the proposed work is the first photograph based virtual garment creation technique that is based on the pattern drafting.
In the graphics field, there have been various studies for creating virtual garments. Turquin et al. [2] proposed a sketch-based framework, in which the user sketches the silhouette lines in 2D with respect to the body, which are then converted to the 3D garment. Decaudin et al. [3] proposed a more comprehensive technique that improved Turquin et al.'s work with the developability approximation and geometrical modeling of fabric folds.
The recent sketch-based method [4] is based on context-aware interpretation of the sketch strokes. We note that the above techniques are targeted to novel garment creation, not to capturing existing garments.
Some researchers used implicit markers (i.e., printed patterns) in order to capture the 3D shape of the garment [5, 6, 7]. Tanie et al. [5] presented a method for capturing detailed human motion and garment mesh from a suit covered with the meshes which are created with retro-reflective tape. Scholz et al. [6] used the garment on which a specialized color pattern is printed, which enabled reproduction of the 3D garment shape by establishing the correspondence among multi-view images. White et al. [7] used the color pattern of tessellated triangles to capture the occluded part as well as the folded geometry of the garment. We note that the above techniques are applicable to specially created clothes but not to the clothes in the consumers' closet.
A number of marker-free approaches have been also proposed for capturing garments from multi-view video capture [8, 9, 10, 11]. Bradley et al. [8] proposed a method that is based on the establishment of temporally coherent parameterization between the time-steps. Vlasic et al. [9] performed the skeletal pose estimation of the articulated figure, which was then used to estimate the mesh shape by processing the multi-view silhouettes. Aguiar et al. [10] took the approach of taking the full-body laser scan prior to the video-recording. Then, for each frame of the video, the method recovered the avatar pose and captured the surface details. Popa et al. [12] proposed a method to reintroduce high frequency folds, which tend to disappear in the video-based reconstruction of the garment. We note that the above multi-view techniques call for somewhat professional setup for the capture.
Zhou et al. [13] presented a method that generates the garment from a single image. Since the method assume the garment is symmetric in front part and rear part, it is hard to generate realistic rear part of the garment. The result can be useful if the clothing expert applies some additional processing, but not quite sufficient for the graphical coordination of the garments.
Our virtual garment creation is based on the drafts. Conventionally, there exists a draft (note that draft is different from the pattern; a draft is a collection of points and lines that are essential for obtaining the patterns or the cloth panels) for each garment type.
We note that in fact the drafting can be done from the input of just a few parameters [14]. For the case of the one-piece dress draft shown in
This section presents each of the steps overviewed in
Our photographing setup (
The first step of the GarmCap is the garment silhouette extraction, that is based on GrabCut [15] method. We already have the mannequin mask MM obtained from the mannequin image. We can get the exposed mask ME, the non-garment region of the input photograph. Subtracting ME from MM gives us the base mask MB.
With the garment silhouette extracted in Section 4.2, we identify the garment type from the choices in the current garment type DB (shirt, skirt, pants and one-piece dress) by searching the closest match with
where SI is the input garment silhouette image (e.g.,
A few points on the surface of the mannequin are pre-registered as the mannequin-silhouette landmark points (MSLPs). Garmcap identifies them and labels them with red circles as shown in
To identify the GSLPs from the garment silhouette, we search the candidate spots of the silhouette image according to
where MF is one of the filters shown in
This section describes how we extract one-repeat texture from the input image. Texture is a significant part of the garment without which the captured result would look monotonous. Note that our work is not based on vision-based reconstruction of the original surface, but it reproduces the garment by pattern-based construction and simulation.
In that approach, the conventional texture extraction (i.e., extracting the texture of the whole garment) produces poor results. The proposed method calls for extraction of an undistorted one-repeat texture. We pro-pose a simple texture extraction method that can approximately produce visual impression of the original garment in the limited cases of regular patterns consisting of straight lines.
We eliminate the distortion first and then extract one-repeat texture from undistorted image. We extract the lines by applying the Sobel filter, then construct a 2D triangle mesh based on the extracted lines as shown in
T={tilde over (V)}V
−1 (3)
To apply the deformation transfer method, we define the affine transformation T as
for each triangle, where V and V− represent undeformed and deformed triangle matrices, respectively. Using only the smoothness term ES and the identity term EI,
we formulate the optimization problem as
where wS and wI are the user controlled weights, Lh and Lv are horizontal and vertical lines, respectively, and yV-i is y coordinate of vertex i. We use weights wS=1.0 and wI=0.001 as in [16]. The optimization produces straightened results as shown in
After we get the garment type and the PBSs, we create the panels by supplying them to the parameterized drafting module. We map the one-repeat texture on the panels. Each garment type has the information on how to position and create seams between the panels. Each panel has the 3D coordinate for positioning. We has the index of the line pairs for stitching. After positioning and seaming panels, we perform the physically based clothing simulation [17, 18, 19].
We implemented the proposed garment capture method on a 3.2 GHz Intel Core™ i7-960 processor with 8 GB memory and a Nvidia GeForce GTX 560Ti video card. We ran the method to the left images of
Our experiments included three dresses (
There can exist some discrepancies between captured and real garments. We measured the discrepancies in the corresponding PBSs (of the captured and real garments). For the garments experimented in this paper, the discrepancy was bounded by 3 cm.
The proposed method reproduces the shoulder strap (
Intrinsically, the proposed method can not capture the input garment accurately when its draft does not exist in the database. In
In this work, we proposed a novel method GarmCap that generates the virtual garment from a single photograph of a real garment. The method got the insight from the drafting of the garments in the pattern-making study. GarmCap abstracted the drafting process into a computer module, which takes the garment type and PBSs to produce the draft as the output. For identifying the garment type, GarmCap matched the photographed garment silhouette with the selections in the database. The method extracted the PBSs based on the distances between the garment silhouette landmark points. GarmCap also extracted the one-repeat texture in some limited cases based on the deformation transfer technique.
The virtual garment captured from the input photograph looks quite similar to the real garment. The method did not require any panel-flattening procedure, which contributed to obtaining realistic results. Although we created the virtual garment based on the front image, the result is plausible even when it is viewed from an arbitrary view.
The proposed method is based on the silhouette of the garment. Therefore, the method is difficult to represent the non-silhouette details of the garment such as wrinkles, collars, stitches, pleats and pockets. Therefore it would be challenging for the method to represent complex dresses (including traditional costumes). In the future, we plan to investigate the methods for more comprehensive garment capture techniques that can represent the above features.
While the invention has been shown and described with reference to different embodiments thereof, it will be appreciated by those skilled in the art that variations in form, detail, compositions and operation may be made without departing from the spirit and scope of the invention as defined by the accompanying claims.