This application claims the benefit under 35 U.S.C. §119(a) of a Korean patent application filed on Apr. 9, 2013 in the Korean Intellectual Property Office and assigned Serial number 10-2013-0038630, the entire disclosure of which is hereby incorporated by reference.
The present disclosure relates to a three-Dimensional (3D) image conversion apparatus for converting a two-Dimensional (2D) image into a 3D image and a method for controlling the 3D conversion apparatus.
With the recent increasing interest in three-Dimensional (3D) images, research has been conducted into generation of 3D images. From the early stage of research on 3D graphics, the ultimate object of researchers is to generate a realistic graphic screen like a real image. Thus, research has been carried out using a polygonal model in the field of traditional modeling technology, and as a result, modeling and rendering have been developed enough to provide a very realistic 3D environment. However, a process of generating a complex model needs much effort and time of experts. Moreover, a realistic and complex environment needs a significant amount of information, and causes low efficiency in storage and transmission.
To address this issue, various 3D image rendering technologies have been developed.
However, when a 3D image is generated using different methods of the related art, depth information has to be given to objects of an image and such an operation performed for each frame increases time and computation. Moreover, in segmentation for identifying respective objects, a boundary of each object has to be designated.
Therefore, a need exists for a 3D image conversion apparatus for converting a 2D image into a 3D image and a method for controlling the 3D conversion apparatus.
The above information is presented as background information only to assist with an understanding of the present disclosure. No determination has been made, and no assertion is made, as to whether any of the above might be applicable as prior art with regard to the present disclosure.
Aspects of the present disclosure are to address at least the above-mentioned problems and/or disadvantages and to provide at least the advantages described below. Accordingly, an aspect of the present disclosure is to provide a three-Dimensional (3D) image conversion apparatus for converting a 3D image into a two-Dimensional (2D) image by designating an object through a user input and a method for controlling the 3D conversion apparatus.
In accordance with an aspect of the present disclosure, a method for controlling a 3D image conversion apparatus that converts a 2D image into a 3D image is provided. The method includes displaying the 2D image to be converted into the 3D image, receiving a user input designating at least one object included in the 2D image, obtaining boundaries of the at least one object included in the 2D image based on the received user input to identify each of the at least one object, analyzing the 2D image including the at least one object to obtain depth information of each of the at least one object, and arranging the identified each of the at least one object based on the obtained depth information to generate the 3D image.
In accordance with another aspect of the present disclosure, a 3D image conversion apparatus for converting a 2D image into a 3D image is provided. The 3D image conversion apparatus includes a touch screen configured to display the 2D image to be converted into the 3D image and receive a user input designating at least one object included in the 2D image and a controller configured to obtain boundaries of the at least one object included in the 2D image based on the received user input to identify each of the at least one object, to analyze the 2D image including the at least one object to obtain depth information of each of the at least one object, and to arrange the identified each of the at least one object based on the obtained depth information to generate the 3D image.
In accordance with another aspect of the present disclosure, a control method for a server transmitting data to and receiving data from a first terminal and a second terminal is provided. The control method includes receiving a 2D image from the first terminal, receiving a user input designating at least one object included in the 2D image, obtaining boundaries of the at least one object included in the 2D image based on the received user input to identify each of the at least one object, analyzing the 2D image including the at least one object to obtain depth information of each of the at least one object, and arranging the identified each of the at least one object based on the obtained depth information to generate a 3D image.
In accordance with another aspect of the present disclosure, a server for transmitting data to and receiving data from a first terminal and a second terminal is provided. The server includes a communication unit configured to receive a 2D image from the first terminal, a touch screen configured to display the 2D image and receive a user input designating at least one object included in the 2D image, and a controller configured to obtain boundaries of the at least one object included in the 2D image based on the received user input to identify each of the at least one object, to analyze the 2D image including the at least one object to obtain depth information of each of the at least one object, and to arrange the identified each of the at least one object based on the obtained depth information to generate a 3D image.
Other aspects, advantages, and salient features of the disclosure will become apparent to those skilled in the art from the following detailed description, which, taken in conjunction with the annexed drawings, discloses various embodiments of the present disclosure.
The above and other aspects, features, and advantages of certain embodiments of the present disclosure will be more apparent from the following description taken in conjunction with the accompanying drawings, in which:
Throughout the drawings, like reference numerals will be understood to refer to like parts, components, and structures.
The following description with reference to the accompanying drawings is provided to assist in a comprehensive understanding of various embodiments of the present disclosure as defined by the claims and their equivalents. It includes various specific details to assist in that understanding but these are to be regarded as merely exemplary. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the various embodiments described herein can be made without departing from the scope and spirit of the present disclosure. In addition, descriptions of well-known function and constructions may be omitted for clarity and conciseness.
The terms and words used in the following description and claims are not limited to the bibliographical meanings, but, are merely used by the inventor to enable a clear and consistent understanding of the present disclosure. Accordingly, it should be apparent to those skilled in the art that the following description of various embodiments of the present disclosure is provided for illustration purpose only and not for the purpose of limiting the present disclosure as defined by the appended claims and their equivalents.
It is to be understood that the singular forms “a,” “an,” and “the” include plural referents unless the context clearly dictates otherwise. Thus, for example, reference to “a component surface” includes reference to one or more of such surfaces.
By the term “substantially” it is meant that the recited characteristic, parameter, or value need not be achieved exactly, but that deviations or variations, including for example, tolerances, measurement error, measurement accuracy limitations and other factors known to skill in the art, may occur in amounts that do not preclude the effect the characteristic was intended to provide.
Referring to
Referring to
The controller 110 may include a Central Processing Unit (CPU) 111, a Read Only Memory (ROM) 112 in which a control program for controlling the 3D image conversion apparatus 100 is stored, and a Random Access Memory (RAM) 113 which memorizes a signal or data input from the 3D image conversion apparatus 100 or is used as a memory region for a task performed in the 3D image conversion apparatus 100. The CPU 111 may include a single core, a dual core, a triple core, or a quad core processor. The CPU 111, the ROM 112, and the RAM 113 may be interconnected through an internal bus.
The controller 110 may control the communication module 120, the sub communication module 130, the multimedia module 140, the camera module 150, the GPS module 155, the input/output module 160, the sensor module 170, the storage 175, the power supply unit 180, a first touch screen 190a, a second touch screen 190b, and the touch screen controller 195.
The mobile communication module 120 may facilitate the connection between the 3D image conversion apparatus 100 and an external device through mobile communication by using one or more antennas (not illustrated) under control of the controller 110. The mobile communication module 120 transmits/receives a wireless signal for a voice call, a video call, a text message (a Short Messaging Service (SMS)), and/or a multimedia message (a Multi Media Service (MMS)) with a cellular phone (not illustrated), a smart phone (not illustrated), a tablet PC, or another electronic device (not illustrated) which has a phone number input into the 3D image conversion apparatus 100.
The sub communication module 130 may include at least one of the WLAN module 131 and the short-range communication module 132. Alternatively, the sub communication module 130 may include either the WLAN module 131 or the short-range communication module 132, or both.
The WLAN module 131 may be connected to the Internet in a place where a wireless AP (not illustrated) is installed, under control of the controller 110. The WLAN module 131 supports the wireless LAN standard IEEE802.11x of the Institute of Electrical and Electronics Engineers (IEEE). The short-range communication module 132 may wirelessly perform short-range communication between the 3D image conversion apparatus 100 and an external electronic device under control of the controller 110. The short-range communication may include Bluetooth, Infrared Data Association (IrDA), or the like.
The 3D image conversion apparatus 100 may include at least one of the mobile communication module 120, the WLAN module 131, and the short-range communication module 132. For example, the 3D image conversion apparatus 100 may include a combination of the mobile communication module 120, the WLAN module 131, and the short-range communication module 132.
The multimedia module 140 may include the broadcast communication module 131, the audio playback module 142, or the video playback module 143. The broadcast communication module 141 receives a broadcast signal (for example, a TV broadcast signal, a radio broadcast signal, or a data broadcast signal) and broadcast additional information (for example, Electric Program Guide (EPG) or Electric Service Guide (ESG)) transmitted from a broadcasting station (not shown) via a broadcast communication antenna (not illustrated) under control of the controller 110. The audio playback module 142 may play a digital audio file (for example, a file having a file extension, such as ‘mp3’, ‘wma’, ‘ogg’, or ‘wav’) stored in the storage 175 or received under control of the controller 110. The video playback module 143 may play a digital video file (for example, a file having a file extension, such as ‘mpeg’, ‘mpg’, ‘mp4’, ‘avi’, ‘mov’, or ‘mkv’) stored or received under control of the controller 110. The video playback module 143 may play a digital audio file.
The multimedia module 140 includes the audio playback module 142 or the video playback module 143. The audio playback module 142 or the video playback module 143 of the multimedia module 140 may be included in the controller 110.
The camera module 150 may include the first camera 151 and the second camera 152 which capture a still image or a video under control of the controller 110. The camera module 150 may include an auxiliary light source (for example, a flash, not illustrated) for providing light necessary for photographing. The first camera 151 may be positioned on the front surface of the 3D image conversion apparatus 100, and the second camera 152 may be positioned on the rear surface of the 3D image conversion apparatus 100. In another embodiment, the first camera 151 and the second camera 152 may be positioned adjacent to each other (for example, a space between the first camera 151 and the second camera 152 is greater than 1 cm and less than 8 cm) to capture a 3D still image or a 3D moving image.
The GPS module 155 receives electric waves from a plurality of GPS satellites (not illustrated) in the Earth's orbit, and calculates a location of the portable device 100 by using a time of arrival from the GPS satellite (not illustrated) to the 3D image conversion apparatus 100.
The input/output module 160 may include a plurality of buttons 161, the microphone 162, the speaker 163, the vibration motor 164, the connector 165, and the keypad 166.
The buttons 161161 may be formed on at least one of a front surface, a side surface, and a rear surface of a housing (or case) of the 3D image conversion apparatus 100, and may include at least one of a power/lock button, a volume button, a menu button, a home button, a back button, and a search button.
The microphone 162 receives voice or sound and generates a corresponding electric signal under control of the controller 110.
The speaker 163 outputs sound corresponding to various signals or data (for example, wireless data, broadcast data, digital audio data, digital video data, or the like) under control of the controller 110. The speaker 163 may output sound corresponding to a function executed by the 3D image conversion apparatus 100 (for example, button manipulation sound corresponding to a phone call, a ring back tone, or voice of a counterpart user). One or more speakers 163 may be formed in a proper position or proper positions of the housing of the 3D image conversion apparatus 100.
The vibration motor 164 converts an electric signal into mechanical vibration under control of the controller 110. For example, in the 3D image conversion apparatus 100, in a vibration mode, if a voice call or a video call from another device (not illustrated) is received, the vibration motor 164 operates. One or more of the vibration motor 164 may be disposed in the housing of the 3D image conversion apparatus 100. The vibration motor 164 may operate in response to user input generated through the touch screen 190.
The connector 165 may be used as an interface for connecting the 3D image conversion apparatus 100 with an external device (not illustrated) or a power source (not illustrated). Under control of the controller 110, data stored in the storage 175 of the 3D image conversion apparatus 100 may be transmitted to an external electronic device or data may be received from the external electronic device through a wired cable connected to the connector 165. The 3D image conversion apparatus 100 receives power from the power source through the wired cable connected to the connector 165 or may charge a battery (not illustrated) by using the power source.
The keypad 166 receives key input from the user for control of the 3D image conversion apparatus 100. The keypad 166 includes a physical keypad (not illustrated) formed in the 3D image conversion apparatus 100 or a virtual keypad (not illustrated) displayed on the touch screen 190. The physical keypad (not illustrated) formed in the 3D image conversion apparatus 100 may be excluded according to the capability or structure of the 3D image conversion apparatus 100.
The sensor module 170 includes at least one sensor for detecting a state of the 3D image conversion apparatus 100. For example, the sensor module 170 may include at least one of a proximity sensor for detecting the user's proximity with respect to the 3D image conversion apparatus 100, an illumination sensor (not illustrated) for detecting an amount of light around the 3D image conversion apparatus 100, and a motion sensor (not illustrated) for detecting an operation of the 3D image conversion apparatus 100 (for example, rotation of the 3D image conversion apparatus 100 or acceleration or vibration applied to the 3D image conversion apparatus 100).
At least one sensor may detect a state, generate a signal corresponding to the detection, and transmit the signal to the controller 110. The sensors of the sensor module 170 may be removed or sensors may be added thereto according to the capability of the 3D image conversion apparatus 100.
The storage 175 stores a signal or data which is input/output corresponding to operations of the communication module 120, the sub communication module 130, the multimedia module 140, the camera module 150, the GPS module 155, the input/output module 160, the sensor module 170, and the touch screen 190, under control of the controller 110. The storage 175 stores a control program and applications for control of the 3D image conversion apparatus 100 and/or the controller 110.
The term “storing unit” includes the storing unit 175, the ROM 112 and the RAM 113 in the controller 110, or a memory card (not illustrated) mounted in the 3D image conversion apparatus 100 (for example, a Secure Digital (SD) card, a memory stick). The storing unit 175 may include a non-volatile memory, a volatile memory, a Hard Disk Drive (HDD), or a Solid State Drive (SSD).
The power supply unit 180 supplies power to one or more batteries disposed in the housing of the portable terminal 100 under control of the controller 110. The one or more batteries supply power to the 3D image conversion apparatus 100. The power supply unit 180 may also supply power input from an external power source through the wired cable connected with the connector 165 to the 3D image conversion apparatus 100. The power supply unit 180 may also supply power, which is wirelessly input from an external power source using a wireless charging technique, to the 3D image conversion apparatus 100.
The touch screen 190 provides a user graphic interface corresponding to various services (for example, call, data transmission, broadcasting, picture taking) to users. The touch screen 190 outputs an analog signal, which corresponds to at least one input to the user graphic interface, to the touch screen controller 195. The touch screen 190 receives at least one user inputs through a user's body (for example, a finger including a thumb) or an input unit (for example, a stylus pen 1). The touch screen 190 also receives a continuous movement of one of at least one touches. The touch screen 190 outputs an analog signal corresponding to the received continuous movement of the touch to the touch screen controller 195.
In an embodiment of the present disclosure, a touch is not limited to a contact between the touch screen 190 and a user's body or a touchable input unit, and may also include a non-contact touch (for example, when the user input unit is positioned within a distance of, for example, 1 cm) in which the user input unit may be detected. A distance or interval from the touch screen 190 within which the user input unit may be detected may be changed according to the capability or structure of the 3D image conversion apparatus 100.
The touch screen 190 may be implemented as, for example, a resistive type, a capacitive type, an infrared type, an acoustic wave type, or a combination thereof.
The touch screen controller 195 converts the analog signal received from the touch screen 190 into a digital signal (for example, X and Y coordinates) and transmits the digital signal to the controller 110. The controller 110 controls the touch screen 190 by using the digital signal received from the touch screen controller 195. For example, the controller 110 may control a shortcut icon (not illustrated) displayed on the touch screen 190 to be selected or executed in response to a touch. The touch screen controller 195 may be included in the controller 110.
The 3D image conversion apparatus 100 loads and displays a 2D image on a touch screen in operation S201. The 3D image conversion apparatus 100 captures an object to store a 2D image or receives a 2D image through communication from another device and stores the 2D image. For example, the 3D image conversion apparatus 100 may display a 2D image 300 as illustrated in
Referring to
Referring to
Referring to
The 3D image conversion apparatus 100 may store the preset object designation order. For example, the 3D image conversion apparatus 100 may store an object designation order of person-horizon-sky-ground-building with respect to an outdoor image. The 3D image conversion apparatus 100 indicates an object to be designated on the designating-object indication window 310 based on the stored object designation order. In the embodiment illustrated in
Referring to
Referring to
The 3D image conversion apparatus 100 may receive a user input 313 as a drag gesture input for the horizon object as illustrated in
As illustrated in
Herein, the edges may be made by points that form a line of a boundary between regions having a difference in pixel value therebetween. For example, the edges may be formed of points having non-zero first-order partial derivative values of a captured image. The edges may be obtained by calculating a partial derivative of a visible-ray captured image and using the calculated partial derivative value.
The corner points may be formed of a set of points having extremums of a captured image, and may be points having zero first-order partial derivative values and non-zero second-order partial derivative values of the captured image. Points that cannot be differentiated with respect to the captured image may also be regarded as extremums and determined as the corner points. The corner points may be obtained as unique values of a Hessian matrix introduced by Harris corner detection. The entire Hessian matrix may be formed of second-order partial derivatives of a continuous function.
The blob indicates a region having a larger or smaller pixel value than that of the periphery. The blob may be obtained using a Laplacian or Laplace operator of a second-order partial derivative with respect to each dimension (an x dimension and a y dimension) in a visible-ray captured image.
Referring back to
Referring to
The 3D image conversion apparatus 100 may also obtain depth information based on a start position of the ground object 303. For example, it can be seen from
The 3D image conversion apparatus 100 may also obtain depth information based on the object designation order. For example, the 3D image conversion apparatus 100 may obtain depth information of “10” for a person object which is designated first. For example, the 3D image conversion apparatus 100 may store a precedence relationship among objects in a general outdoor landscape image, and indicate designation of an object based on the precedence relationship.
As described above, the 3D image conversion apparatus 100 may obtain depth information for respective objects in various ways. However, the scope of the present disclosure is not limited by the way to obtain depth information, as will be understood by those of ordinary skill in the art.
Referring to
Referring to
Referring to
The 3D image conversion apparatus 100 receives a viewpoint change command from the user in operation S423. Herein, the viewpoint change command may be a command for changing a viewpoint from which the 3D image is viewed to have a different angle, and may include at least one of a viewpoint changing direction and a viewpoint changing level. For example, a user 2 may input a drag gesture 10 directed to the left from a point on the touch screen. The 3D image conversion apparatus 100 may recognize that a viewpoint is changed into a right viewpoint, based on the direction of the input drag gesture, the left direction. The 3D image conversion apparatus 100 may also recognize a viewpoint changing level from a touch distance of the drag gesture. The 3D image conversion apparatus 100 generates and displays the 3D image having the changed viewpoint based on the recognized the viewpoint changing direction and changing level in operation S425. For example, the 3D image conversion apparatus 100 may generate and display the viewpoint-changed 3D image as illustrated in
Moreover, as can be seen from
Referring to
Referring to
The 3D image conversion apparatus 100 provides a different object designation option according to whether the 2D image is the indoor scenery/outdoor landscape image in operation S703, and receives a user input based on the object designation option in operation S705. To be more specific, the 3D image conversion apparatus 100 may display a message on a different designating-object indication window according to whether the 2D image is the indoor scenery/outdoor landscape image. For example, in
The 3D image conversion apparatus 100 performs segmentation based on the received user input in operation S707 and obtains depth information for segmented objects in operation S709. The 3D image conversion apparatus 100 generates a 3D image based on the obtained depth information in operation S711.
Referring to
Referring to
The 3D image conversion apparatus 100 may store the preset object designation order. For example, the 3D image conversion apparatus 100 may store an object designation order of person-ceiling-wall-ground for the indoor scenery image. The 3D image conversion apparatus 100 indicates an object to be designated on the designating-object indication window 810 based on the stored object designation order. In the embodiment illustrated in
Referring to
Referring to
As described above, the 3D image conversion apparatus 100 according to the present disclosure determines the indoor scenery and the outdoor landscape, and provides a different object designation option based on the determination result.
Referring to
The 3D image conversion apparatus 100 loads a 2D image to be converted into a 3D image in operation S901. The 3D image conversion apparatus 100 performs object recognition in operation S903. For example, the 3D image conversion apparatus 100 may store a face recognition program or algorithm, and recognize a face of a person in a 2D image based on the stored face recognition program. Once the 3D image conversion apparatus 100 recognizes a face of a person, it may recognize an object connected to the recognized face as a person object. Moreover, the 3D image conversion apparatus 100 may recognize a sky object based on color determination. Based on various schemes described above, the 3D image conversion apparatus may recognize various objects in a 2D image.
The 3D image conversion apparatus 100 performs segmentation based on the recognition result in operation S905, and obtains depth information for each segmented object in operation S907. The 3D image conversion apparatus 100 generates a 3D image based on the obtained depth information in operation S909.
Referring to
For example, the 3D image conversion apparatus 100 may display a segmentation result as illustrated in
Referring to
Referring to
Referring to
Referring to
Referring to
Referring to
The object edition command is not limited to the drag-up gesture, and implemented by drag gestures in various directions. For example, if a downward drag gesture (or drag-down gesture) is input, the 3D image conversion apparatus 100 may reduce depth information of the object or lower the position of the object. The object edition command is not limited to a drag gesture and may be implemented with various forms of gestures, such as a flick gesture, tilting, a rotation gesture, and the like. The object edition command is not limited to a gesture input and may be implemented by manipulation of a button, such as a soft key.
Referring to
The first terminal 100 and the second terminal 1300 transmit and receive data to and from each other. For example, the first terminal 100 may transmit and receive data to and from the second terminal 1300 by using the mobile communication module 120 or the sub communication module 130. For example, the first terminal 100 transmits and receives data to and from the second terminal 1300 by using a MMS, the Internet, or short-range communication.
Referring to
Referring to
The operations of the server 1310 and the first to fourth terminals 100, 1300, 1320, and 1330 will be described below with reference to
Referring to
The server 1310 receives the generated 3D image from the first terminal 100 in operation S1603. The server 1310 stores the received 3D image in operation S1605, and transmits the stored 3D image to the second to fourth terminals 1300, 1320, and 1330 in operation S1607. The second to fourth terminals 1300, 1320, and 1330 display the received 3D image in operation S1609.
Referring to
Referring to
Certain aspects of the present disclosure can also be embodied as computer readable code on a non-transitory computer readable recording medium. A non-transitory computer readable recording medium is any data storage device that can store data which can be thereafter read by a computer system. Examples of the non-transitory computer readable recording medium include Read-Only Memory (ROM), Random-Access Memory (RAM), Compact Disc-ROMs (CD-ROMs), magnetic tapes, floppy disks, and optical data storage devices. The non-transitory computer readable recording medium can also be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion. In addition, functional programs, code, and code segments for accomplishing the present disclosure can be easily construed by programmers skilled in the art to which the present disclosure pertains.
At this point it should be noted that the various embodiments of the present disclosure as described above typically involve the processing of input data and the generation of output data to some extent. This input data processing and output data generation may be implemented in hardware or software in combination with hardware. For example, specific electronic components may be employed in a mobile device or similar or related circuitry for implementing the functions associated with the various embodiments of the present disclosure as described above. Alternatively, one or more processors operating in accordance with stored instructions may implement the functions associated with the various embodiments of the present disclosure as described above. If such is the case, it is within the scope of the present disclosure that such instructions may be stored on one or more non-transitory processor readable mediums. Examples of the processor readable mediums include a ROM, a RAM, CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices. The processor readable mediums can also be distributed over network coupled computer systems so that the instructions are stored and executed in a distributed fashion. In addition, functional computer programs, instructions, and instruction segments for accomplishing the present disclosure can be easily construed by programmers skilled in the art to which the present disclosure pertains.
The 3D image conversion apparatus may receive and store the program from a program providing device connected in a wired or wireless manner. The program providing device may include a memory for storing a program including instructions for instructing the 3D image conversion apparatus to execute the claimed method for controlling the 3D image conversion apparatus, information necessary for the method for controlling the 3D image conversion apparatus, a communication unit for performing wired or wireless communication with the 3D image conversion apparatus, and a controller for transmitting a corresponding program to the 3D image conversion apparatus at the request of the 3D image conversion apparatus or automatically.
According to various embodiments of the present disclosure, the 3D image conversion apparatus and the method for controlling the same are provided in which an object is designated by a simple user input for conversion of a 2D image into a 3D image. Thus, the user may designate an object by simply inputting the user input, such that the 3D image may be readily generated and viewed. Moreover, an object designation order suitable for an indoor scenery/outdoor landscape image is disclosed and object designation appropriate for a corresponding environment may be performed.
Other effects that may be obtained or expected from the embodiment of the present disclosure are explicitly or implicitly disclosed in the detailed description of the embodiment of the present disclosure. For example, various effects expected from the embodiment of the present disclosure have been disclosed in the detailed description of the present disclosure.
While the present disclosure has been shown and described with reference to various embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the present disclosure as defined by the appended claims and their equivalents.
Number | Date | Country | Kind |
---|---|---|---|
10-2013-0038630 | Apr 2013 | KR | national |