Pursuant to 35 U.S.C. §119(a), this application claims the benefit of earlier filing date and right of priority to Korean Application No. 10-2013-0095053, filed on Aug. 9, 2013, the contents of which are hereby incorporated by reference herein in its entirety.
1. Field of the Invention
The present invention relates to a mobile terminal capable of taking pictures without displaying a preview image of a camera and a method of controlling therefor.
2. Discussion of the Related Art
A mobile terminal is a device which may be configured to perform various functions. Examples of such functions include data and voice communications, capturing images and video via a camera, recording audio, playing music files and outputting music via a speaker system, and displaying images and video on a display.
Generally, terminals can be classified into mobile terminals and stationary terminals according to a presence or non-presence of mobility. And, the mobile terminals can be further classified into handheld terminals and vehicle mount terminals according to availability for hand-carry.
There are ongoing efforts to support and increase the functionality of mobile terminals. Such efforts include software and hardware improvements, as well as changes and improvements in the structural components which form the mobile terminal.
In general, a camera is installed in a mobile terminal for the purpose of performing a video call and taking pictures. When a user activates the camera to take pictures, the camera displays a preview image in a display to provide the user with a composition of a picture to be taken and information on a subject, which will be included in the picture to be taken.
Yet, if the user takes a picture while checking a preview image, it may take time, which is not trivial, until an actual picture is taken no matter how fast a processor may be. As an example, since a time taken for activating the camera and displaying the preview image and a time taken for capturing an actual picture after inputting a capturing command are nontrivial, it may take minimum 2˜3 seconds to take a picture no matter how fast the user intends to take a picture. Moreover, if the user further performs such a process as adjusting a configuration value of the camera in accordance with a capturing environment or setting a subject to be taken to a focus area, the time taken for capturing a picture cannot but increase.
Hence, a current mobile terminal may be inappropriate for capturing a subject which is needed to be captured in a moment.
Accordingly, embodiments of the present invention are directed to a mobile terminal and controlling method thereof that substantially obviate one or more problems due to limitations and disadvantages of the related art.
The present invention is contrived by the aforementioned necessity. An object of the present invention is to provide a mobile terminal for enhancing user convenience.
Specifically, an object of the present invention is to provide a mobile terminal capable of quickly capturing pictures without displaying a preview image.
Moreover, another object of the present invention is to provide a mobile terminal capable of automatically focusing a camera on a subject or adjusting a camera setting to correspond to an intention of a user.
To achieve these objects and other advantages and in accordance with the purpose of the invention, as embodied and broadly described herein, according to one embodiment, a mobile terminal includes a camera configured to capture a picture, a display configured to display information, and a controller configured to control the camera to capture a picture in response to a user input and without displaying a preview image received via the camera, and the controller to subsequently control the captured picture to be displayed on the display simultaneously with the preview image after the picture is captured.
To further achieve these objects and other advantages and in accordance with the purpose of the invention, as embodied and broadly described herein, according to one embodiment, a method of controlling a mobile terminal includes receiving a user input, in response to receiving the user input, capturing a picture without displaying a preview image received via a camera, and subsequently displaying, on the display, the captured picture together with the preview image after the picture is captured.
Effects obtainable from the present invention may be non-limited by the above mentioned effect. And, other unmentioned effects can be clearly understood from the following description by those having ordinary skill in the technical field to which the present invention pertains.
It is to be understood that both the foregoing general description and the following detailed description of the present invention are exemplary and explanatory and are intended to provide further explanation of the invention as claimed.
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this application, illustrate embodiment(s) of the invention and together with the description serve to explain the principle of the invention. The above and other aspects, features, and advantages of the present invention will become more apparent upon consideration of the following description of preferred embodiments, taken in conjunction with the accompanying drawing figures. In the drawings:
a to
a and
a to
a to
a and
a and
a and
In the following detailed description, reference is made to the accompanying drawing figures which form a part hereof, and which show by way of illustration specific embodiments of the invention. It is to be understood by those of ordinary skill in this technological field that other embodiments may be utilized, and structural, electrical, as well as procedural changes may be made without departing from the scope of the present invention. Wherever possible, the same reference numbers will be used throughout the drawings to refer to the same or similar parts.
As used herein, the suffixes ‘module’, ‘unit’ and ‘part’ are used for elements in order to facilitate the disclosure only. Therefore, significant meanings or roles are not given to the suffixes themselves and it is understood that the ‘module’, ‘unit’ and ‘part’ can be used together or interchangeably.
The present invention can be applicable to a various types of mobile terminals. Examples of such terminals include mobile phones, user equipments, smart phones, digital broadcast receivers, personal digital assistants, laptop computers, portable multimedia players (PMP), navigators and the like.
Yet, it is apparent to those skilled in the art that a configuration according to an embodiment disclosed in this specification is applicable to such a fixed terminal as a digital TV, a desktop computer and the like as well as a mobile terminal.
In the following description, the above elements of the mobile terminal 100 are explained in sequence.
First of all, the wireless communication unit 110 typically includes one or more components which permits wireless communication between the mobile terminal 100 and a wireless communication system or network within which the mobile terminal 100 is located. For instance, the wireless communication unit 110 can include a broadcast receiving module 111, a mobile communication module 112, a wireless internet module 113, a short-range communication module 114, a position-location module 115 and the like.
The broadcast receiving module 111 receives a broadcast signal and/or broadcast associated information from an external broadcast managing server via a broadcast channel. The broadcast channel may include a satellite channel and a terrestrial channel. At least two broadcast receiving modules 111 can be provided to the mobile terminal 100 in pursuit of simultaneous receptions of at least two broadcast channels or broadcast channel switching facilitation.
The broadcast managing server generally refers to a server which generates and transmits a broadcast signal and/or broadcast associated information or a server which is provided with a previously generated broadcast signal and/or broadcast associated information and then transmits the provided signal or information to a terminal. The broadcast signal may be implemented as a TV broadcast signal, a radio broadcast signal, and a data broadcast signal, among others. If desired, the broadcast signal may further include a broadcast signal combined with a TV or radio broadcast signal.
The broadcast associated information includes information associated with a broadcast channel, a broadcast program, a broadcast service provider, etc. And, the broadcast associated information can be provided via a mobile communication network. In this case, the broadcast associated information can be received by the mobile communication module 112.
The broadcast associated information can be implemented in various forms. For instance, broadcast associated information may include an electronic program guide (EPG) of digital multimedia broadcasting (DMB) and electronic service guide (ESG) of digital video broadcast-handheld (DVB-H).
The broadcast receiving module 111 may be configured to receive broadcast signals transmitted from various types of broadcast systems. By non-limiting example, such broadcasting systems include digital multimedia broadcasting-terrestrial (DMB-T), digital multimedia broadcasting-satellite (DMB-S), digital video broadcast-handheld (DVB-H), Convergence of Broadcasting and Mobile Service (DVB-CBMS), Open Mobile Alliance-BroadCAST (OMA-BCAST), China Multimedia Mobile Broadcasting (CMMB), Mobile Broadcasting Business Management System (MBBMS), the data broadcasting system known as media forward link only (MediaFLO®) and integrated services digital broadcast-terrestrial (ISDB-T). Optionally, the broadcast receiving module 111 can be configured suitable for other broadcasting systems as well as the above-explained digital broadcasting systems.
The broadcast signal and/or broadcast associated information received by the broadcast receiving module 111 may be stored in a suitable device, such as a memory 160.
The mobile communication module 112 transmits/receives wireless signals to/from one or more network entities (e.g., base station, external terminal, server, etc.) via a mobile network such as GSM (Global System for Mobile communications), CDMA (Code Division Multiple Access), WCDMA (Wideband CDMA) and so on. Such wireless signals may represent audio, video, and data according to text/multimedia message transceivings, among others.
The wireless internet module 113 supports Internet access for the mobile terminal 100. This module may be internally or externally coupled to the mobile terminal 100. In this case, the wireless Internet technology can include WLAN (Wireless LAN) (Wi-Fi), Wibro (Wireless broadband), Wimax (World Interoperability for Microwave Access), HSDPA (High Speed Downlink Packet Access), GSM, CDMA, WCDMA, LTE (Long Term Evolution) etc.
Wireless internet access by Wibro, HSPDA, GSM, CDMA, WCDMA, LTE or the like is achieved via a mobile communication network. In this aspect, the wireless internet module 113 configured to perform the wireless internet access via the mobile communication network can be understood as a sort of the mobile communication module 112.
The short-range communication module 114 facilitates relatively short-range communications. Suitable technologies for implementing this module include radio frequency identification (RFID), infrared data association (IrDA), ultra-wideband (UWB), as well at the networking technologies commonly referred to as Bluetooth and ZigBee, to name a few.
The position-location module 115 identifies or otherwise obtains the location of the mobile terminal 100. If desired, this module may be implemented with a global positioning system (GPS) module. According to the current technology, the GPS module 115 is able to precisely calculate current 3-dimensional position information based on at least one of longitude, latitude and altitude and direction (or orientation) by calculating distance information and precise time information from at least three satellites and then applying triangulation to the calculated information. Currently, location and time informations are calculated using three satellites, and errors of the calculated location position and time informations are then amended using another satellite. Besides, the GPS module 115 is able to calculate speed information by continuously calculating a real-time current location.
Referring to
The image frames processed by the camera 121 can be stored in the memory 160 or can be externally transmitted via the wireless communication unit 110. Optionally, at least two cameras 121 can be provided to the mobile terminal 100 according to environment of usage.
The microphone 122 receives an external audio signal while the portable device is in a particular mode, such as phone call mode, recording mode and voice recognition. This audio signal is processed and converted into electric audio data. The processed audio data is transformed into a format transmittable to a mobile communication base station via the mobile communication module 112 in case of a call mode. The microphone 122 typically includes assorted noise removing algorithms to remove noise generated in the course of receiving the external audio signal.
The user input unit 130 generates input data responsive to user manipulation of an associated input device or devices. Examples of such devices include a button 136 provided to front/rear/lateral side of the mobile terminal 100 and a touch sensor (constant pressure/electrostatic) 137 and may further include a key pad, a dome switch, a jog wheel, a jog switch and the like [not shown in the drawing].
The sensing unit 140 provides sensing signals for controlling operations of the mobile terminal 100 using status measurements of various aspects of the mobile terminal. For instance, the sensing unit 140 may detect an open/close status of the mobile terminal 100, relative positioning of components (e.g., a display and keypad) of the mobile terminal 100, a change of position of the mobile terminal 100 or a component of the mobile terminal 100, a presence or absence of user contact with the mobile terminal 100, orientation or acceleration/deceleration of the mobile terminal 100. By non-limiting example, such sensing unit 140 include, gyro sensor, accelerate sensor, geomagnetic sensor.
As an example, consider the mobile terminal 100 being configured as a slide-type mobile terminal. In this configuration, the sensing unit 140 may sense whether a sliding portion of the mobile terminal is open or closed. Other examples include the sensing unit 140 sensing the presence or absence of power provided by the power supply 190, the presence or absence of a coupling or other connection between the interface unit 170 and an external device. And, the sensing unit 140 can include a proximity sensor 141.
The output unit 150 generates outputs relevant to the senses of sight, hearing, touch and the like. And, the output unit 150 includes the display 151, an audio output module 152, an alarm unit 153, and a haptic module 154 and the like.
The display 151 is typically implemented to visually display (output) information associated with the mobile terminal 100. For instance, if the mobile terminal is operating in a phone call mode, the display will generally provide a user interface (UI) or graphical user interface (GUI) which includes information associated with placing, conducting, and terminating a phone call. As another example, if the mobile terminal 100 is in a video call mode or a photographing mode, the display 151 may additionally or alternatively display images which are associated with these modes, the UI or the GUI.
The display module 151 may be implemented using known display technologies including, for example, a liquid crystal display (LCD), a thin film transistor-liquid crystal display (TFT-LCD), an organic light-emitting diode display (OLED), a flexible display and a three-dimensional display. The mobile terminal 100 may include one or more of such displays.
Some of the above displays can be implemented in a transparent or optical transmittive type, which can be named a transparent display. As a representative example for the transparent display, there is TOLED (transparent OLED) or the like. A rear configuration of the display 151 can be implemented in the optical transmittive type as well. In this configuration, a user is able to see an object in rear of a terminal body via the area occupied by the display 151 of the terminal body.
At least two displays 151 can be provided to the mobile terminal 100 in accordance with the implemented configuration of the mobile terminal 100. For instance, a plurality of displays can be arranged on a single face of the mobile terminal 100 in a manner of being spaced apart from each other or being built in one body. Alternatively, a plurality of displays can be arranged on different faces of the mobile terminal 100.
In case that the display 151 and the touch sensor 137 configures a mutual layer structure (hereinafter called ‘touch screen’), it is able to use the display 151 as an input device as well as an output device. In this case, the touch sensor can be configured as a touch film, a touch sheet, a touch pad or the like.
The touch sensor 137 can be configured to convert a pressure applied to a specific portion of the display 151 or a variation of a capacitance generated from a specific portion of the display 151 to an electric input signal. Moreover, it is able to configure the touch sensor 137 to detect a pressure of a touch as well as a touched position or size.
If a touch input is made to the touch sensor 137, signal(s) corresponding to the touch is transferred to a touch controller. The touch controller processes the signal(s) and then transfers the processed signal(s) to the controller 180. Therefore, the controller 180 is able to know whether a prescribed portion of the display 151 is touched.
Referring to
The proximity sensor can include one of a transmittive photoelectric sensor, a direct reflective photoelectric sensor, a mirror reflective photoelectric sensor, a radio frequency oscillation proximity sensor, an electrostatic capacity proximity sensor, a magnetic proximity sensor, an infrared proximity sensor and the like. In case that the touchscreen includes the electrostatic capacity proximity sensor, it is configured to detect the proximity of a pointer using a variation of electric field according to the proximity of the pointer. In this case, the touchscreen (touch sensor) can be classified as the proximity sensor.
For clarity and convenience of the following description, as a pointer becomes proximate to a touchscreen without coming into contact with the touchscreen, if the pointer is perceived as situated over the touchscreen, such an action shall be named ‘proximity touch’. If a pointer actually comes into contact with a touchscreen, such an action shall be named ‘contact touch’. A proximity-touched position over the touchscreen with the pointer may mean a position at which the pointer vertically opposes the touchscreen when the touchscreen is proximity-touched with the pointer.
The proximity sensor detects a proximity touch and a proximity touch pattern (e.g., a proximity touch distance, a proximity touch duration, a proximity touch position, a proximity touch shift state, etc.). And, information corresponding to the detected proximity touch action and the detected proximity touch pattern can be outputted to the touchscreen.
The audio output module 152 functions in various modes including a call-receiving mode, a call-placing mode, a recording mode, a voice recognition mode, a broadcast reception mode and the like to output audio data which is received from the wireless communication unit 110 or is stored in the memory 160. During operation, the audio output module 152 outputs audio relating to a particular function (e.g., call received, message received, etc.). The audio output module 152 is often implemented using one or more speakers, buzzers, other audio producing devices, and combinations thereof.
The alarm unit 153 is output a signal for announcing the occurrence of a particular event associated with the mobile terminal 100. Typical events include a call received event, a message received event and a touch input received event. The alarm unit 153 is able to output a signal for announcing the event occurrence by way of vibration as well as video or audio signal. The video or audio signal can be outputted via the display 151 or the audio output unit 152. Hence, the display 151 or the audio output module 152 can be regarded as a part of the alarm unit 153.
The haptic module 154 generates various tactile effects that can be sensed by a user. Vibration is a representative one of the tactile effects generated by the haptic module 154. Strength and pattern of the vibration generated by the haptic module 154 are controllable. For instance, different vibrations can be outputted in a manner of being synthesized together or can be outputted in sequence.
The haptic module 154 is able to generate various tactile effects as well as the vibration. For instance, the haptic module 154 generates the effect attributed to the arrangement of pins vertically moving against a contact skin surface, the effect attributed to the injection/suction power of air though an injection/suction hole, the effect attributed to the skim over a skin surface, the effect attributed to the contact with electrode, the effect attributed to the electrostatic force, the effect attributed to the representation of hold/cold sense using an endothermic or exothermic device and the like.
The haptic module 154 can be implemented to enable a user to sense the tactile effect through a muscle sense of finger, arm or the like as well as to transfer the tactile effect through a direct contact. Optionally, at least two haptic modules 154 can be provided to the mobile terminal 100 in accordance with the corresponding configuration type of the mobile terminal 100.
The memory unit 160 is generally used to store various types of data to support the processing, control, and storage requirements of the mobile terminal 100. Examples of such data include program instructions for applications operating on the mobile terminal 100, contact data, phonebook data, messages, audio, still pictures (or photo), moving pictures, etc. And, a recent use history or a cumulative use frequency of each data (e.g., use frequency for each phonebook, each message or each multimedia) can be stored in the memory unit 160. Moreover, data for various patterns of vibration and/or sound outputted in case of a touch input to the touchscreen can be stored in the memory unit 160.
The memory 160 may be implemented using any type or combination of suitable volatile and non-volatile memory or storage devices including hard disk, random access memory (RAM), static random access memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disk, multimedia card micro type memory, card-type memory (e.g., SD memory, XD memory, etc.), or other similar memory or data storage device. And, the mobile terminal 100 is able to operate in association with a web storage for performing a storage function of the memory 160 on Internet.
The interface unit 170 is often implemented to couple the mobile terminal 100 with external devices. The interface unit 170 receives data from the external devices or is supplied with the power and then transfers the data or power to the respective elements of the mobile terminal 100 or enables data within the mobile terminal 100 to be transferred to the external devices. The interface unit 170 may be configured using a wired/wireless headset port, an external charger port, a wired/wireless data port, a memory card port, a port for coupling to a device having an identity module, audio input/output ports, video input/output ports, an earphone port and/or the like.
The identity module is the chip for storing various kinds of information for authenticating a use authority of the mobile terminal 100 and can include User Identify Module (UIM), Subscriber Identify Module (SIM), Universal Subscriber Identity Module (USIM) and/or the like. A device having the identity module (hereinafter called ‘identity device’) can be manufactured as a smart card. Therefore, the identity device is connectible to the mobile terminal 100 via the corresponding port.
When the mobile terminal 110 is connected to an external cradle, the interface unit 170 becomes a passage for supplying the mobile terminal 100 with a power from the cradle or a passage for delivering various command signals inputted from the cradle by a user to the mobile terminal 100. Each of the various command signals inputted from the cradle or the power can operate as a signal enabling the mobile terminal 100 to recognize that it is correctly loaded in the cradle.
The controller 180 typically controls the overall operations of the mobile terminal 100. For example, the controller 180 performs the control and processing associated with voice calls, data communications, video calls, etc. The controller 180 may include a multimedia module 181 that provides multimedia playback. The multimedia module 181 may be configured as part of the controller 180, or implemented as a separate component.
Moreover, the controller 180 is able to perform a pattern (or image) recognizing process for recognizing a writing input and a picture drawing input carried out on the touchscreen as characters or images, respectively.
The power supply unit 190 provides power required by the various components for the mobile terminal 100. The power may be internal power, external power, or combinations thereof.
A battery may include a built-in rechargeable battery and may be detachably attached to the terminal body for a charging and the like. A connecting port may be configured as one example of the interface 170 via which an external charger for supplying a power of a battery charging is electrically connected.
Various embodiments described herein may be implemented in a computer-readable medium using, for example, computer software, hardware, or some combination thereof.
For a hardware implementation, the embodiments described herein may be implemented within one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, other electronic units designed to perform the functions described herein, or a selective combination thereof. Such embodiments may also be implemented by the controller 180.
For a software implementation, the embodiments described herein may be implemented with separate software modules, such as procedures and functions, each of which perform one or more of the functions and operations described herein. The software codes can be implemented with a software application written in any suitable programming language and may be stored in memory such as the memory 160, and executed by a controller or processor, such as the controller 180.
The mobile terminal 100 shown in the drawing has a bar type terminal body. Yet, the mobile terminal 100 may be implemented in a variety of different configurations. Examples of such configurations include folder-type, slide-type, rotational-type, swing-type and combinations thereof. For clarity, further disclosure will primarily relate to a bar-type mobile terminal 100. However such teachings apply equally to other types of mobile terminals.
Referring to
Occasionally, electronic components can be mounted on a surface of the rear case 102. The electronic part mounted on the surface of the rear case 102 may include such a detachable part as a battery, a USIM card, a memory card and the like. In doing so, the rear case 102 may further include a backside cover 103 configured to cover the surface of the rear case 102. In particular, the backside cover 103 has a detachable configuration for user's convenience. If the backside cover 103 is detached from the rear case 102, the surface of the rear case 102 is exposed.
Referring to
The cases 101, 102 and 103 are formed by injection molding of synthetic resin or can be formed of metal substance such as stainless steel (STS), titanium (Ti) or the like for example.
A display 151, an audio output unit 152, a camera 121, user input units 130/131 and 132, a microphone 122, an interface 180 and the like can be provided to the case 101 or 102.
The display 151 occupies most of a main face of the front case 101. The audio output unit 152 and the camera 121 are provided to an area adjacent to one of both end portions of the display 151, while the user input unit 131 and the microphone 122 are provided to another area adjacent to the other end portion of the display 151. The user input unit 132 and the interface 170 can be provided to lateral sides of the front and rear cases 101 and 102.
The input unit 130 is manipulated to receive a command for controlling an operation of the terminal 100. And, the input unit 130 is able to include a plurality of manipulating units 131 and 132. The manipulating units 131 and 132 can be named a manipulating portion and may adopt any mechanism of a tactile manner that enables a user to perform a manipulation action by experiencing a tactile feeling.
Content inputted by the first or second manipulating unit 131 or 132 can be diversely set. For instance, such a command as start, end, scroll and the like is inputted to the first manipulating unit 131. And, a command for a volume adjustment of sound outputted from the audio output unit 152 and the like can be inputted to the second manipulating unit 132, a command for a switching to a touch recognizing mode of the display 151 and the like can be inputted to the second manipulating unit 133.
Referring to
Preferably, for instance, the front camera 121 has low pixels enough to capture and transmit a picture of user's face for a video call, while the rear camera 121′ has high pixels for capturing a general subject for photography without transmitting the captured subject. And, each of the cameras 121 and 121′ can be installed at the terminal body to be rotated or popped up.
A flash 123 and a mirror 124 are additionally provided adjacent to the camera 121′. The flash 123 projects light toward a subject in case of photographing the subject using the camera 121′. In case that a user attempts to take a picture of the user (self-photography) using the camera 121′, the mirror 124 enables the user to view user's face reflected by the mirror 124.
An additional audio output unit 152′ can be provided to the backside of the terminal body. The additional audio output unit 152′ is able to implement a stereo function together with the former audio output unit 152 shown in
A broadcast signal receiving antenna 116 can be additionally provided to the lateral side of the terminal body as well as an antenna for communication or the like. The antenna 116 constructing a portion of the broadcast receiving module 111 shown in
For clarity, assume that the mobile terminal 100 according to the present invention includes at least one of the configuration elements depicted in
The mobile terminal 100 according to the present invention may include a single camera 121 or may include the front camera 121 and the rear camera 121′ as depicted in the examples of
And, if the display 151 of the mobile terminal 100 according to the present invention corresponds to a touch screen, which plays a role of an input device for receiving a user input as well as an output device for outputting information, the present invention may be more easily implemented. Hence, assume that the display 151 corresponds to the touch screen in the following description. Yet, it is not mandatory that the display 151 according to the present invention corresponds to the touch screen. If the mobile terminal 100 according to the present invention does not correspond to the touch screen, the mobile terminal 100 may receive a user input via separate input devices configured to receive the user input (of course, although the display 151 corresponds to the touch screen, the mobile terminal can include separate input devices except the touch screen).
The mobile terminal 100 according to the present invention is explained in detail with reference to drawings in the following.
In this case, various methods can be applied to the first and the second user input. Regarding this, it shall be described with reference to
a to
As depicted in an example of
As a different example, as depicted in an example
As depicted in an example
As a different example, as depicted in
As depicted in an example
In order to increase accessibility of a user, an icon (or button) for inputting the first and the second user input may be displayed all the time in the display when the display 151 is turned on irrespective of an output change of the display 151. As an example, as depicted in
As depicted in the examples of
An operation of the mobile terminal 100 in case of receiving the first and the second user input is explained in detail with reference to following drawings.
As depicted in an example of
First of all, the menu items depicted in
The setting button 622 is used for adjusting various settings of the camera. Specifically, a user can adjust a capturing mode of the camera, resolution of a picture to be captured, and the like via the setting button 622. In this case, the capturing mode of the camera is used for calling a suitable setting value of the camera in accordance with a type of a subject. When the capturing mode is adjusted, the controller 180 can control a preset setting value of the camera to be applied. In this case, the setting value of the camera may indicate an aperture value, an exposure value, a shutter speed, an ISO speed, and the like of the camera. In general, the capturing mode of the camera can be set to a portrait mode (a mode for capturing a person), a landscape mode (a mode for capturing a landscape), a close-up mode (a mode for capturing an object positioned in a short distance), a night mode (a mode for capturing in environment of less light), a fast mode (a mode for capturing a subject moving fast), and the like, by which the present invention may be non-limited.
The camera 121 switching mode 623 is used for determining which camera is going to be used when two or more cameras are included in the mobile terminal 100. The mobile terminal 100 according to the present invention can take a picture using one of a front camera 121 and a rear camera 121′ only. Or, the mobile terminal 100 according to the present invention may take a picture using the front camera 121 and the rear camera 121′ at the same time.
The toggle button 624 may be used for selecting a type of data to be taken among a picture and a video. In case of capturing a picture, a momentary image is captured when a capturing command is inputted. On the contrary, in case of capturing a video, images can be contiguously captured from a point on which a capturing command is inputted to a point on which the capturing is completed.
The flash button 625 is used for controlling On, Off, and automatic adjustment of a flash. In case of using the flash, a picture can be efficiently captured with a less amount of light.
The thumbnail button 626 may display a thumbnail of a recently captured picture. If the thumbnail button 626 is touched, a list of pictures and videos captured by a user can be displayed by entering a gallery.
As depicted in
Subsequently, an operation of the mobile terminal 100 is explained in detail when a second user input is received.
a and
As depicted in an example of
As depicted in an example of
As depicted in
a to
If the camera 121 icon is long touched, the controller 180 can control a picture to be immediately captured without displaying a preview image of the camera. As depicted in examples of
As an example, as depicted in an example of
As depicted in
As a different example, as depicted in
When a picture is captured by a second user input, since a process of displaying the preview image is omitted, a user cannot have a chance to check an object to be captured. Hence, the user cannot manually focus on a preferred subject. As a result, the subject preferred by the user may be defocused. In this case, the controller 180 can automatically adjust a focus in accordance with a situation or based on a voice indication of the user in case of capturing a picture by the second user input.
a to
In case of capturing a picture by the second user input, the controller 180 can control a focus to be adjusted to a subject positioned in the middle. Specifically, as depicted in an example of
As a different example, the controller 180 can control a focus to be adjusted to a subject positioned in the nearest from the camera 121. For clarity, assume that a proximity extent of subjects located in a capturing direction is a person 910>a tree 920>a cat 930 in sequence. In case of capturing a picture by the second user input, as depicted in
As a different example, the controller 180 can control a focus to be automatically adjusted to a person in case of capturing a picture by the second user input.
The controller 180 may focus on an object indicated by a voice command of a user. As an example, if a second user input corresponds to an input of a voice command while the camera 121 button is touched, the controller 180 can control a picture to be captured in a manner of focusing on the object indicated by the voice command. As an example, as depicted in
If the user says a name of a person, the controller 180 can take a picture in a manner of focusing on the person designated by the user. As an example, as depicted in an example of
When a picture is captured by a first user input, various setting of the camera can be changed via the menu items provided with the preview image. Yet, when a picture is captured by a second user input, a chance for manipulating the menu items is not provided. Hence, a user may have a problem in that the picture is captured without changing the setting of the camera. Accordingly, the controller 180 can control the setting of the camera to be appropriately adjusted according to a situation when the picture is captured by the second user input. Regarding this, it shall be described in detail with reference to following drawings.
a and
As depicted in
As depicted in an example of
Although it is not depicted, in case of capturing a picture in an environment where the mobile terminal 100 moves fast, the controller 180 can control the camera setting to be adjusted to make the camera setting to be appropriate for capturing a subject that moves fast (e.g., a shutter speed is reduced to the minimum). In particular, the controller 180 can control the camera setting to be modified in accordance with a capturing mode (i.e., fast mode) suitable for capturing the subject that moves fast. The controller 180 may determine whether the mobile terminal 100 is moving fast based on a speed and the like measured using a signal detected by the sensing unit 140 (specifically, an acceleration sensor), GPS, and the like.
As mentioned in the foregoing description, the controller 180 can select a capturing mode appropriate for a subject to be captured or a capturing environment. In order to inform a user of a capturing mode of a picture, the controller 180 can control feedbacks different from each other to be outputted according to the capturing mode which is selected when the picture is captured. For instance, if a picture is captured in the portrait mode, the controller controls such a voice as ‘person’ to be outputted as a feedback. If a picture is captured in the fast mode, the controller controls such a voice as ‘fast’ to be outputted as a feedback. The controller 180 can control sound signals different from each other to be outputted in accordance with a capturing mode. Besides, the controller 180 may output a feedback in a form of vibration, LED flickering, or the like.
a and
The controller 180 touches the camera 121 icon and may be then able to determine a capturing mode of the camera based on a flicking direction of the pointer. As an example, as depicted in an example of
In particular, as depicted in the example of
As depicted in the examples of
If a preview image is displayed by inputting a first user input, a user can take a picture via menu items provided together with the preview image in a manner of selecting at least one of a front camera 121 and a rear camera 121′. Yet, when a picture is captured by a second user input, it is not clear which one is used for capturing the picture among the front camera 121 and the rear camera 121′ since a chance for manipulating the menu items is not provided. Accordingly, the controller 180 captures a picture selectively using one of the front camera 121 and the rear camera 121′ according to a default setting value. Or, the controller 180 may take a picture using the front camera 121 and the rear camera 121′ at the same time. In case of capturing a picture using the front camera 121 and the rear camera 121′ at the same time, the controller 180 may generate two picture files (a picture captured by the front camera 121 and a picture captured by the rear camera 121′) or may generate one picture file (merging the picture captured by the front camera 121 and the picture captured by the rear camera 121′ with each other)
The mobile terminal 100 according to the present invention can take a picture in a manner of appropriately using at least one of the front camera 121 and the rear camera 121′ in accordance with a situation when the picture is captured by a second user input. Regarding this, it shall be described in detail with reference to
a and
The controller 180 can select at least one of the front camera 121 and the rear camera 121′ based on whether the mobile terminal 100 is tilted. As an example, as depicted in an example of
The controller 180 may select at least one of the front camera 121 and the rear camera 121′ based on a state that the mobile terminal 100 is gripped. As an example, as depicted in an example of
As depicted in the examples of
A user may take a picture while watching a preview image after a first user input is inputted. Or, the user may take a picture without outputting the preview image due to a second user input. Capturing a picture by the second user input may be helpful in quickly taking a picture compared to capturing a picture by the first user input since a step of outputting the preview image is omitted. In particular, the user may operate the mobile terminal 100 to perform a prompt picture capturing by the second user input in case of capturing a fast moving subject or in an urgent situation. As mentioned in the foregoing description, the controller 180 can capture a plurality of pictures in case of capturing a picture by the second user input in a manner of considering that a picture capturing will be initiated by the second user input in the situation as urgent as omitting the preview image. The controller 180 can capture a plurality of the pictures as many as a predetermined number or may take a plurality of the pictures in a manner of continuously capturing pictures until a contact of a pointer is released from a prescribed icon. An output of the mobile terminal 100 in case of capturing a plurality of the pictures is described in detail with reference to
a and
If at least one of a plurality of the pictures is selected, the controller 180 can control remaining pictures except the selected picture to be automatically deleted. In particular, as depicted in an example of
The controller 180 can capture a plurality of pictures in a manner of varying a focal length. Regarding this, it shall be described in detail with reference to
In case of capturing a plurality of pictures in response to a second user input, as depicted in an example of
If the second user input is inputted, a picture is captured without displaying a preview image. Hence, the user cannot precisely expect the picture to be captured. Accordingly, if distortion for an output (i.e., a captured picture) of capturing a picture, which is captured by the second used input, is gross, the controller 180 can control a feedback configured to inform a user of the distortion of the captured picture to be outputted. Regarding this, it shall be described in detail with reference to
Although it is not depicted, if sharpness of the captured picture is less than a predetermined value due to a movement of the mobile terminal 100 or a movement of a subject, if brightness of the captured picture is less than a predetermined value since surrounding is too dark, or the like, the controller 180 can determine the captured picture is distorted.
If the captured picture is determined as distorted, the controller 180 can output a feedback configured to inform a user of the distortion of the captured picture. The feedback can be outputted in a form of sound data, vibration, LED flickering, or the like.
As a different example, if the captured picture is determined as distorted, the controller 180 may automatically take a picture one more time. In this case, the controller 180 may automatically adjust the camera 121 setting according to a reason for the distortion of the captured picture. For instance, if the sharpness of the captured picture is low, the controller can capture a picture again by reducing a shutter speed. If the brightness of the captured picture is low, the controller can capture a picture again by increasing the extent of opening the aperture or the sensitivity of the camera.
As mentioned in the foregoing description, the mobile terminal 100 according to the present invention may capture a picture after a preview image is displayed or may quickly capture a picture without displaying the preview image. In the following, various embodiments capable of being expanded by the mobile terminal 100 according to the present invention are described.
<In-App Capturing>
The mobile terminal 100 according to the present invention can output a preview image while an application except a camera 121 application is executing or without displaying the preview image. Regarding this, it shall be described in detail with reference to
In the middle of executing an application, a user can display a preview image by inputting a first user input or a second user input or the user can take a picture without displaying the preview image. In this case, as depicted in an example of
As depicted in an example of
Unlike the aforementioned description, if a second input is inputted (e.g., a long touch on the camera 121 icon 1610 of the virtual keypad), the controller 180 immediately captures a picture without displaying the preview image and may control a captured picture 1630 to be automatically inserted into a document in editing as depicted in an example of
If a picture is captured (for instance, the picture is captured via the preview image or the picture is directly captured by the second user input without displaying the preview image) while the mobile terminal 100 is executing a messenger application (e.g., Kakao Talk, WhatsApp, etc.), the controller 180 can control the captured picture to be immediately transmitted to a conversation counterpart. If a picture is captured while the mobile terminal 100 is executing a message application or an e-mail application, the controller 180 can control the captured picture to be attached to a message or an e-mail in the course of writing the message or the e-mail.
As mentioned in the foregoing description, if a picture is captured while an application is executing, the controller 180 can control the captured picture to be applied to the currently executing application.
If the mobile terminal 100 is executing a plurality of applications at the same time, the controller 180 may control the captured picture to be applied to a single application only among a plurality of the applications. Regarding this, it shall be described in detail with reference to
As depicted in an example of
Unlike the aforementioned description, if a second user input is inputted, the control unit 180 may capture a picture without displaying a preview image. As depicted in an example of
As an example, as depicted in an example of
As mentioned in the foregoing description, if a picture is captured while an application is executing, the controller 180 can control the captured picture to be immediately applied to the currently executing application. By doing so, a user can easily apply the captured picture to the application without performing a separate loading process.
<A Position in Which a Captured Picture is Displayed>
As mentioned earlier in
As mentioned in the foregoing description, the position in which the captured picture is displayed may depend on the position in which the preview image is displayed. A user may preferentially determine a position in which a picture to be captured is inserted by adjusting the position of the preview image before the picture is captured.
<Configuring a Captured Picture as a Hyperlink Object>
Having selected a prescribed data by a user, if a picture is captured by a first user input or a second user input, the controller 180 can control the captured picture to be configured as an object linked to the selected prescribed data. Regarding this, it shall be described in detail with reference to
As an example, as depicted in an example of
As depicted in the examples of
<Liaison Between a Captured Picture and Schedule>
If a picture is captured via a preview image or a second user input while a scheduling application is executing, the controller 180 can control the captured picture to be directly inserted into a schedule. In this case, the controller 180 may automatically determine a date to which a schedule is added via a text recognition of the captured picture. Regarding this, it shall be described in detail with reference to
If a user long touches a specific date on the calendar, as depicted in an example of
If a first user input (e.g., a short touch on the ‘picture add’ item 2014) is inputted on the menu item, the controller 180 can control a preview image 2020, which is used for capturing a picture to be added to the selected date, to be outputted as depicted in an example of
If a picture capturing command is inputted via the preview image 2020 or if a picture is captured by receiving a second user input, the controller 180 can control a captured picture 2030 to be added to the specific date selected by the user as depicted in an example of
a) to (c) show an example that the captured picture can be added to the selected specific date in case of capturing the picture after the user selects the specific date. On the contrary, the controller 180 captures a picture and may be then able to control the captured picture to be added to the specific date in a manner of waiting for a selection of the specific date selected by the user.
Subsequently, an example that a date to which a picture to be added is automatically determined based on a text extracted from the captured picture is explained in the following.
As depicted in an example of
If a picture is captured via the preview image 2120 or the second user input, the controller 180 extracts a text from the captured picture 2130 and may be then able to determine whether the extracted text indicates a prescribed date. As an example, as depicted in an example of
As depicted in
<Focusing on Data Selected by User>
If a user inputs a first user input or a second user input after selecting a prescribed data, the controller 180 can display a preview image or take a picture in a manner of focusing on an object capable of being indicated by the prescribed data selected by the user. Regarding this, it shall be described in detail with reference to
A user can select data for indicating an object, which is intended to be focused, before inputting a first or a second user input. Subsequently, if a preview image or a picture is captured by the first user input or the second user input, the controller 180 can control a focus to be adjusted to the object indicated by the data which is selected by the user. As an example, if the first user input (e.g., a short touch on an assistive object 2210) is inputted while such a phrase as ‘jogging person’ 2220 is selected, as depicted in an example of
If the data selected by the user corresponds to a name of a specific person, the controller 180 may display a preview image where a focus is adjusted to the person indicated by the selected data or may control a picture to be captured in a manner of focusing on the person indicated by the selected data. As an example, as depicted in an example of
If the data selected by the user indicates a plurality of objects, the controller 180 may control a plurality of pictures to be captured in a manner of focusing on each of a plurality of the objects. Regarding this, it shall be described in detail with reference to
If a first user input or a second user input is inputted, the controller 180 can determine an object indicated by the data, which is selected by the user, in a manner of analyzing the data selected by the user. As an example, in
If the first user input is inputted while the data selected by the user is indicating a plurality of objects, the controller 180 displays a preview image 2420 and may be then able to control a focus to be adjusted to a randomly selected object among a plurality of the objects as depicted in an example of
If the second user input is inputted while the data selected by the user is indicating a plurality of objects, the controller 180 may control a plurality of pictures to be captured in a manner of focusing on each of a plurality of the indicated objects without displaying the preview image. As an example, as depicted in an example of FIG. 24(c), the controller 180 can capture a picture 2430 where a focus is adjusted to ‘John’ indicated by the data which is selected by the user and a picture 2440 where a focus is adjusted to ‘Jane’.
In the course of editing a document, a user can select an object on which a focus will be adjusted in a manner of inputting a text and selecting the inputted text by the user. If a picture where a focus is adjusted to the data selected by the user is captured in the middle of editing the text, as mentioned earlier in
As mentioned in the foregoing description with reference to the examples of
<Size Adjustment of a Captured Picture>
When a captured picture is applied to an application, the controller 180 can adjust a size of the captured picture to make the size of the captured picture appropriate for an attribute of the application. As an example, in case of attaching the captured picture to a text message (e.g., MMS), the controller 180 resizes the captured picture less than a first size (e.g., 700 kb) and can control the resized picture to be attached to the text message. As a different example, in case of attaching the captured picture to an e-mail, the controller 180 resizes the captured picture less than a second size (e.g., 1 Mb) and can control the resized picture to be attached to the e-mail.
In particular, the controller 180 can control the size of the captured picture to be automatically adjusted in accordance with the characteristic of the application to which the captured picture is applied.
<Capturing a Picture Using a Front Camera 121 and a Rear Camera 121′ at the Same Time>
The mobile terminal 100 according to the present invention can take a picture selectively using either the front camera 121 or the rear camera 121′. Or, the mobile terminal can take a picture simultaneously using the front camera 121 and the rear camera 121′. In case of capturing a picture using the front camera 121 and the rear camera 121′ at the same time, a preview image or a captured picture of the front camera 121 (or rear camera 121′) can be provided as a sub image of a preview image or a captured picture of the rear camera 121′ (or front camera 121) or both images can be provided independently. Regarding this, it shall be described in detail with reference to
If a first user input is received in a state that both the front camera 121 and the rear camera 121′ are configured to use, the controller 180 can control a preview image of the front camera and a preview image of the rear camera to be displayed at the same time. In this case, as depicted in an example of
As a different example, as depicted in an example of
As depicted in the example of
In this case, the pictures 2610/2620 captured by the front camera 121 and the rear camera 121′ can be merged into a single file or both pictures can be regarded as separate files. If the both pictures are merged into a single file, the controller 180 can generate a merged file in a manner of configuring one of the both picture 2610/2620 as a main image 2610 and configuring the other picture as a sub image 2620 as depicted in an example of
<Capturing a Picture with an Output of a Display 151 in the Background>
In case of capturing a picture, the mobile terminal 100 according to the present invention can take a picture with an output of the display 151 in the background. Regarding this, it shall be described in detail with reference to
As depicted in
<Editing a Captured Picture>
The mobile terminal 100 according to the present invention can provide an editing screen capable of editing a picture which is captured via a first user input or a second user input. Specifically, the controller 180 can provide a user interface (UI) for determining a picture frame of the captured picture, a UI for attaching a sticker (i.e., clip art) to an editing item, and the like via the editing screen.
As depicted in an example of
<Switching to a Panorama Mode>
If a user input for expanding a capturing area is received from a user while a preview image is displaying, the controller 180 can control a capturing mode of the camera to be switched to a panorama capturing mode. On the contrary, if a user input for reducing the capturing area is received from the user when the capturing mode of the mobile terminal 100 is set to the panorama capturing mode, the controller 180 can control the capturing mode of the camera to be switched from the panorama capturing mode to a general capturing mode.
Although it is not depicted, the user input for expanding the capturing area may correspond to a pinch out input for the preview image and the user input for reducing the capturing area may correspond to a pinch in input for the preview image.
Accordingly, embodiments of the present invention provide various effects and/or features.
According to one of the embodiments, the present invention can provide a mobile terminal for enhancing user convenience.
Specifically, the present invention can provide a mobile terminal capable of quickly capturing a picture without displaying a preview image.
Moreover, the present invention can provide a mobile terminal capable of automatically adjusting a focus and controlling a camera setting to correspond to intention of a user.
It will be appreciated by those skilled in the art that the present invention can be specified into other form(s) without departing from the spirit or scope of the inventions.
In addition, the above-described methods can be implemented in a program recorded medium as processor-readable codes. The processor-readable media may include all kinds of recording devices in which data readable by a processor are stored. The processor-readable media may include ROM, RAM, CD-ROM, magnetic tapes, floppy discs, optical data storage devices, and the like for example and also include carrier-wave type implementations (e.g., transmission via Internet).
It will be appreciated by those skilled in the art that various modifications and variations can be made in the present invention without departing from the spirit or scope of the inventions. Thus, it is intended that the present invention covers the modifications and variations of this invention provided they come within the scope of the appended claims and their equivalents.
Number | Date | Country | Kind |
---|---|---|---|
10-2013-0095053 | Aug 2013 | KR | national |