The present technology relates to an information processing apparatus and an imaging apparatus. In detail, it relates to an information processing apparatus, an imaging apparatus, and an imaging system which perform a control related to an imaging operation, control methods of these, and programs for causing these methods to be executed by a computer.
In related art, imaging apparatuses such as digital still cameras and digital video cameras (for example, a camera-integrated recorder) are widespread, which generate an image (image data) by capturing a photographic subject, and record this image as a content. Further, wireless communication technology exists where an exchange of various types of data is performed by using wireless communication.
Further, technology exists where an imaging apparatus is operated by another apparatus by using wireless communication. For example, an electronic device has been proposed, when a contact operation is detected for an operation switch image displayed in a display unit, for causing an operation, which corresponds to the operation switch image for which this contact operation is performed, to be executed in an imaging apparatus (for example, refer to Patent Literature 1).
In the above described related art, since an imaging apparatus can be operated by an electronic device by using wireless communication, the imaging apparatus can be operated by using the electronic device, for example, even in the case where the imaging apparatus and the electronic device are separated.
Here, for example, it is assumed that an imaging operation is performed, in an information processing apparatus capable of operating an imaging apparatus by using wireless communication, by attaching this imaging apparatus. In this case, it is assumed that the use states of the imaging apparatus and the information processing apparatus will be different compared to the case where performing an imaging operation by setting to a location separated from the imaging apparatus. Accordingly, in the case where an imaging operation is performed by using an imaging apparatus and an information processing apparatus, it will be important to provide an appropriate user interface in accordance with these use states.
The present technology is created by considering such a situation, and an object of the present technology is to provide an appropriate user interface corresponding to a use state of an apparatus.
The present technology is created to solve the above-described problem. A first aspect of the present technology is an information processing apparatus, a control method thereof, and a program for causing a computer to execute the control method, the information processing apparatus including: a control unit which performs a control for switching a display state of a display screen for operating an imaging apparatus based on a relative position relationship with the imaging apparatus. In this way, an action is brought about for switching a display state of a display screen for operating an imaging apparatus based on a relative position relationship with the imaging apparatus.
According to the first aspect, the control unit may perform a control for switching a display state of the display screen based on a distance between the information processing apparatus and the imaging apparatus. In this way, an action is brought about for switching a display state of a display screen based on a distance between an information processing apparatus and an imaging apparatus.
According to the first aspect, the control unit may perform a control for switching a display state of the display screen based on whether or not the imaging apparatus is mounted on the information processing apparatus. In this way, an action is brought about for switching a display state of a display screen based on whether or not an imaging apparatus is mounted on an information processing apparatus.
According to the first aspect, the control unit may perform a control for switching a display state of the display screen based on whether or not the imaging apparatus is mounted on a display surface of the information processing apparatus. In this way, an action is brought about where a control for switching a display state of a display screen is performed based on whether or not an imaging apparatus is mounted on a display surface of an information processing apparatus.
According to the first aspect, in the case where the imaging apparatus is mounted on a display surface of the information processing apparatus, the control unit may perform a control for switching a display state of the display screen based on a position of the imaging apparatus on a display surface of the information processing apparatus. In this way, an action is brought about, in the case where an imaging apparatus is mounted on a display surface of an information processing apparatus, for switching a display state of a display screen based on a position of the imaging apparatus on the display surface of the information processing apparatus.
According to the first aspect, the control unit may cause the display screen which includes an operation object for operating the imaging apparatus to be displayed, and may perform a control for changing a display state of the operation object based on the relative position relationship. In this way, an action is brought about which causes a display screen which includes an operation object for operating an imaging apparatus to be displayed, and changes a display state of the operation object based on a relative position relationship.
According to the first aspect, the control unit may cause the display screen which includes an operation object for operating the imaging apparatus to be displayed, and may perform a control, in the case where the imaging apparatus is not mounted on the information processing apparatus, for changing a display state of the operation object based on a change in a posture of the information processing apparatus, and in the case where the imaging apparatus is mounted on the information processing apparatus, for not changing a display state of the operation object based on a change in a posture of the information processing apparatus. In this way, an action is brought about, in the case where an imaging apparatus is not mounted on an information processing apparatus, which changes a display state of an operation object based on a change in a posture of the information processing apparatus, and in the case where an imaging apparatus is mounted on an information processing apparatus, which does not change a display state of an operation object based on a change in a posture of the information processing apparatus.
A second aspect of the present technology is an imaging apparatus, a control method thereof, and a program for causing a computer to execute the control method, the imaging apparatus including: a control unit which performs a control related to an imaging operation based on an operation input performed in the information processing apparatus in which a display screen is displayed for a display state to be switched based on a relative position relationship of the imaging apparatus and the information processing apparatus. In this way, an action is brought about where a control related to an imaging operation is performed based on an operation input performed in an information processing apparatus in which a display screen is displayed for a display state to be switched based on a relative position relationship of an imaging apparatus and the information processing apparatus.
According to the second aspect, a display state of the display screen may be switched based on a distance between the information processing apparatus and the imaging apparatus. In this way, an action is brought about for a display state of a display screen to be switched based on a distance between an information processing apparatus and an imaging apparatus.
According to the second aspect, a display state of the display screen may be switched based on whether or not the imaging apparatus is mounted on the information processing apparatus. In this way, an action is brought about for a display state of a display screen to be switched based on whether or not an imaging apparatus is mounted on an information processing apparatus.
According to the second aspect, a display state of the display screen may be switched based on whether or not the imaging apparatus is mounted on a display surface of the information processing apparatus. In this way, an action is brought about for a display state of a display screen to be switched based on whether or not an imaging apparatus is mounted on a display surface of an information processing apparatus.
According to the second aspect, in the case where the imaging apparatus is mounted on a display surface of the information processing apparatus, a display state of the display screen may be switched based on a position of the imaging apparatus on a display surface of the information processing apparatus. In this way, an action is brought about, in the case where an imaging apparatus is mounted on a display surface of an information processing apparatus, for a display state of a display screen to be switched based on a position of the imaging apparatus on the display surface of the information processing apparatus.
According to the second aspect, the information processing apparatus may cause the display screen which includes an operation object for operating the imaging apparatus to be displayed, and may change a display state of the operation object based on the relative position relationship. In this way, an action is brought about where an information processing apparatus causes a display screen which includes an operation object for operating an imaging apparatus to be displayed, and changes a display state of the operation object based on a relative position relationship.
According to the second aspect, the information processing apparatus may cause the display screen which includes an operation object for operating the imaging apparatus to be displayed, and in the case where the imaging apparatus is not mounted on the information processing apparatus, may change a display state of the operation object based on a change in a posture of the information processing apparatus, and in the case where the imaging apparatus is mounted on the information processing apparatus, does not have to change a display state of the operation object based on a change in a posture of the information processing apparatus. In this way, an action is brought about where an information processing apparatus, in the case where an imaging apparatus is not mounted on the information processing apparatus, changes a display state of an operation object based on a change in the posture of the information processing apparatus, and in the case where an imaging apparatus is mounted on the information processing apparatus, does not change a display state of an operation object based on a change in the posture of the information processing apparatus.
A third aspect of the present technology is an imaging system, a control method thereof, and a program for causing a computer to execute the control method, the imaging system including: an imaging apparatus in which a control related to an imaging operation is performed based on an operation input performed in an information processing apparatus by connecting to the information processing apparatus by using wireless communication; and an information processing apparatus which performs a control for switching a display state of a display screen for operating the imaging apparatus based on a relative position relationship with the imaging apparatus. In this way, an action is brought about where an imaging apparatus has a control related to an imaging operation performed based on an operation input performed in an information processing apparatus by connecting to the information processing apparatus by using wireless communication, and the information processing apparatus switches a display state of a display screen for operating the imaging apparatus based on a relative position relationship with the imaging apparatus.
The present technology can accomplish a superior effect such as being able to provide an appropriate user interface corresponding to a use state of an apparatus. Note that, the effect described here is not necessarily limited, and may be any effect described within the present disclosure.
Hereinafter, the modes for executing the present technology (hereinafter, called the embodiments) will be described. Note that, the description will be given in the following order.
1. First embodiment (example where a display state of a display screen is switched based on a distance between an imaging apparatus and an information processing apparatus) 2. Second embodiment (example where a display state of a display screen is switched based on an attachment position of an imaging apparatus) 3. Third embodiment (example where switching of a display state is controlled based on a change in the posture of an information processing apparatus, based on the presence or absence of mounting of an imaging apparatus and an information processing apparatus)
Further, in the embodiments of the present technology, an example will be shown where the imaging apparatus 100 is set to a cylindrical (columnar shaped) imaging apparatus. That is, an example will be shown where the shape of the imaging apparatus 100 is set to a shape so that only a lens portion of a general imaging apparatus (for example, an integrated camera) is taken out. Note that, while the imaging apparatus 100 includes operation members such as a zoom lever and a shutter button, these illustrations are omitted in
The imaging apparatus 100 includes a lens barrel 160, and attachment members 171 and 172. The lens barrel 160 accommodates each of the members of an optical system, an imaging system or the like.
The attachment members 171 and 172 are attachments used at the time when attaching the imaging apparatus 100 to another apparatus (for example, an information processing apparatus 200 shown in
In this way, for example, the imaging apparatus 100 is capable of performing a usual imaging operation, and can be used by mounting on another apparatus (for example, a smartphone). Further, in the case where used by mounting on another apparatus, the imaging apparatus 100 can be operated by a remote operation using the other apparatus.
Each of
An example is shown in a of
The information processing apparatus 200 includes operation members 221 through to 223, an input-output unit 240, a voice output unit 280, a light emitting unit 291, and an imaging unit 292. Note that, the information processing apparatus 200 is implemented, for example, by an information processing apparatus such as a smartphone or a tablet terminal.
The operation members 221 through to 223 are operation members used at the time when performing various types of operation inputs.
The input-output unit 240 displays various types of images, and receives an operation input from a user based on a detection state of an object close to or in contact with the display surface of the input-output unit 240.
The voice output unit 280 outputs various types of voice information.
The light emitting unit 291 is a light emitting apparatus for causing light to be emitted to a photographic subject. The light emitting unit 291 is used, for example, at the time when performing an imaging operation by using the information processing apparatus 200 in an environment where sufficient brightness is not able to be expected such as at night or within a room.
The imaging unit 292 generates an image (image data) by capturing a photographic subject.
As shown in
The imaging apparatus 100 includes an imaging unit 110, an image processing unit 120, a storage unit 130, a control unit 140, and a wireless communication unit 150.
The imaging unit 110 generates an image (image data) by capturing a photographic subject, and outputs the generated image to the image processing unit 120. The imaging unit 110 is constituted, for example, by an optical system (a plurality of lens), and an imaging element. Further, the imaging unit 110 has each unit (for example, a zoom lens, a focus lens, a diaphragm) controlled, based on a control of the control unit 140. Note that, in the embodiments of the present technology, in the case where described with “image”, it includes both meanings of this image, and image data for displaying this image.
The image processing unit 120 applies a prescribed image process (for example, a de-mosaic process) for an image output from the imaging unit 110, based on a control of the control unit 140, and causes the image to which this image process has been applied to be stored in the storage unit 130. Note that, an image to which an image process has been applied by the image processing unit 120 may be stored in the storage unit 270 by transmitting to the information processing apparatus 200 by using wireless communication.
The storage unit 130 is a recording medium which stores an image to which an image process has been applied by the image processing unit 120 as a content (for example, a still image file, a moving image file). Note that, the storage unit 130 may be built into the imaging apparatus 100, or may be detachable from the imaging apparatus 100.
The control unit 140 controls each of the units in the imaging apparatus 100 based on a control program. For example, the control unit 140 controls each of the units based on an operation input received by operation members such as a zoom lever and a shutter button (not illustrated) included in the imaging apparatus 100. Further, the control unit 140 controls each of the units based on control information from the information processing apparatus 200 received via the wireless communication unit 150. That is, the imaging apparatus 100 can be remotely operated by using the information processing apparatus 200.
The wireless communication unit 150 performs transmission and reception of each information (for example, control data, image data) with other information processing apparatuses (for example, the information processing apparatus 200), by using wireless communication.
Here, a wireless Local Area Network (LAN) can be used, for example, as wireless communication. Wireless Fidelity (Wi-Fi) can be used, for example, as this wireless LAN. Further, wireless communication such as Near Field Communication (NFC), Bluetooth (registered trademark), infrared rays, or portable electric waves can be used, for example, as wireless communication.
Further, a plurality of wireless communication systems may be used. For example, at the time of the start of wireless communication, an exchange of data (for example, a Service Set Identifier (SSID)) related to Wi-Fi when turning on a power source is performed by NFC. Then, the exchange of data from here onwards can be performed by Wi-Fi.
The information processing apparatus 200 includes a posture detection unit 210, an operation reception unit 220, a wireless communication unit 230, an input-output unit 240, a control unit 250, an image processing unit 260, a storage unit 270, and a voice output unit 280.
The posture detection unit 210 detects a change in the posture of the information processing apparatus 200 by detecting an acceleration, movement, orientation or the like of the information processing apparatus 200, and outputs posture information related to the detected change in posture to the control unit 250. Note that, various types of sensors such as a gyro sensor or an acceleration sensor can be used, for example, as the posture detection unit 210.
The operation reception unit 220 is an operation reception unit which receives an operation performed by a user, and outputs control information (operation information) corresponding to the received operation content to the control unit 250. Note that, the operation reception unit 220 corresponds, for example, to the operation members 221 through to 223 shown in
The wireless communication unit 230 performs transmission and reception of each information (for example, control data, image data) with other information processing apparatuses (for example, the imaging apparatus 100) by wireless communication, based on a control of the control unit 250. The above described wireless LAN (for example, Wi-Fi), NFC, Bluetooth, infrared rays, or portable electric waves can be used, for example, as wireless communication. Further, a plurality of wireless communication systems may be used.
The input-output unit 240 has an input unit 241 and a display unit 242 constituted as one body. Further, the input-output unit 240 displays various types of images in a display unit 242 based on a control of the control unit 250, and receives an operation input from a user by the input unit 241 based on a detection state of an object close to or in contact with the display surface of the display unit 242. Further, the input unit 241 outputs control information corresponding to the received operation input to the control unit 250.
For example, an electrostatic type (electrostatic capacitance system) touch panel, which detects contact or closeness of an object having conductivity (for example, a person's finger) based on a change in electrostatic capacitance, can be used as the input unit 241. Further, for example, a display panel such as a Liquid Crystal Display (LCD) or an organic Electro Luminescence (EL) panel can be used as the display unit 242. Also, the input-output unit 240 is constituted, for example, by overlapping a transparent touch panel on the display surface of a display panel.
For example, an operation of the information processing apparatus 200 or the imaging apparatus 100 becomes possible by having a user perform a contact operation (or, a closeness operation) of an operation object or the like displayed in the display unit 242. Here, for example, operation objects are displayed in the input-output unit 240, such as operation objects 301 through to 306 shown in b of
The control unit 250 controls each of the units in the information processing apparatus 200 based on a control program. For example, the control unit 250 decides the posture of the information processing apparatus 200, based on posture information from the posture detection unit 210, and switches a display state of a display screen for displaying in the input-output unit 240, based on this decision result. For example, the control unit 250 decides the up-down direction of the posture of the information processing apparatus 200, based on posture information from the posture detection unit 210, and switches the up-down direction of a display screen for displaying in the input-output unit 240, based on this decision result.
Further, for example, in the case where the information processing apparatus 200 and the imaging apparatus 100 are connected via the wireless communication unit 230, the control unit 250 causes a display screen for operating the imaging apparatus 100 to be displayed in the input-output unit 240. In this case, the control unit 250 performs a control for switching a display state of a display screen for operating the imaging apparatus 100, based on a relative position relationship with the imaging apparatus 100. For example, the control unit 250 performs a control for switching a display state of a display screen based on a distance between the information processing apparatus 200 and the imaging apparatus 100. In this case, for example, the control unit 250 causes a display screen which includes operation objects for operating the imaging apparatus 100 (for example, the operation objects 301 through to 306 shown in
Further, for example, the control unit 250 performs a control for switching a display state of a display screen based on whether or not the imaging apparatus 100 is mounted on the information processing apparatus 200. In this case, the control unit 250 performs a control for switching a display state of a display screen based on whether or not the imaging apparatus 100 is mounted on the display surface of the input-output unit 240. Further, in the case where the imaging apparatus 100 is mounted on the display surface of the input-output unit 240, the control unit 250 performs a control for switching a display state of a display screen based on the position of the imaging apparatus 100 on the display surface of the input-output unit 240. Note that, the control for switching a display state of a display screen based on whether or not the imaging apparatus 100 is mounted on the display surface of the input-output unit 240 will be shown in a second embodiment of the present technology.
In this way, in the information processing apparatus 200, a display screen is switched based on a relative position relationship of the imaging apparatus 100 and the information processing apparatus 200 connected by using wireless communication. Further, the control unit 140 of the imaging apparatus 100 performs a control related to an imaging operation based on an operation input using a display screen performed in the information processing apparatus 200 in which this display screen is displayed.
The image processing unit 260 applies a prescribed image process for an image generated by the imaging apparatus 100 or an image generated by the imaging unit 292 (shown in b of
The storage unit 270 is a recording medium which stores each information, based on a control of the control unit 250. For example, an image generated by the imaging apparatus 100 or an image generated by the imaging unit 292 (shown in b of
The voice output unit 280 outputs voice information, based on a control of the control unit 250. The voice output unit 280 can be implemented, for example, by a speaker.
An example is shown in a of
An example is shown in b of
Here, it is assumed that the use state of the information processing apparatus 200 (for example, the way to hold) is different, in the case where the imaging apparatus 100 is mounted on the information processing apparatus 200, and in cases other than this. Accordingly, in a first embodiment of the present technology, an example is shown where a display state of a display screen is switched based on a distance between the imaging apparatus 100 and the information processing apparatus 200. Further, in a second embodiment of the present technology, an example is shown where a display state of a display screen is switched based on an attachment position of the imaging apparatus 100.
Here, a method will be described which acquires a distance between the imaging apparatus 100 and the information processing apparatus 200.
For example, a distance between the imaging apparatus 100 and the information processing apparatus 200 can be estimated by using a received electric wave intensity. For example, a table showing a relationship between a received electric wave intensity and a distance is created beforehand, and stored in the storage unit 270. Then, the control unit 250 of the information processing apparatus 200 can acquire, from this table, a distance corresponding to a received electric wave intensity acquired by the wireless communication unit 230, and can use this distance as a distance between the imaging apparatus 100 and the information processing apparatus 200.
For example, a distance between the imaging apparatus 100 and the information processing apparatus 200 can be estimated by using the input-output unit 240. Here, a case is assumed where a touch panel of a projection-type electrostatic capacitance system is used, as the input-output unit 240. In this case, a material which reacts with a touch panel of a projection-type electrostatic capacitance system is adopted on a mounting surface 175 (such as shown in b of
For example, as shown in a of
Then, in the case where the size of this detected object and the size stored in the storage unit 270 match or approximately match, the control unit 250 can decide that the imaging apparatus 100 has been mounted on the display surface of the input-output unit 240. On the other hand, in the case where the size of this detected object and the size stored in the storage unit 270 do not match, or do not approximately match, the control unit 250 can decide that the imaging apparatus 100 is not mounted on the display surface of the input-output unit 240.
Further, the input-output unit 240 can detect the position of the imaging apparatus 100 on the display surface. Accordingly, the control unit 250 can switch a display state of a display screen, based on the position of the imaging apparatus 100 on the display surface. This switching example of the display screen will be shown in a second embodiment of the present technology.
[Example where a Member for Detecting the Mounting is Used]
For example, the mounting of the imaging apparatus 100 can be detected by using a member for detecting that the imaging apparatus 100 has been mounted on the information processing apparatus 200. A switch can be installed in at least one of the imaging apparatus 100 and the information processing apparatus 200, for example, as this member.
For example, it is assumed to be a case where a switch is installed in the imaging apparatus 100. In this case, at the time when the imaging apparatus 100 has been mounted on the information processing apparatus 200, this mounting is detected by this switch, and information showing that this mounting has been detected is output to the control unit 140. Then, in the case where the information showing that this mounting has been detected is acquired, the control unit 140 transmits this to the information processing apparatus 200 via the wireless communication unit 150. In this way, the control unit 250 of the information processing apparatus 200 can detect that the imaging apparatus 100 has been mounted on the information processing apparatus 200.
Further, for example, it is assumed to be a case where a switch is installed in the information processing apparatus 200. In this case, at the time when the imaging apparatus 100 has been mounted on the information processing apparatus 200, this mounting is detected by this switch, and information showing that this mounting has been detected is output to the control unit 250. In this way, the control unit 250 of the information processing apparatus 200 can detect that the imaging apparatus 100 has been mounted on the information processing apparatus 200.
Further, a distance between the imaging apparatus 100 and the information processing apparatus 200 may be detected, by using another sensor. For example, a distance sensor (for example, a sensor which detects a distance by using infrared rays or ultrasonic waves) can be used. Further, for example, a distance between the imaging apparatus 100 and the information processing apparatus 200 may be detected by using a Global Positioning System (GPS). For example, the respective positions of the imaging apparatus 100 and the information processing apparatus 200 can be acquired by using GPS, and a distance between the imaging apparatus 100 and the information processing apparatus 200 can be calculated based on these positions.
Note that, the above described detection method of a distance and detection method of the mounting are examples, and are not limited to these, and other detection methods may be used.
For example, in the case where a distance between the imaging apparatus 100 and the information processing apparatus 200 becomes a threshold or more, it is assumed that a frequent setting change while photographing is comparatively small. Accordingly, it is preferable for operation objects displayed in the input-output unit 240 (for example, an operation key) to be made simple by limiting to a minimum. For example, as shown in
Here, for example, a portion easy for a user to operate means the region of the right side in the input-output unit 240, in the case of a right-handed user, and means the region of the left side in the input-output unit 240, in the case of a left-handed user.
Further, for example, in the case where a distance between the imaging apparatus 100 and the information processing apparatus 200 becomes less than a threshold (for example, in the case where the imaging apparatus 100 is mounted on the information processing apparatus 200), it is assumed that a setting change while photographing is comparatively large. In particular, in the case where the imaging apparatus 100 is mounted on the information processing apparatus 200, it is assumed that a user supports the lens barrel 160 of the imaging apparatus 100 with his or her left hand, or performs a ring operation or a zoom operation, and performs a body operation with his or her right hand. Accordingly, it is preferable to arrange operation objects such as various types of setting buttons in the input-output unit 240. For example, as shown in
Note that, these display switching examples are examples, and may be set to display switching other than these. Hereinafter, these display switching examples will be shown.
Display examples are shown in
An example is shown in
Further, examples are shown in
An example is shown in
An example is shown in
Note that, the operation object 301 is an operation object pressed at the time when performing an on-off operation of the light emitting unit 291 of the information processing apparatus 200. Note that, the operation object 302 is an operation object pressed at the time when switching a photography mode of the imaging apparatus 100. Further, the operation object 303 is an operation object pressed at the time when performing a shutter operation of the imaging apparatus 100. Further, the operation object 304 is an operation object pressed at the time when performing various types of setting operations of the imaging apparatus 100. Further, the operation object 305 is an operation object pressed at the time when switching an exposure mode of the imaging apparatus 100. Further, the operation object 306 is an operation object for displaying scene information set in the imaging apparatus 100.
For example, in the case where a distance between the imaging apparatus 100 and the information processing apparatus 200 becomes a threshold or more, it is assumed that light from the light emitting unit 291 of the information processing apparatus 200 does not reach a photographic subject which becomes an imaging target of the imaging apparatus 100. Accordingly, it is assumed that light output from the light emitting unit 291 of the information processing apparatus 200 is hardly used as photography auxiliary light. Accordingly, as shown in a of
Further, for example, in the case where a distance between the imaging apparatus 100 and the information processing apparatus 200 becomes less than a threshold, it is assumed that light from the light emitting unit 291 of the information processing apparatus 200 reaches a photographic subject which becomes an imaging target of the imaging apparatus 100. Accordingly, since light output from the light emitting unit 291 of the information processing apparatus 200 can be used as photography auxiliary light, it is assumed that the light emitting unit 291 of the information processing apparatus 200 is used. Accordingly, as shown in b of
An example is shown in
An example is shown in
Note that, the operation objects 311 through to 313 correspond to the operation objects 301 through to 303 shown in
As described above, an operation member (zoom lever) for performing a zoom operation is included in the main body of the imaging apparatus 100. Therefore, for example, in the case where a distance between the imaging apparatus 100 and the information processing apparatus 200 becomes a threshold or more, it is assumed that a user does not reach the operation member (zoom lever) included in the main body of the imaging apparatus 100. In this case, a user is not able to operate the operation member (zoom lever) included in the main body of the imaging apparatus 100. Accordingly, as shown in a of
Further, for example, in the case where a distance between the imaging apparatus 100 and the information processing apparatus 200 becomes less than a threshold, it is assumed that a user reaches the operation member (zoom lever) included in the main body of the imaging apparatus 100. In this case, a user can operate the operation member (zoom lever) included in the main body of the imaging apparatus 100. Accordingly, as shown in b of
An example is shown in
a of
For example, the operation object 321 is an operation object pressed at the time when performing a shutter operation of the imaging apparatus 100. Further, the operation object 322 is an operation object used at the time when performing a trace operation. The operation object 322 corresponds, for example, to a mode dial, and can perform various types of operations, by performing an operation (trace operation) which moves the operation object 322 in a left-right direction, such as a mode dial.
Further, the operation object 323 is an operation object pressed at the time when performing an operation which starts moving image recording of the imaging apparatus 100. The operation object 323 corresponds, for example, to a moving image REC button. Further, the operation object 324 is an operation object pressed at the time when displaying a menu screen for performing various types of settings related to the imaging apparatus 100. The operation object 324 corresponds, for example, to a menu button.
Further, the operation object 325 is an operation object used at the time when performing a trace operation or a determination operation. The operation object 325 corresponds, for example, to a control wheel, and can perform a movement operation of a cursor displayed in the input-output unit 240 or various types of determination operations. Further, the operation object 326 is an operation object pressed at the time when performing reproduction of an image.
Further, a live view image, various types of operation objects (for example, icons) or the like are displayed in a display region 327 surrounded by a rectangular dotted line. Note that, an example is shown in b of
Here, for example, in the case where a distance between the imaging apparatus 100 and the information processing apparatus 200 becomes less than a threshold (in the case where the imaging apparatus 100 is mounted on the information processing apparatus 200), it is preferable to use the imaging apparatus 100 and the information processing apparatus 200 as a general imaging apparatus. In particular, in the case where the imaging apparatus 100 is mounted on the information processing apparatus 200, it is assumed that a user expects an operativeness to an extent that there is no sense of discomfort compared with a general imaging apparatus (for example, an integrated camera). Accordingly, it is preferable to perform an arrangement of operation members to an extent approximately the same as that of a general imaging apparatus (for example, an integrated camera). Accordingly, an example is shown in b of
First, the control unit 250 detects a distance between the imaging apparatus 100 and the information processing apparatus 200 (step S901). This detection method of a distance can use any of the above described methods.
To continue, the control unit 250 compares the distance between the imaging apparatus 100 and the information processing apparatus 200, and a threshold (step S902). Several cm (for example, 1 cm) can be used, for example, as this threshold. In this case, for example, if the distance between the imaging apparatus 100 and the information processing apparatus 200 is less than the threshold, it can be considered that the imaging apparatus 100 is mounted on the information processing apparatus 200.
To continue, the control unit 250 decides whether or not a change has occurred in a comparison result between the distance between the imaging apparatus 100 and the information processing apparatus 200 and the threshold (step S903). The case where a change has occurred in this comparison result means, for example, the case where this distance is changed from a condition of less than the threshold to a condition of the threshold or more, or the case where this distance is changed from a condition of the threshold or more to a condition of less than the threshold. Then, in the case where a change has not occurred in this comparison result (step S903), it proceeds to step S907.
Further, in the case where a change has occurred in this comparison result (step S903), the control unit 250 decides whether or not the condition after this change is a condition where the distance between the imaging apparatus 100 and the information processing apparatus 200 is the threshold or more (step S904). Then, in the case where the condition after this change is a condition where the distance between the imaging apparatus 100 and the information processing apparatus 200 is the threshold or more (step S904), the control unit 250 causes a display screen of the case where this distance becomes the threshold or more to be displayed in the input-output unit 240 (step S905). For example, the display screens shown in a of
On the other hand, in the case where the condition after this change is a condition where the distance between the imaging apparatus 100 and the information processing apparatus 200 is less than the threshold (step S904), the control unit 250 causes a display screen of the case where this distance becomes less than the threshold to be displayed in the input-output unit 240 (step S906). For example, the display screens shown in b of
To continue, the control unit 250 decides whether or not there is an end instruction of the imaging operation (step S907), and in the case where there is no end instruction of the imaging operation, it returns to step S901, and in the case where there is an end instruction of the imaging operation, the operation of the display control process ends.
In the first embodiment of the present technology, an example has been shown where a display state of a display screen is switched based on a distance between an imaging apparatus and an information processing apparatus. Here, for example, a case is assumed where a distance between an imaging apparatus and an information processing apparatus is less than a threshold, and the imaging apparatus is mounted on the display surface of an input-output unit of the information processing apparatus. In this case, a user is not able to view the portion where the imaging apparatus is attached, from within the display screen displayed in the input-output unit. Accordingly, in such a case, it is preferable for a display state of a display screen to be switched based on this attachment position.
Accordingly, in a second embodiment of the present technology, an example is shown where a display state of a display screen is switched based on an attachment position of an imaging apparatus. Note that, the imaging apparatus and the information processing apparatus in the second embodiment of the present technology are the same as the imaging apparatus 100 and the information processing apparatus 200 shown in
First, a determination method will be described which determines an attachment position of the imaging apparatus 100.
For example, a detection member for detecting the mounting of the imaging apparatus 100 is installed in the information processing apparatus 200. For example, detection members are installed at both end parts (both end parts of the longer direction of the information processing apparatus 200) on each of the front surface (the surface on which the input-output unit 240 is installed) and the rear surface (the surface on which the imaging unit 292 is installed) of the information processing apparatus 200. A switch or sensor for detecting the mounting of the imaging apparatus 100 can be used, for example, as this detection member. Also, by using this detection member, it can be detected at what portion of the front surface and the rear surface of the information processing apparatus 200 the imaging apparatus 100 is mounted.
For example, in the case where the input-output unit 240 is a touch panel of a light sensor system, a position and size of the imaging apparatus 100 mounted on the display surface of the input-output unit 240 can be detected. Further, for example, in the case where the input-output unit 240 is a touch panel of a projection-type electrostatic capacitance system, a position and size of the imaging apparatus 100 mounted on the display surface of the input-output unit 240 can also be detected. For example, by adopting a material which reacts with a touch panel of a projection-type electrostatic capacitance system, on the mounting surface 175 (shown in b of
Accordingly, the control unit 250 decides whether or not a distance between the imaging apparatus 100 and the information processing apparatus 200 is less than a threshold (for example, several cm), and in the case where this distance is less than a threshold, can decide that the imaging apparatus 100 is mounted on the information processing apparatus 200. In this case, the control unit 250 can determine an attachment position on the display surface of the input-output unit 240, based on a detection state on the display surface of the input-output unit 240. An example of this determination process is shown in
First, the control unit 250 detects a distance between the imaging apparatus 100 and the information processing apparatus 200 (step S911). This detection method of a distance can use any of the above described methods.
To continue, the control unit 250 compares the distance between the imaging apparatus 100 and the information processing apparatus 200 and a threshold, and decides whether or not this distance is less than the threshold (step S912). Several cm (for example, 1 cm) can be used, for example, as this threshold. Then, in the case where this distance is the threshold or more (step S912), the control unit 250 determines that the imaging apparatus 100 is not mounted on the information processing apparatus 200 (step S913).
Further, in the case where this distance is less than the threshold (step S912), the control unit 250 determines that the imaging apparatus 100 is mounted on the information processing apparatus 200. In this case, the control unit 250 determines an attachment position of the imaging apparatus 100 in the information processing apparatus 200.
Specifically, the control unit 250 acquires a position and size of an object in contact with the display surface of the input-output unit 240 (step S914). To continue, the control unit 250 decides whether or not the size of this acquired object is the same (or, approximately the same) as a setting value (step S915). This decision of approximately the same can be set to a range in which an error detection can be allowed. Further, the setting value can be set, for example, to a value the same (or, approximately the same) as the size of the mounting surface 175 (shown in b of
In the case where the size of this acquired object is the same (or, approximately the same) as the setting value (step S915), the control unit 250 determines that the imaging apparatus 100 is mounted on the display surface of the input-output unit 240. In this case, the control unit 250 determines an attachment position of the imaging apparatus 100 on the display surface of the input-output unit 240.
Specifically, the control unit 250 determines whether or not the position of this acquired object (the position of the object in contact with the display surface of the input-output unit 240) is on the left side on the display surface of the input-output unit 240 (step S917). Then, in the case where the position of this acquired object is on the left side on the display surface of the input-output unit 240 (step S917), the control unit 250 determines that an attachment position of the imaging apparatus 100 is on the left side on the display surface of the input-output unit 240 (step S918). On the other hand, in the case where the position of this acquired object is not on the left side on the display surface of the input-output unit 240 (in the case where it is on the right side) (step S917), the control unit 250 decides that an attachment position of the imaging apparatus 100 is on the right side on the display surface of the input-output unit 240 (step S919).
Note that, examples are shown in
An example is shown in
A display example is shown in a of
A display example is shown in b of
For example, as shown in a of
For example, as shown in a of
Examples are shown in
A display example is shown in a of
A display example is shown in b of
First, the control unit 250 detects an attachment position of the imaging apparatus 100 in the information processing apparatus 200 (step S921). This detection method of an attachment position can use any of the above described methods.
To continue, the control unit 250 decides whether or not a change has occurred in the attachment position of the imaging apparatus 100 in the information processing apparatus 200 (step S922). The case where a change has occurred in this attachment position means, for example, the case where this attachment position has moved on the display surface of the input-output unit 240, or the case where this attachment position has changed to another surface (for example, the case where moved to the surface on the opposite side, from the surface on which the input-output unit 240 is installed). Then, in the case where a change has not occurred in this attachment position (step S922), it proceeds to step S928.
Further, in the case where a change has occurred in this attachment position (step S922), the control unit 250 decides whether or not the attachment position of the imaging apparatus 100 is on the display surface of the information processing apparatus 200 (the surface of the input-output unit 240) (step S923). Then, in the case where the attachment position of the imaging apparatus 100 is not on the display surface of the information processing apparatus 200 (step S923), the control unit 250 causes a display screen of the case where the attachment position of the imaging apparatus 100 is on the rear side of the information processing apparatus 200 to be displayed in the input-output unit 240 (step S924). For example, the display screen shown in a of
In the case where the attachment position of the imaging apparatus 100 is on the display surface of the information processing apparatus 200 (step S923), the control unit 250 decides whether or not the attachment position of the imaging apparatus 100 is on the left side on the display surface of the input-output unit 240 (step S925).
In the case where the attachment position of the imaging apparatus 100 is on the left side on the display surface of the input-output unit 240 (step S925), the control unit 250 causes a display screen of the case where the attachment position of the imaging apparatus 100 is on the left side on the display surface of the input-output unit 240 to be displayed in the input-output unit 240 (step S926). For example, the display screen shown in a of
In the case where the attachment position of the imaging apparatus 100 is on the right side on the display surface of the input-output unit 240 (step S925), the control unit 250 causes a display screen of the case where the attachment position of the imaging apparatus 100 is on the right side on the display surface of the input-output unit 240 to be displayed in the input-output unit 240 (step S927). For example, the display screen shown in b of
Note that, in the second embodiment of the present technology, an example has been shown where it is determined which of both sides on the display surface of the input-output unit 240 (the left side, the right side), as an attachment position of the imaging apparatus 100. However, for example, it may be determined which out of 3 or more regions on the display surface of the input-output unit 240, as an attachment position of the imaging apparatus 100. In this case, a live view image and operation objects are displayed, so as to avoid the attachment position of the imaging apparatus 100.
In the first and the second embodiments of the present technology, examples have been shown where a display state of a display screen is switched based on a distance between an imaging apparatus and an information processing apparatus, and an attachment position of an imaging apparatus. Here, it is possible for the information processing apparatus shown in the first and the second embodiments of the present technology to switch a display state of a display screen based on the posture of the information processing apparatus, such as described above. For example, a display state of a display screen can be switched, in accordance with whether the information processing apparatus is in a horizontally long condition (a condition where the longer direction of the information processing apparatus becomes approximately parallel with a horizontal direction), or a vertically long condition (a condition where the longer direction of the information processing apparatus becomes approximately parallel with a vertical direction). However, for example, in the case where an imaging operation is performed in a condition where the imaging apparatus is mounted on the information processing apparatus, it is assumed that the use states of the imaging apparatus and the information processing apparatus will be different compared to the case where an imaging operation is performed in a condition where the imaging apparatus and the information processing apparatus are separated. Accordingly, in the case where an imaging operation is performed by using an imaging apparatus and an information processing apparatus, it will be important to provide an appropriate user interface in accordance with these use states.
Accordingly, in a third embodiment of the present technology, an example is shown where switching of a display state is controlled based on a change in the posture of an information processing apparatus, based on the presence or absence of mounting of an imaging apparatus and an information processing apparatus.
The imaging apparatus 101 includes a posture detection unit 180.
The posture detection unit 180 detects a change in the posture of the imaging apparatus 101 by detecting an acceleration, movement, inclination or the like of the imaging apparatus 101, and outputs posture information related to the detected change in posture to the control unit 140. For example, the posture detection unit 180 can detect a rotation angle, in the case where an optical axis direction of the imaging apparatus 101 is set to a rotation axis, as a change in the posture of the imaging apparatus 101. Note that, various types of sensors such as a gyro sensor or an acceleration sensor can be used, for example, as the posture detection unit 180.
The control unit 140 transmits the posture information output from the posture detection unit 180 to the information processing apparatus 200 via the wireless communication unit 150. In this way, the control unit 250 of the information processing apparatus 200 can comprehend the posture of the imaging apparatus 101. Accordingly, in the case where an image generated by the imaging unit 110 is displayed in the input-output unit 240 as a live view image, for example, the control unit 250 can change this display state, based on the posture of the imaging apparatus 101. This change example of a display state will be shown in
[Example where the Mounting of the Imaging Apparatus is Detected by Using Short-Distance Wireless Communication]
Here, another detection method other than the above described detection method of mounting will be described. For example, the mounting to the information processing apparatus 200 of the imaging apparatus 101 can be detected by using short-distance wireless communication. Here, an example will be shown which uses Near Field Communication (NFC) as short-distance wireless communication.
For example, in NFC, regular communication is performed between apparatuses. For example, a regular Polling Command is issued from the information processing apparatus 200. This Polling Command is a command for distinguishing an NFC Tag specified by an NFC specification (for example, refer to NFC Forum Type3 Tag Operation Specification NFC Forum-TS-Type-3-Tag_1.1).
For example, in the case where there is a response (Polling Responce) to a regularly issued Polling Command, the information processing apparatus 200 can decide that an adjoining apparatus is present. Here, the distance at which data communication using NFC becomes possible is approximately 1 through to 10 cm. Accordingly, in the case where there is a Polling Responce, the control unit 250 of the information processing apparatus 200 can decide that the imaging apparatus 101 has been mounted on the information processing apparatus 200.
Further, in the case where there is no response (Polling Responce) to a Polling Command, the information processing apparatus 200 can decide that an adjoining apparatus is not present. In this case, the control unit 250 of the information processing apparatus 200 can decide that the imaging apparatus 101 is not mounted on the information processing apparatus 200.
Here, NFC is widespread. Further, NFC is often included in an information processing apparatus such as a smartphone. Accordingly, in the case of an information processing apparatus which includes NFC, by performing a detection of mounting using NFC, it may not be necessary to newly install hardware for performing this detection. In this way, the manufacturing cost of an information processing apparatus can be reduced.
Here, for example, it is assumed to be a case such as an imaging apparatus and an information processing apparatus are adjoining, without the imaging apparatus being mounted on the information processing apparatus. In such a case, when performing a mounting detection by using short-distance wireless communication other than NFC, there will the possibility that an erroneous detection is made when the imaging apparatus is mounted. Accordingly, in the embodiments of the present technology, NFC will be used as short-distance wireless communication. In this way, the accuracy of a mounting detection of an imaging apparatus can be improved.
[Example where the Mounting of the Imaging Apparatus is Detected by Using a Check Command and a Check Responce]
Heretofore, an example has been shown where the mounting of an imaging apparatus is detected by using NFC. In this way, in the case where the mounting of an imaging apparatus is detected by using NFC, the mounting of the imaging apparatus may be detected by using a Check Command and a Check Responce.
For example, in the case where an exchange of a Check Command and a Check Responce is performed, and specific information is included in the Check Responce, it can be decided that the imaging apparatus 101 is mounted on the information processing apparatus 200. Here, specific information is information (identification information) for specifying the imaging apparatus 101. Further, a check command and a check response are commands for reading a content of an NFC tag. These commands are defined in an NFC Forum Type3 Tag Operation Specification.
For example, an issue of a Polling Command, and an exchange of a response (Polling Responce) to the Polling Command, are performed, such as described above.
Then, when the Polling Responce is received, the control unit 250 of the information processing apparatus 200 transmits a Check Command. When this Check Command is received, the control unit 140 of the imaging apparatus 101 transmits a response (Check Responce) to the Check Command. In this case, the control unit 140 of the imaging apparatus 101 performs transmission by including specific information (information (identification information) for specifying the imaging apparatus 101) in the Check Responce.
For example, it can be transmitted by including information representing [ABC DSC/Lens-Style Camera] in the Check Responce as specific information. Here, for example, [ABC] is information showing a company name which manufactures the imaging apparatus 101, and [DSC] is information showing that it is an imaging apparatus. Further, [Lens-Style Camera] is information showing that it is a lens-style camera.
In this way, the control unit 140 of the imaging apparatus 101 performs transmission by including specific information in the Check Responce. In this way, the information processing apparatus 200 receiving the Check Responce can comprehend that the apparatus transmitting the Check Responce is the imaging apparatus 101, based on the specific information included in the Check Responce. That is, it can be comprehended that the apparatus transmitting the Check Responce is a lens-style camera (imaging apparatus 101) made by [AAA] company.
When the Check Responce is received, the control unit 250 of the information processing apparatus 200 acquires a content of the Check Responce. To continue, the control unit 250 of the information processing apparatus 200 decides whether or not specific information is included in the Check Responce. Then, in the case where specific information is included in the Check Responce, the control unit 250 of the information processing apparatus 200 determines that the imaging apparatus 101 is mounted. On the other hand, in the case where specific information is not included in the Check Responce, the control unit 250 of the information processing apparatus 200 determines that the imaging apparatus 101 is not mounted.
Note that, this detection method is an example, and is not limited to this, and another detection method may be used.
A captured image 501 is shown in a of
A captured image 502 is shown in b of
A captured image 503 is shown in c of
A display example is shown in a of
The control unit 250 decides the posture of the information processing apparatus 200, based on posture information output from the posture detection unit 210. Then, in the case where the longer direction of the information processing apparatus 200 becomes approximately the same as a horizontal direction, such as shown in a of
Further, a display example is shown in d of
Transition examples are shown in a through to d of
For example, as shown in a through to d of
Further, it becomes similar to the transition example of a display screen shown in a through to d of
Transition examples are shown in a through to d of
For example, similar to the example shown in a through to d of
Further, it becomes similar to the transition example of the display screen shown in
In this way, in the case where the imaging apparatus 101 is not mounted on the information processing apparatus 200, a display state of a live view image and operation objects can be changed, based on a change in the posture of the information processing apparatus 200. However, in the case where the imaging apparatus 101 is mounted on the information processing apparatus 200, it is assumed that a user is holding the imaging apparatus 101 and the information processing apparatus 200 with the same way of holding, even if the imaging apparatus 101 and the information processing apparatus 200 have some posture. Accordingly, when a display state of a live view image and operation objects is changed, based on a change in the posture of the information processing apparatus 200, it is assumed that it will be difficult for a user to operate. Accordingly, in the third embodiment of the present technology, in the case where the imaging apparatus 101 is mounted on the information processing apparatus 200, a change of a display state of a live view image and operation objects is not performed based on a change in the posture of the information processing apparatus 200.
The example shown in a of
A display example is shown in b of
In this way, in the case where the imaging apparatus 101 is mounted on the information processing apparatus 200, a display state of operation objects displayed in the input-output unit 240 is not changed, even in the case where the posture of the information processing apparatus 200 is changed. In this way, operation objects are displayed with the same display state regardless of the posture of the imaging apparatus 101 and the information processing apparatus 200.
Further, in the case where the imaging apparatus 101 is mounted on the information processing apparatus 200, a display state of a live view image displayed in the input-output unit 240 is not changed, even in the case where the posture of the imaging apparatus 101 is changed. In this way, a live view image can be confirmed with the same display state regardless of the posture of the imaging apparatus 101 and the information processing apparatus 200.
For example, it is assumed to be a case where a user who is holding the imaging apparatus 101 and the information processing apparatus 200 performs an operation (zoom operation) of an operation object 527 with the thumb of his or her right hand, in the condition shown in a of
Further, for example, it is assumed to be a case where a user who is holding the imaging apparatus 101 and the information processing apparatus 200 performs an operation (shutter operation) of an operation object 522 with the thumb of his or her right hand, in the condition shown in a of
In this way, at the time of the mounting of the imaging apparatus 101 and the information processing apparatus 200, a display state of operation objects and a live view image displayed in the input-output unit 240 is not changed. In this way, it can prevent a user sensing that it is difficult to operate, by having the position of operation objects for performing a zoom operation or a shutter operation changed, every time the posture of the information processing apparatus 200 changes. Further, it can prevent a user sensing that it is difficult to confirm a photographic subject, by having a live view image changed, every time the posture of the information processing apparatus 200 changes. That is, the operativeness can be improved at the time of an imaging operation.
Note that, here, an example has been shown where a display state of all operation objects is not changed, at the time of the mounting of the imaging apparatus 101 and the information processing apparatus 200. However, a display state of a part of the operation objects (for example, operation objects not particularly used during an imaging operation) may change, at the time of the mounting of the imaging apparatus 101 and the information processing apparatus 200. In this case, the other operation objects (for example, operation objects frequently used during an imaging operation) are not changed. Further, for example, at least 1 operation object may have only this orientation changed, without changing this position. For example, the operation object 522 may have this orientation changed based on a change in the posture of the information processing apparatus 200, without changing the position on the display surface of the input-output unit 240. For example, it may be changed so that a camera mark of the operation object 522 becomes upwards, based on a change in the posture of the information processing apparatus 200. Further, a user may set the presence or absence of a change of a display state of an operation object for each operation object, and the presence or absence of a change may be decided based on this setting content.
First, the control unit 250 causes a display screen which includes a live view image and operation objects to be displayed in the input-output unit 240 (step S931). For example, the display screen shown in a of
To continue, the control unit 250 detects a change in the posture of the information processing apparatus 200, based on posture information from the posture detection unit 210 (step S932). Then, the control unit 250 decides whether or not this change is a threshold or more (step S933). For example, as shown in
Further, in the case where this change is a threshold or more (step S933), the control unit 250 decides whether or not the imaging apparatus 101 is mounted on the information processing apparatus 200 (step S934). Then, in the case where the imaging apparatus 101 is mounted on the information processing apparatus 200 (step S934), it proceeds to step S936. Further, in the case where the imaging apparatus 101 is not mounted on the information processing apparatus 200 (step S934), the control unit 250 changes the live view image and the operation objects and causes them to be displayed in the input-output unit 240, based on the posture of the information processing apparatus 200 (step S935). For example, as shown in
To continue, the control unit 250 decides whether or not there is an end instruction of the imaging operation (step S936), and in the case where there is no end instruction of the imaging operation, it returns to step S931, and in the case where there is an end instruction of the imaging operation, the operation of the display control process ends.
In this way, the control unit 250 causes a display screen which includes operation objects for operating the imaging apparatus 101 to be displayed in the input-output unit 240. Also, in the case where the imaging apparatus 101 is not mounted on the information processing apparatus 200, the control unit 250 performs a control for changing a display state of a live view image and operation objects based on a change in the posture of the information processing apparatus 200. On the other hand, in the case where the imaging apparatus 101 is mounted on the information processing apparatus 200, the control unit 250 performs a control so as not to change a display state of a live view image and operation objects based on a change in the posture of the information processing apparatus 200.
Note that, the display control processes shown in the first and the second embodiments of the present technology (switching a display state of a display screen based on a distance between an imaging apparatus and an information processing apparatus, and an attachment position of an imaging apparatus) may be performed in combination in the display control process shown in
In this way, according to the embodiments of the present technology, an appropriate user interface corresponding to the use states of the imaging apparatuses 100 and 101 and the information processing apparatus 200 can be provided.
Further, in the embodiments of the present technology, while cylindrical (columnar shaped) imaging apparatuses 100 and 101 have been described as an example, the embodiments of the present technology can also be applied to an imaging apparatus with another shape capable of being attached to another apparatus. Further, in the embodiments of the present technology, while information processing apparatuses 200 through to 203 such as a smartphone or a tablet terminal have been described, the embodiments of the present technology can also be applied to another apparatus capable of connecting to an imaging apparatus by using wireless communication.
Further, in the embodiments of the present technology, while a description has been made by setting, as an example, a relationship of an imaging apparatus and an information processing apparatus, the embodiments of the present technology can also be applied to that between other apparatuses. For example, the embodiments of the present technology can also be applied to a light emitting apparatus and an imaging apparatus (for example, a digital still camera) used at the time of still image photography. Further, the embodiments of the present technology can also be applied to a sound collection apparatus (for example, a microphone) and an imaging apparatus (for example, a camera integrated recorder) used at the time of moving image photography. Further, the embodiments of the present technology can also be applied to an audio output apparatus (for example, a speaker) and an imaging apparatus (for example, a camera integrated recorder) used at the time of moving image reproduction.
Further, for example, the embodiments of the present technology can also be applied to a sound collection apparatus (for example, a microphone) and an information processing apparatus. Further, the embodiments of the present technology can also be applied to an audio output apparatus (for example, a speaker) and an information processing apparatus used at the time of moving image reproduction. That is, the embodiments of the present technology can be comprehended as an example of an information processing apparatus (electronic device) including a control unit which performs a control for switching a display state of a display screen for operating another apparatus based on a relative position relationship with the other apparatus.
Note that, the above-described embodiments show examples for embodying the present technology, and the matters in the embodiments have a corresponding relationship with invention-specific matters in the claims. Similarly, the invention-specific matters in the claims, and the matters in the embodiments of the present technology to which the same names as these are attached, have a corresponding relationship with each other. However, the present technology is not limited to the embodiments, and various modifications of the embodiments may be embodied in the scope of the present technology without departing from the spirit of the present technology.
Further, the above described process procedures described in the embodiments may be handled as a method having this series of procedures, or may be handled as a program for causing a computer to execute this series of procedures and a recording medium storing this program. A Compact Disc (CD), a MiniDisc (MD), a Digital Versatile Disc (DVD), a memory card, a Blu-ray (registered trademark) Disc or the like can be used, for example, as this recording medium.
Note that, the effect described herein is not limited to that illustrated, and there may be additional effects.
Additionally, the present technology may also be configured as below.
(1)
An information processing apparatus, including:
a control unit which performs a control for switching a display state of a display screen for operating an imaging apparatus based on a relative position relationship with the imaging apparatus.
(2)
The information processing apparatus according to (1),
wherein the control unit performs a control for switching a display state of the display screen based on a distance between the information processing apparatus and the imaging apparatus.
(3)
The information processing apparatus according to (1), wherein the control unit performs a control for switching a display state of the display screen based on whether or not the imaging apparatus is mounted on the information processing apparatus.
(4)
The information processing apparatus according to (3),
wherein the control unit performs a control for switching a display state of the display screen based on whether or not the imaging apparatus is mounted on a display surface of the information processing apparatus.
(5)
The information processing apparatus according to (4),
wherein, in the case where the imaging apparatus is mounted on a display surface of the information processing apparatus, the control unit performs a control for switching a display state of the display screen based on a position of the imaging apparatus on a display surface of the information processing apparatus.
(6)
The information processing apparatus according to any of (1) to (5), wherein the control unit causes the display screen which includes an operation object for operating the imaging apparatus to be displayed, and performs a control for changing a display state of the operation object based on the relative position relationship.
(7)
The information processing apparatus according to any of (1) to (6),
wherein the control unit causes the display screen which includes an operation object for operating the imaging apparatus to be displayed, and performs a control, in the case where the imaging apparatus is not mounted on the information processing apparatus, for changing a display state of the operation object based on a change in a posture of the information processing apparatus, and in the case where the imaging apparatus is mounted on the information processing apparatus, for not changing a display state of the operation object based on a change in a posture of the information processing apparatus.
(8)
An imaging apparatus, including:
a control unit which performs a control related to an imaging operation based on an operation input performed in the information processing apparatus in which a display screen is displayed for a display state to be switched based on a relative position relationship of the imaging apparatus and the information processing apparatus.
(9)
The imaging apparatus according to (8),
wherein a display state of the display screen is switched based on a distance between the information processing apparatus and the imaging apparatus.
(10)
The imaging apparatus according to (8),
wherein a display state of the display screen is switched based on whether or not the imaging apparatus is mounted on the information processing apparatus.
(11)
The imaging apparatus according to (10),
wherein a display state of the display screen is switched based on whether or not the imaging apparatus is mounted on a display surface of the information processing apparatus.
(12)
The imaging apparatus according to (11),
wherein, in the case where the imaging apparatus is mounted on a display surface of the information processing apparatus, a display state of the display screen is switched based on a position of the imaging apparatus on a display surface of the information processing apparatus.
(13)
The imaging apparatus according to any of (8), to (12)
wherein the information processing apparatus causes the display screen which includes an operation object for operating the imaging apparatus to be displayed, and changes a display state of the operation object based on the relative position relationship.
(14)
The imaging apparatus according to any of (8) to (13),
wherein the information processing apparatus causes the display screen which includes an operation object for operating the imaging apparatus to be displayed, and in the case where the imaging apparatus is not mounted on the information processing apparatus, changes a display state of the operation object based on a change in a posture of the information processing apparatus, and in the case where the imaging apparatus is mounted on the information processing apparatus, does not change a display state of the operation object based on a change in a posture of the information processing apparatus.
(15)
An imaging system, including:
an imaging apparatus in which a control related to an imaging operation is performed based on an operation input performed in an information processing apparatus by connecting to the information processing apparatus by using wireless communication; and
an information processing apparatus which performs a control for switching a display state of a display screen for operating the imaging apparatus based on a relative position relationship with the imaging apparatus.
(16)
A control method of an information processing apparatus which performs a control for switching a display state of a display screen for operating an imaging apparatus based on a relative position relationship with the imaging apparatus.
(17)
A control method of an imaging apparatus which performs a control related to an imaging operation based on an operation input performed in an information processing apparatus in which a display screen is displayed for a display state to be switched based on a relative position relationship of the imaging apparatus and the information processing apparatus.
(18)
A program for causing a computer to execute a control for switching a display state of a display screen for operating an imaging apparatus based on a relative position relationship with the imaging apparatus.
(19)
A program for causing a computer to execute a control related to an imaging operation based on an operation input performed in an information processing apparatus in which a display screen is displayed for a display state to be switched based on a relative position relationship of an imaging apparatus and the information processing apparatus.
Number | Date | Country | Kind |
---|---|---|---|
2013-210118 | Oct 2013 | JP | national |
2013-237988 | Nov 2013 | JP | national |
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/JP2014/069547 | 7/24/2014 | WO | 00 |