The present invention relates to hand-held electronic devices. More particularly, the invention relates to hand-held electronic devices each having a stereoscopic display section for displaying an image which is stereoscopically visible with naked eyes.
In recent years, hand-held electronic devices each having a stereoscopic display section for displaying an image which is stereoscopically visible with naked eyes have been proposed. For example, an electronic device disclosed in Patent Literature 1 is provided with a compound-eye camera having two imaging sensors. In this electronic device, two images taken by the compound-eye camera at the time of shooting are combined to create an image for stereoscopically visible display, and the image is three-dimensionally displayed on a monitor.
[PTL 1] Japanese Laid-Open Patent Publication No. 2010-56737
In Patent Literature 1, however, two imaging sensors are provided, and the functions thereof are limited. Further, in Patent Literature 1, an image taken by the compound-eye camera and a stored image are merely displayed on a display device. In order to take an image in a direction different from the imaging direction of the compound-eye camera, the direction of the compound-eye camera must be changed. In order to stereoscopically view a 3D image displayed on the display device, a user needs to keep close watch on a screen of the display device and focus his/her eyes on the 3D image. However, if the user takes his/her eyes from the screen to change the imaging direction of the compound-eye camera and then turns the eyes back to the screen, the eyes are out of focus, and the screen may look blurred to the user's eyes. Thus, in the electronic device disclosed in Patent Literature 1, the direction of the compound-eye camera must be changed when taking an image in a direction different from the imaging direction of the compound-eye camera, which makes it difficult for the user to view the image displayed on the screen.
Therefore, an object of the present invention is to provide a hand-held electronic device having a variety of functions.
To achieve the above object, the present invention has the following configurations.
A hand-held electronic device of the present invention includes a stereoscopic display section capable of displaying an image which is stereoscopically visible with naked eyes, and three imaging sections.
According to the invention, a variety of functions can be realized by the stereoscopic display section capable of displaying an image which is stereoscopically visible with naked eyes, and the three imaging sections.
In the present invention, two of the three imaging sections may be arranged so as to be horizontally symmetrical with respect to a center of a screen of the stereoscopic display section.
In the present invention, one of the three imaging sections may be aligned with a horizontal center of a screen of the stereoscopic display section.
In the present invention, the three imaging sections may be linearly arranged along the horizontal direction of the stereoscopic display section.
In the present invention, the three imaging sections may be assembled as a single unit, and the unit may be housed in a storage section of a housing.
In the present invention, the three imaging sections may be arranged so that the imaging directions of two of the three imaging sections are in a first direction, and the imaging direction of the other one imaging section is in a second direction opposite to the first direction.
In the present invention, two of the three imaging sections are capable of stereoscopic shooting, and the stereoscopic display section may display a stereoscopic image by using images taken by and outputted from the two imaging sections.
In the present invention, the stereoscopic display section may be located on a main surface of a housing of the electronic device. Two of the three imaging sections are arranged on a rear surface of the housing, which is reverse to the main surface, so as to be aligned in parallel with the horizontal direction of the display screen of the stereoscopic display section, thereby configuring an outer imaging section. The other one of the three imaging sections is arranged on the main surface of the housing, thereby configuring an inner imaging section. The electronic device further includes selection means and display control means. The selection means selects either of the outer imaging section and the inner imaging section. When the selection means selects the outer imaging section, the display control means causes the stereoscopic display section to display the stereoscopic image by using a right-eye image and a left-eye image taken by the outer imaging section.
In the above configuration, a user can take a stereoscopic image by using the outer imaging section while viewing the display screen of the stereoscopic display section from the front thereof. In addition, when the user views the display screen of the stereoscopic display section from the front thereof, the inner imaging section becomes able to take an image of the user's face. Therefore, when a stereoscopic image taken by the outer imaging section is displayed on the display screen of the stereoscopic display section and the user views the stereoscopic image, the inner imaging section can take an image of the user's face.
In the present invention, the stereoscopic display section may be capable of displaying the stereoscopic image and a planar image. In this case, when the selection means selects the outer imaging section, the display control means may cause the stereoscopic display section to display the stereoscopic image by using a right-eye image and a left-eye image taken by the outer imaging section. Further, when the selection means selects the inner imaging section, the display control means may cause the stereoscopic display section to planarly display an image taken by the inner imaging section.
In the above configuration, an image taken by the outer imaging section can be stereoscopically displayed when the outer imaging section is selected, and an image taken by the inner imaging section can be planarly displayed when the inner imaging section is selected.
In the present invention, the display control means may cause the stereoscopic display section to display, in real time, the image taken by the imaging section selected by the selection means.
In the above configuration, an image taken by the imaging section selected by the selection means can be displayed on the stereoscopic display section in real time. Accordingly, a user can view the taken image in real time.
In the present invention, the three imaging sections may be located on either of the foldable two housings.
In the present invention, the stereoscopic display section may be capable of switching the display mode between a stereoscopic view mode and a planar view mode. Further, the electronic device may include a hardware operation switch for performing switching between the stereoscopic view mode and the planar view mode. The operation switch is a slider switch, and the degree of stereoscopic effect of the stereoscopic display section is adjustable by sliding the slider switch. The display mode is switched to the planar view mode when the slider switch is slid to an end position in the direction in which the stereoscopic effect deteriorates.
In the present invention, stereoscopically visible display may be performed on the stereoscopic display section in the stereoscopic view mode by performing stereo-shooting with the two imaging sections, and planarly visible display may be performed on the stereoscopic display section in the planarly view mode by performing monaural-shooting with the single imaging section.
In the present invention, stereoscopically visible display may be performed on the stereoscopic display section in the stereoscopic view mode by performing stereo-shooting with the two imaging sections facing the first direction, and planarly visible display may be performed on the stereoscopic display section in the planarly view mode by performing monaural-shooting with the single imaging section facing the second direction.
In the present invention, stereoscopically visible display may be performed on the stereoscopic display section in a stereoscopic view mode by using images taken by the two imaging sections facing in the first direction, and planarly visible display may be performed on the stereoscopic display section in a planar view mode by using an image taken by the single imaging sections facing in the second direction. Alternatively, stereoscopically visible display or planarly visible display may be performed on the stereoscopic display section by combining at least two images among the images taken by the two imaging sections facing the first direction and the image taken by the single imaging section facing the second direction.
In the present invention, stereoscopically visible display may be performed on the stereoscopic display section in a stereoscopic view mode by using images taken by the two imaging sections facing in the first direction, and planarly visible display may be performed on the stereoscopic display section in a planar view mode by using an image taken by the single imaging sections facing in the second direction. Alternatively, stereoscopically visible display of a real image may be performed on the stereoscopic display section in the stereoscopic view mode by using images taken by the two imaging sections facing in the first direction and, simultaneously, a virtual object may be taken by a virtual camera for a left eye and a virtual camera for a right eye by using, as a texture of the virtual object, an image taken by the single imaging section facing in the second direction, and then a stereoscopically visible virtual image thus obtained may be combined with the stereoscopically visible real image displayed on the stereoscopic display section in the stereoscopic view mode.
In the present invention, the stereoscopic display section may be capable of switching the display mode between a stereoscopic display mode for displaying a stereoscopic image which is stereoscopically visible with naked eyes, by using a right-eye image and a left-eye image, and a planar display mode for planarly displaying an image. In this case, when the selection means selects the outer imaging section, the display control means switches the display mode of the stereoscopic display section to the stereoscopic display mode, and causes the stereoscopic display section to display the stereoscopic image by using a right-eye image and a left-eye image taken by the outer imaging section. When the selection means selects the inner imaging section, the display control means switches the display mode of the stereoscopic display section to the planar display mode, and causes the stereoscopic display section to planarly display an image taken by the inner imaging section.
In the above configuration, when the outer imaging section is selected, the display mode of the stereoscopic display section is switched to the stereoscopic display mode to stereoscopically display an image taken by the outer imaging section. Further, when the inner imaging section is selected, the display mode of the stereoscopic display section is switched to the planar display mode to planarly display an image taken by the inner imaging section. For example, if the stereoscopic display section is a parallax barrier type liquid crystal display device, ON/OFF of a parallax barrier can be controlled by switching the imaging sections.
In the present invention, the hand-held electronic device may further include a second housing. The second housing has a holding part to be held by a user, and is foldably connected to a lower side portion of the housing at its upper side portion.
In the above configuration, the second housing to be held by the user is provided. Therefore, when the user holds the hand-held electronic device, the user's hand and fingers do not hang over the display screen of the stereoscopic display section. Thus, when the user views an image displayed on the display screen of the stereoscopic display section, the user's hand and fingers do not interfere with the user's stereoscopic viewing.
In the present invention, the hand-held electronic device may further include imaging instruction means. The imaging instruction section is located on the second housing, and instructs storage of an image taken by the outer imaging section or the inner imaging section. The “storage of a taken image” means temporarily storing the image in volatile storage means or permanently storing the image in non-volatile storage means.
In the above configuration, since the second housing is provided with the imaging instruction means for instructing storage of an image, when a user instructs storage of an image, the user's hand and fingers do not hang over the display screen of the stereoscopic display section. Therefore, when the user instructs storage of an image, the user's hand and fingers do not interfere with the user's stereoscopic viewing. For example, when an image taken by the outer imaging section is displayed in real time on the stereoscopic display section, the user can store the image while stereoscopically viewing the image displayed on the stereoscopic display section.
In the present invention, the main surface of the housing may be located inside when the housing and the second housing are folded. In this case, the electronic device further includes a planar display section for planarly displaying an image. The planar display section is located on an inner side surface of the second housing, which is located inside when the housing and the second housing are folded.
In the above configuration, an image can be planarly displayed on the planar display section while an image is stereoscopically displayed on the stereoscopic display section. Accordingly, information such as images and text, which need not be stereoscopically displayed, can be displayed on the planar display section while an image is stereoscopically displayed on the stereoscopic display section.
In the present invention, the hand-held electronic device may further include contact position detection means. The contact position detection means is located on the display screen of the planar display section, and detects a contact position on the display screen of the planarly display section.
In the above configuration, the detection means for detecting a contact on the touch panel or the like is provided on the display screen of the planar display section. Thereby, for example, when an operation screen is displayed on the planar display section, a user is allowed to perform an operation using the screen.
In the present invention, the hand-held electronic device may further include position designation means. The position designation means designates a position on the display screen of the stereoscopic display section, which position corresponds to the contact position on the display screen of the planarly display section, which is detected by the contact position detection means.
In the above configuration, a user can designate a position on the display screen of the stereoscopic display section by touching the position on the display screen of the planar display section. When a stereoscopic image is displayed on the stereoscopic display section, if the user designates a position on the display screen of the stereoscopic display section by directly touching the screen with a finger or a touch pen, the finger or the touch pen comes in the sight of the user, which interferes with the user's stereoscopic viewing. However, as described above, when the user touches a position on the display screen of the planar display section to designate a position, corresponding to the touched position, on the display screen of the stereoscopic display section, the finger or the touch pen does not come in the sight of the user and therefore does not interfere with the user's stereoscopic viewing.
In the present invention, the area of the display screen of the stereoscopic display section may be larger than the area of the display screen of the planar display section.
In the above configuration, since the area of the display screen of the stereoscopic display section is larger than the area of the display screen of the planar display section, a user's attention can be drawn to the display screen of the stereoscopic display section. In addition, since the stereoscopic display section and the planar display section have different screen sizes, the user can recognize the correct manner of holding the hand-held electronic device.
In the present invention, the resolution of the display screen of the stereoscopic display section in the horizontal direction may be higher than the resolution of the display screen of the planar display section in the horizontal direction.
In the above-configuration, since the resolution of the display screen of the stereoscopic display section in the horizontal direction is higher than that of the display screen of the planar display section, a user's attention can be drawn to the display screen of the stereoscopic display section. That is, when a stereoscopic image is displayed on the display screen of the stereoscopic display section, a reduction in the resolution in the horizontal direction is avoided, and thus the stereoscopic image can be displayed more clearly.
In the present invention, the inner imaging section may be located at a horizontal center of the display screen of the stereoscopic display section.
In the above configuration, when a user directly confronts the display screen of the stereoscopic display section, an image of the user's face can be taken from the front thereof by using the inner imaging section.
In the present invention, two imaging means of the outer imaging section may be located so as to be symmetrical with respect to a line which divides the display screen of the stereoscopic display section into two equal parts in the horizontal direction.
In the above configuration, when a user views the display screen of the stereoscopic display section from the front thereof, the imaging directions of the two imaging means of the outer imaging section can be made equal to the direction of the line of sight of the user.
In the present invention, the two imaging means of the outer imaging section may be located on an upper portion of the rear surface of the housing, at a position reverse of a position above an upper edge of the display screen of the stereoscopic display section.
In the above configuration, since the two imaging means of the outer imaging section and the stereoscopic display section do not interfere with each other inside the housing, the thickness of the housing can be reduced. In addition, since the outer imaging section is located at a position far from the user's hand when the user holds the second housing, the user's hand can be prevented from hanging over the two imaging means of the outer imaging section.
In the present invention, the inner imaging section may be located at a position reverse of a middle position between the two imaging means of the outer imaging section.
In the above configuration, when a user views the display screen of the stereoscopic display section from the front thereof, an image of the user's face can be taken by the inner imaging section from the front thereof. In addition, by arranging the two imaging means of the outer imaging section and the inner imaging section in this way, the two imaging means of the outer imaging section and the stereoscopic display section do not interfere with each other inside the housing.
In the present invention, the stereoscopic display section may be capable of switching the display mode between a stereoscopic display mode for displaying a stereoscopic image which is stereoscopically visible with naked eyes, by using a right-eye image and a left-eye image, and a planar display mode for planarly displaying an image. In this case, the hand-held electronic device further includes a switch which is located on the main surface of the housing, and switches the display mode between the stereoscopic display mode and the planar display mode.
In the above configuration, the display mode of the stereoscopic display section can be changed by using the switch. In addition, since the switch is located on the main surface of the housing, a user can view the switch when the user views the display screen of the stereoscopic display section.
In the present invention, the stereoscopic display section may be capable of switching the display mode between a stereoscopic display mode for displaying a stereoscopic image which is stereoscopically visible with naked eyes, by using a right-eye image and a left-eye image, and a planar display mode for planarly displaying an image. In this case, the hand-held electronic device further includes a switch which is located on the main surface of the housing, and switches the display mode between the stereoscopic display mode and the planar display mode.
In the above configuration, the switch for switching the display mode of the stereoscopic display section is located on the housing, in the case where the second housing to be held by a user is provided. Thereby, a user is prevented from operating the switch accidentally while holding the second housing.
In the present invention, the switch may be a slider, and the slider may be movable in a predetermined range between an first position and a second position, and to a third position. In this case, the stereoscopic display section is switched to the planar display mode when the slider is located at the third position, and switched to the stereoscopic display mode when the slider is located in the predetermined range. The hand-held electronic device further includes adjustment means. When the slider is located in the predetermined range, the adjustment means adjusts a shift in the horizontal direction between the position of the right-eye image and the position of the left-eye image on the display screen of the stereoscopic display section, in accordance with the position of the slider in the predetermined range.
In the above configuration, the display mode of the stereoscopic display section can be changed by using the switch. In addition, the positions of a right-eye image and a left-eye image in the horizontal direction can be adjusted by using the switch. Thereby, a manner in which the stereoscopic image is visible can be adjusted by using the switch.
In the present invention, the hand-held electronic device may further includes a first indicator which is located on the main surface of the housing, and indicates whether or not the display mode of the stereoscopic display section is the stereoscopic display mode.
In the above configuration, since the first indicator is located on the main surface of the housing, when a user views the display screen of the stereoscopic display section, the user can view the first indicator. Thereby, the user can know the display mode of the stereoscopic display section.
In the present invention, the hand-held electronic device may further includes a second indicator which is located on a surface different from the main surface of the housing, and indicates a predetermined state of the electronic device, which is different from the state indicated by the first indicator.
In the above configuration, for example, an indicator indicating ON/OFF of a power of the hand-held electronic device is located on a surface different from the main surface of the housing. Thereby, when a user views the display screen of the stereoscopic display section, the second indicator does not come in the sight of the user and therefore does not interfere with the user's stereoscopic viewing.
In the present invention, the entire main surface of the housing, including the display screen of the stereoscopic display section, may be covered with a transparent screen cover.
In the above configuration, the entire main surface of the housing and the display screen of the stereoscopic display section seem to be united as one body, and thus a user's attention can be drawn to the display screen of the stereoscopic display section.
In the present invention, the hand-held electronic device may further includes an analog stick which is positioned such that a finger of a user contacts the analog stick when the user holds the electronic device to view the display screen of the stereoscopic display section from the front thereof.
In the above configuration, a user can designate any direction by using the analog stick. For example, when a 3D virtual space is displayed on the stereoscopic display section and the user performs an operation to designate a direction in the 3D virtual space, the user can easily perform this operation by using the analog stick.
In the present invention, the hand-held electronic device may further include a cross key located beneath the analog stick.
In the above configuration, a user can perform a variety of operations depending on a displayed image by using, for example, the analog stick when an image is stereoscopically displayed on the stereoscopic display section, or the cross key when an image is planarly displayed on the stereoscopic display section.
In the present invention, the hand-held electronic device may further include an analog stick, and a cross key. The analog stick is positioned such that a finger of a user contacts the analog stick when the user holds the second housing to view the display screen of the stereoscopic display section from the front thereof The cross key is located beneath the analog stick.
In the above configuration, in the case where the planar display section is provided beneath the stereoscopic display section, the analog stick is located at the upper side while the cross key is located at the lower side. Thereby, for example, the stereoscopic display section and the planar display section can be related to the analog stick and the cross key, respectively.
In the present invention, the hand-held electronic device may further include attitude detection means, and attitude annunciation means. The attitude detection means detects an attitude of the hand-held electronic device. The attitude annunciation means announces whether or not the attitude detected by the attitude detection means is horizontal.
In the above configuration, the attitude of the hand-held electronic device is detected by the attitude detection means, and whether the device is horizontal or not can be announced to the user.
In the present invention, the attitude annunciation means may display, on the stereoscopic display section, whether or not the attitude detected by the attitude detection means is horizontal.
In the above configuration, when a user views the display screen of the stereoscopic display section, the user can know the attitude of the hand-held electronic device. Thereby, when the user takes a stereoscopic image by using the outer imaging section while viewing the display screen of the stereoscopic display section, the user can know whether the hand-held electronic device is horizontal or not.
In the present invention, the hand-held electronic device may further include a second housing, shooting instruction means, and a planar display section. The second housing is connected to a lower side portion of the housing, and has a main surface facing the same direction as the main surface of the housing, and a holding part to be held by a user. The shooting instruction means is located on the second housing, and instructs the outer imaging section or the inner imaging section to shoot an image. The planar display section is located on the main surface of the second housing, and planarly displays an image.
In the above configuration, a user can take an image using the outer imaging section or the inner imaging section while holding the second housing.
According to the present invention, in an electronic device having a stereoscopic display section capable of displaying an image which is stereoscopically visible with naked eyes, a variety of functions can be realized. A user can take a stereoscopic image by using the outer imaging section while viewing the display screen of the stereoscopic display section from the front thereof. When the user views the display screen of the stereoscopic display section from the front thereof, the inner imaging section can take an image of the user's face.
(Structure of Game Apparatus)
Hereinafter, a game apparatus as according to one embodiment of the present invention will be described.
Initially, an external structure of the game apparatus 10 will be described with reference to
As shown in
(Description of Lower Housing)
Initially, a structure of the lower housing 11 will be described. As shown in
As shown in
As shown in
The operation buttons 14A to 14L are each an input device for making a predetermined input. As shown in
The analog stick 15 is a device for indicating a direction, and is provided to the left of the lower LCD 12 in an upper portion of the inner side surface of the lower housing 11. As shown in
Four buttons, that is, the button 14B, the button 14C, the button 14D, and the button 14E, which are positioned so as to form a cross shape, are positioned such that a thumb of a right hand with which the lower housing 11 is held is naturally positioned on the positions of the four buttons. Further, the four buttons and the analog stick 15 sandwich the lower LCD 12, so as to be bilaterally symmetrical in position with respect to each other. Thus, depending on a game program, for example, a left-handed person can make a direction instruction input by using these four buttons.
Further, the microphone hole 18 is provided on the inner side surface of the lower housing 11. Under the microphone hole 18, a microphone (see
As shown in
Further, as shown in
Further, as shown in
A rechargeable battery (not shown) acting as a power supply for the game apparatus 10 is accommodated in the lower housing 11, and the battery can be charged through a terminal provided on a side surface (for example, the upper side surface) of the lower housing 11.
(Description of Upper Housing)
Next, a structure of the upper housing 21 will be described. As shown in
As shown in
The screen of the upper LCD 22 is provided on the inner side surface (main surface) 21B of the upper housing 21, and the screen of the upper LCD 22 is exposed at an opening of the upper housing 21. Further, as shown in
The upper LCD 22 is a display device capable of displaying a stereoscopically visible image. Further, in the present embodiment, a left-eye image and a right-eye image are displayed by using substantially the same display area. Specifically, the upper LCD 22 may be a display device using a method in which the left-eye image and the right-eye image are alternately displayed in the horizontal direction in predetermined units (for example, every other line). Alternatively, a display device using a method in which the left-eye image and the right-eye image are alternately displayed may be used. Further, in the present embodiment, the upper LCD 22 is a display device capable of displaying an image which is stereoscopically visible with naked eyes. A lenticular lens type display device or a parallax barrier type display device is used which enables the left-eye image and the right-eye image, which are alternately displayed in the horizontal direction, to be separately viewed by the left eye and the right eye, respectively. In the present embodiment, the upper LCD 22 of a parallax barrier type is used. The upper LCD 22 displays, by using the right-eye image and the left-eye image, an image (a stereoscopically visible image) which is stereoscopically visible with naked eyes. That is, the upper LCD 22 allows a user to view the left-eye image with her/his left eye, and the right-eye image with her/his right eye by utilizing a parallax barrier, so that a stereoscopic image (a stereoscopically visible image) exerting a stereoscopic effect for a user can be displayed. Further, the upper LCD 22 may disable the parallax barrier. When the parallax barrier is disabled, an image can be displayed in a planar manner (it is possible to display a planar visible image which is different from a stereoscopically visible image as described above. Specifically, a display mode is used in which the same displayed image is viewed with a left eye and a right eye.). Thus, the upper LCD 22 is a display device capable of switching between a stereoscopic display mode for displaying a stereoscopically visible image and a planar display mode (for displaying a planar visible image) for displaying an image in a planar manner. The switching of the display mode is performed by the 3D adjustment switch 25 described below.
Two imaging sections (23a and 23b) provided on the outer side surface (the back surface reverse of the main surface on which the upper LCD 22 is provided) 21D of the upper housing 21 are generically referred to as the outer imaging section 23. The imaging directions of the outer imaging section (left) 23a and the outer imaging section (right) 23b are each the same as the outward normal direction of the outer side surface 21D. Further, these imaging sections are each designed so as to be positioned in a direction which is opposite to the normal direction of the display surface (inner side surface) of the upper LCD 22 by 180 degrees. Specifically, the imaging direction of the outer imaging section (left) 23a and the imaging direction of the outer imaging section (right) 23b are parallel to each other. The outer imaging section (left) 23a and the outer imaging section (right) 23b can be used as a stereo camera depending on a program executed by the game apparatus 10 (in this case, a stereoscopically visible image is displayed on the upper LCD by using a taken image from the outer imaging section (left) 23a and a taken image from the outer imaging section (right) 23b). Further, depending on a program, when any one of the two outer imaging sections (23a and 23b) is used alone, the outer imaging section 23 may be used as a non-stereo camera (in this case, a taken image from either of the outer imaging section (left) 23a and the outer imaging section (right) 23b is planarly displayed on the upper LCD in the planar display mode). Further, depending on a program, images taken by the two outer imaging sections (23a and 23b) may be combined with each other or may compensate for each other, thereby enabling imaging using an extended imaging range. Further, at least two images, among images taken by the two outer imaging sections (23a and 23b) and an image taken by the inner imaging section 24, may be combined to display a composite image on the upper LCD. For example, a stereoscopically visible image is displayed on the upper LCD by using images taken by the two outer imaging sections (23a and 23b), and an image (which may be a still image) taken by the inner imaging section 24 can be combined with the stereoscopically displayed image. More specifically, an image taken by the inner imaging section 24 is stored. Using the image as a texture of a virtual object, the virtual object is taken by a virtual camera for a left eye and a virtual camera for a right eye to generate a stereoscopically visible image in a virtual world. Then, the stereoscopically visible image in the virtual world is combined with a stereoscopically visible image in a real world, which is obtained by using images taken by the two outer imaging sections. In the present embodiment, the outer imaging section 23 is structured so as to include two imaging sections, that is, the outer imaging section (left) 23a and the outer imaging section (right) 23b. Each of the outer imaging section (left) 23a and the outer imaging section (right) 23b includes an imaging device, such as a CCD image sensor or a CMOS image sensor, having a common predetermined resolution, and a lens. The lens may have a zooming mechanism.
As indicated by dashed lines in
In the present embodiment, the outer imaging section (left) 23a and the outer imaging section (right) 23b are secured to the housing, and the imaging directions thereof cannot be changed.
Further, the outer imaging section (left) 23a and the outer imaging section (right) 23b are positioned to the left and to the right, respectively, of the upper LCD 22 (on the left side and the right side, respectively, of the upper housing 21) so as to be horizontally symmetrical with respect to the center of the upper LCD 22. Specifically, the outer imaging section (left) 23a and the outer imaging section (right) 23b are positioned so as to be symmetrical with respect to a line which divides the upper LCD 22 into two equal parts, that is, the left part and the right part. Further, the outer imaging section (left) 23a and the outer imaging section (right) 23b are positioned at positions which are reverse of positions above the upper edge of the screen of the upper LCD 22 and which are on the upper portion of the upper housing 21 in an opened state. Specifically, when the upper LCD 22 is projected on the outer side surface of the upper housing 21, the outer imaging section (left) 23a and the outer imaging section (right) 23b are positioned, on the outer side surface of the upper housing 21, at a position above the upper edge of the screen of the upper LCD 22 having been projected.
As described above, the two imaging sections (23a and 23b) of the outer imaging section 23 are positioned to the left and the right of the upper LCD 22 so as to be horizontally symmetrical with respect to the center of the upper LCD 22. Therefore, when a user views the upper LCD 22 from the front thereof, the imaging direction of the outer imaging section 23 can be made equal to the direction of the line of sight of the user. Further, the outer imaging section 23 is positioned at a position reverse of a position above the upper edge of the screen of the upper LCD 22. Therefore, the outer imaging section 23 and the upper LCD 22 do not interfere with each other inside the upper housing 21. Therefore, the upper housing 21 may have a reduced thickness as compared to a case where the outer imaging section 23 is positioned on a position reverse of a position of the screen of the upper LCD 22.
The inner imaging section 24 is positioned on the inner side surface (main surface) 21B of the upper housing 21, and acts as an imaging section which has an imaging direction which is the same direction as the inward normal direction of the inner side surface. The inner imaging section 24 includes an imaging device, such as a CCD image sensor and a CMOS image sensor, having a predetermined resolution, and a lens. The lens may have a zooming mechanism.
As shown in
As described above, the inner imaging section 24 is used for taking an image in the direction opposite to that of the outer imaging section 23. The inner imaging section 24 is positioned on the inner side surface of the upper housing 21 at a position reverse of the middle position between the left and the right imaging sections of the outer imaging section 23. Thus, when a user views the upper LCD 22 from the front thereof, the inner imaging section 24 can take an image of a face of the user from the front thereof. Further, the left and the right imaging sections of the outer imaging section 23 do not interfere with the inner imaging section 24 inside the upper housing 21, thereby enabling reduction of the thickness of the upper housing 21.
The 3D adjustment switch 25 is a slide switch, and is used for switching a display mode of the upper LCD 22 as described above. Further, the 3D adjustment switch 25 is used for adjusting the stereoscopic effect of a stereoscopically visible image (stereoscopic image) which is displayed on the upper LCD 22. As shown in
As shown in
The program executed by the game apparatus according to the present embodiment includes a program for displaying a stereoscopic photograph, and a program for displaying a stereoscopic CG image. The program for displaying a stereoscopic CG image is used for taking an image of a virtual space by means of a virtual camera for a left eye and a virtual camera for a right eye to generate a left-eye image and a right-eye image. The game apparatus according to the present embodiment adjusts the stereoscopic effect by changing a distance between the two virtual cameras in accordance with the position of the slider 25a of the 3D adjustment switch 25 when executing such a program.
The 3D indicator 26 indicates whether or not the upper LCD 22 is in the stereoscopic display mode. The 3D indicator 26 is implemented as a LED, and is lit up when the stereoscopic display mode of the upper LCD 22 is enabled. The 3D indicator 26 may be lit up only when the program processing for displaying a stereoscopically visible image is performed (namely, image processing in which a left-eye image is different from a right-eye image is performed in the case of the 3D adjustment switch being positioned between the first position and the second position) in a state where the upper LCD 22 is in the stereoscopic display mode. As shown in
Further, a speaker hole 21E is provided on the inner side surface of the upper housing 21. A sound is outputted through the speaker hole 21E from a speaker 43 described below.
(Internal Configuration of Game Apparatus 10)
Next, an internal electrical configuration of the game apparatus 10 will be described with reference to
The information processing section 31 is information processing means which includes a CPU (Central Processing Unit) 311 for executing a predetermined program, a GPU (Graphics Processing Unit) 312 for performing image processing, and the like. In the present embodiment, a predetermined program is stored in a memory (for example, the external memory 44 connected to the external memory I/F 33 or the internal data storage memory 35) inside the game apparatus 10. The CPU 311 of the information processing section 31 executes a shooting process (
To the information processing section 31, the main memory 32, the external memory I/F 33, the external data storage memory I/F 34, and the internal data storage memory 35 are connected. The external memory I/F 33 is an interface for detachably connecting to the external memory 44. The external data storage memory I/F 34 is an interface for detachably connecting to the external data storage memory 45.
The main memory 32 is volatile storage means used as a work area and a buffer area for (the CPU 311 of) the information processing section 31. That is, the main memory 32 temporarily stores various types of data used for the shooting process, and temporarily stores a program acquired from the outside (the external memory 44, another device, or the like), for example. In the present embodiment, for example, a PSRAM (Pseudo-SRAM) is used as the main memory 32.
The external memory 44 is nonvolatile storage means for storing a program executed by the information processing section 31. The external memory 44 is implemented as, for example, a read-only semiconductor memory. When the external memory 44 is connected to the external memory I/F 33, the information processing section 31 can load a program stored in the external memory 44. A predetermined process is performed by the program loaded by the information processing section 31 being executed. The external data storage memory 45 is implemented as anon-volatile readable and writable memory (for example, a NAND flash memory), and is used for storing predetermined data. For example, images taken by the outer imaging section 23 and/or images taken by another device are stored in the external data storage memory 45. When the external data storage memory 45 is connected to the external data storage memory I/F 34, the information processing section. 31 loads an image stored in the external data storage memory 45, and the image can be displayed on the upper LCD 22 and/or the lower LCD 12.
The internal data storage memory 35 is implemented as a non-volatile readable and writable memory (for example, a NAND flash memory), and is used for storing predetermined data. For example, data and/or programs downloaded through the wireless communication module 36 by wireless communication are stored in the internal data storage memory 35.
The wireless communication module 36 has a function of connecting to a wireless LAN by using a method based on, for example, IEEE 802.11.b/g standard. The local communication module 37 has a function of performing wireless communication with the same type of game apparatus in a predetermined communication method (for example, infrared communication). The wireless communication module 36 and the local communication module 37 are connected to the information processing section 31. The information processing section 31 can perform data transmission to and data reception from another device via the Internet by using the wireless communication module 36, and can perform data transmission to and data reception from the same type of another game apparatus by using the local communication module 37.
The acceleration sensor 39 is connected to the information processing section 31. The acceleration sensor 39 detects magnitudes of accelerations (linear accelerations) in the directions of the straight lines along the three axial (xyz axial) directions, respectively. The acceleration sensor 39 is provided inside the lower housing 11. In the acceleration sensor 39, as shown in
The RTC 38 and the power supply circuit 40 are connected to the information processing section 31. The RTC 38 counts time, and outputs the time to the information processing section 31. The information processing section 31 calculates a current time (date) based on the time counted by the RTC 38. The power supply circuit 40 controls power from the power supply (the rechargeable battery accommodated in the lower housing 11 as described above) of the game apparatus 10, and supplies power to each component of the game apparatus 10.
The I/F circuit 41 is connected to the information processing section 31. The microphone 42 and the speaker 43 are connected to the I/F circuit 41. Specifically, the speaker 43 is connected to the I/F circuit 41 through an amplifier which is not shown. The microphone 42 detects a voice from a user, and outputs a sound signal to the I/F circuit 41. The amplifier amplifies a sound signal outputted from the I/F circuit 41, and a sound is outputted from the speaker 43. The touch panel 13 is connected to the I/F circuit 41. The I/F circuit 41 includes a sound control circuit for controlling the microphone 42 and the speaker 43 (amplifier), and a touch panel control circuit for controlling the touch panel. The sound control circuit performs A/D conversion and D/A conversion on the sound signal, and converts the sound signal to a predetermined form of sound data, for example. The touch panel control circuit generates a predetermined form of touch position data based on a signal outputted from the touch panel 13, and outputs the touch position data to the information processing section 31. The touch position data represents a coordinate of a position, on an input surface of the touch panel 13, on which an input is made. The touch panel control circuit reads a signal outputted from the touch panel 13, and generates the touch position data every predetermined time. The information processing section 31 acquires the touch position data, to recognize a position on which an input is made on the touch panel 13.
The operation button 14 includes the operation buttons 14A to 14L described above, and is connected to the information processing section 31. Operation data representing an input state of each of the operation buttons 14A to 14I is outputted from the operation button 14 to the information processing section 31, and the input state indicates whether or not each of the operation buttons 14A to 14I has been pressed. The information processing section 31 acquires the operation data from the operation button 14 to perform a process in accordance with the input on the operation button 14.
The lower LCD 12 and the upper LCD 22 are connected to the information processing section 31. The lower LCD 12 and the upper LCD 22 each display an image in accordance with an instruction from (the GPU 312 of) the information processing section 31. In the present embodiment, the information processing section 31 causes the lower LCD 12 to display an image for operation, and causes the upper LCD 22 to display an image acquired from one of the imaging sections 23 or 24. That is, the information processing section 31 causes the upper LCD 22 to display a stereoscopic image (stereoscopically visible image) using a right-eye image and a left-eye image which are taken by the outer imaging section 23, and causes the upper LCD 22 to display a planar image taken by the inner imaging section 24, for example.
Specifically, the information processing section 31 is connected to an LCD controller (not shown) of the upper LCD 22, and causes the LCD controller to set the parallax barrier to ON or OFF. When the parallax barrier is set to ON in the upper LCD 22, a right-eye image and a left-eye image, (taken by the outer imaging section 23), which are stored in the VRAM 313 of the information processing section 31 are outputted to the upper LCD 22. More specifically, the LCD controller alternately repeats reading of pixel data of the right-eye image for one line in the vertical direction, and reading of pixel data of the left-eye image for one line in the vertical direction, thereby reading, from the VRAM 313, the right-eye image and the left-eye image. Thus, an image to be displayed is divided into the images for a right eye and the images for a left eye each of which is a rectangle-shaped image having one line of pixels aligned in the vertical direction, and an image, in which the rectangle-shaped left-eye image which is obtained through the division, and the rectangle-shaped right-eye image which is obtained through the division are alternately aligned, is displayed on the screen of the upper LCD 22. A user views the images through the parallax barrier in the upper LCD 22, so that the right-eye image is viewed by the user's right eye, and the left-eye image is viewed by the user's left eye. Thus, the stereoscopically visible image is displayed on the screen of the upper LCD 22.
The outer imaging section 23 and the inner imaging section 24 are connected to the information processing section 31. The outer imaging section 23 and the inner imaging section 24 each take an image in accordance with an instruction from the information processing section 31, and output data of the taken image to the information processing section 31. In the present embodiment, the information processing section 31 issues an instruction for taking an image to one of the outer imaging section 23 or the inner imaging section 24, and the imaging section which receives the instruction for taking an image takes an image and transmits data of the taken image to the information processing section 31. Specifically, in the present embodiment, a user selects the imaging section to be used through a touch operation using the touch panel 13. When the information processing section 31 (the CPU 311) detects that the imaging section is selected, the information processing section 31 instructs one of the outer imaging section 32 or the inner imaging section 24 to take an image.
The 3D adjustment switch 25 is connected to the information processing section 31. The 3D adjustment switch 25 transmits, to the information processing section 31, an electrical signal in accordance with the position of the slider 25a.
The 3D indicator 26 is connected to the information processing section 31. The information processing section 31 controls whether or not the 3D indicator 26 is to be lit up. In the present embodiment, the information processing section 31 lights up the 3D indicator 26 when the upper LCD 22 is in the stereoscopic display mode. The game apparatus 10 has the internal configuration as described above.
Next, an example of usage state of the game apparatus 10 will be described with reference to
Since the level line 64 indicating inclination of the game apparatus 10 relative to the horizontal direction is displayed on the screen of the upper LCD 22, the user can easily determine whether or not the game apparatus 10 is inclined relative to the horizontal direction. If a stereoscopic image is taken and saved (shot) with the game apparatus 10 being inclined relative to the horizontal direction, a right-eye image and a left-eye image taken by the outer imaging section 23 are saved in such a manner that each image is rotated about its center.
(Details of Shooting Process)
Next, a shooting process according to the present embodiment will be described in detail with reference to
The camera selection data 71 indicates an imaging section which is currently selected. The camera selection data 71 indicates whether the currently selected imaging section is the outer imaging section 23 or the inner imaging section 24.
The left-eye image position data 72 indicates a display position, on the upper LCD 22, of a left-eye image which is taken by the outer imaging section (left) 23a, and indicates coordinates of an image center of the left-eye image. The right-eye image position data 73 indicates a display position, on the upper LCD 22, of a right-eye image which is taken by the outer imaging section (right) 23b, and indicates coordinates of an image center of the right-eye image.
The acceleration data 74 indicates a latest acceleration detected by the acceleration sensor 39. Specifically, the acceleration data 74 indicates accelerations in the x-, y-, and z-axes directions, which are detected by the acceleration sensor 39. The acceleration sensor 39 detects an acceleration once every predetermined period, and transmits the detected acceleration to the information processing section 31 (CPU 311). The information processing section 31 updates the acceleration data 74 in the main memory 32 every time the acceleration sensor 39 detects an acceleration.
The display mode data 75 indicates whether the display mode of the upper LCD 22 is the stereoscopic display mode or the planar display mode.
The operation data 76 indicates operations performed on the respective operation buttons 14A to 14E and 14G to 14H, and the analog stick 15.
Next, the shooting process will be described in detail with reference to
First, in step S1, the information processing section 31 determines whether or not an imaging section switching instruction is issued. For example, icons indicating the outer imaging section 23 and the internal imaging section 24 are displayed on the screen of the lower LCD 12. The information processing section 31 determines whether or not an imaging section switching instruction is issued, based on a touch position detected by the touch panel 13. Specifically, when a touch position is detected, the information processing section 31 determines that an imaging section switching instruction is issued if the touch position is the display position of the icon of the imaging section that is different from the currently selected imaging section. For example, assuming that the outer imaging section 23 is currently selected, if the touch position detected by the touch panel 13 is the display position of the icon of the internal imaging section 24, the information processing section 31 determines that an imaging section switching instruction is issued. When the result of the determination is positive, the information processing section 31 executes a process of step S2. On the other hand, when the result of the determination is negative, the information processing section 31 executes a process of step S3. Only an icon for switching to the internal imaging section 24 may be displayed when the outer imaging section 23 is selected, and only an icon for switching to the outer imaging section 23 may be displayed when the internal imaging section 24 is selected. The information processing section 31 may determine that an imaging section switching instruction is issued, when any of the operation buttons 14A to 14E is pressed, instead of an operation on the screen of the lower LCD 12 using the touch panel 13.
In step S2, the information processing section 31 selects an imaging section. Specifically, the information processing section 31 switches the imaging section used for image taking to the imaging section touched in step S1, and updates the camera selection data 71 in the main memory 32. Next, the information processing section 31 executes a process in step S3.
In step S3, the information processing section 31 obtains a taken image. Specifically, the information processing section 31 obtains image data indicating an image taken by the currently selected imaging section, and stores the image data in the VRAM 313. For example, when the outer imaging section 23 is selected, the information processing section 31 obtains a right-eye image and a left-eye image which are taken by the outer imaging section 23. When the internal imaging section 24 is selected, the information processing section 31 obtains an image taken by the internal imaging section 24. Next, the information processing section 31 executes a process in step S4.
In step S4, the information processing section 31 detects an orientation of the game apparatus 10. Specifically, the information processing section 31 detects an orientation of the game apparatus 10 relative to the horizontal direction, with reference to the acceleration data 74 stored in the main memory 32. More specifically, the information processing section 31 detects an orientation of the game apparatus 10 based on the value of the acceleration in the x axis direction among the accelerations in the three-axis directions, which have been detected by the acceleration sensor 39.
In
Further, the acceleration sensor 39 detects an acceleration other than the gravity acceleration, in accordance with the movement of the game apparatus 10. For example, when the game apparatus 10 is intentionally shaken by the user, the game apparatus 10 cannot accurately determine the direction of the gravity acceleration. In this case, the magnitude of the acceleration detected by the acceleration sensor 39 is greater (or smaller) than the magnitude of the gravity acceleration. Accordingly, when the magnitude of the detected acceleration is greater than the magnitude of the gravity acceleration by a predetermined threshold, it is not necessary to detect the orientation of the game apparatus 10. Even when the game apparatus 10 is not intentionally shaken by the user, the game apparatus 10 slightly moves and therefore detects an acceleration other than the gravity acceleration. In this case, however, accelerations detected during a predetermined period are averagely equal to the gravity acceleration. Therefore, an orientation of the game apparatus 10 may be detected based on the accelerations detected during the predetermined period.
Step S4 is followed by step S5. In step S5, the information processing section 31 determines whether or not the outer imaging section 23 is selected. Specifically, the information processing section 31 determines whether or not the currently selected imaging section is the outer imaging section 23, with reference to the camera selection data 71. When the result of the determination is positive (when the outer imaging section 23 is selected), the information processing section 31 executes a process of step S6. On the other hand, when the result of the determination is negative, the information processing section 31 executes a process of step S9.
In step S6, the information processing section 31 determines whether or not the slider 25a of the 3D adjustment switch 25 is positioned at the lowermost position. The information processing section 31 determines whether or not the slider 25a is positioned at the lowermost position, based on a signal from the 3D adjustment switch 25. When the result of the determination is negative, the information processing section 31 turns on the 3D indicator 26, and then executes a process in step S7. On the other hand, when the result of the determination is positive, the information processing section 31 turns off the 3D indicator 26, and then executes a process of step S9.
In step S7, the information processing section 31 adjusts the positions of the right-eye image and the left-eye image. Specifically, the information processing section 31 updates the display positions of the right-eye image and the left-eye image, which are obtained in step S3, in the horizontal direction (the left-right direction of the upper LCD 22) in accordance with the position of the slider 25a of the 3D adjustment switch 25. More specifically, the information processing section 31 updates, according to the position of the slider 25a, the left-eye image position data 72 and the right-eye image position data 73 so that the left-eye image and the right-eye image move leftward and rightward, respectively, by a predetermined distance. Since the left-eye image position data 72 and the right-eye image position data 73 are thus updated in step S7, the left-eye image and the right-eye image are displayed in the adjusted positions when executing step S10 (step S23) described later. The amount of movement (the amount of shift) of the left-eye image and the right-eye image is determined according to the position of the slider 25a. When the slider 25a is present at the second position (see
In step S8, the information processing section 31 updates the display mode data 75 to set the display mode of the upper LCD 22 to the stereoscopic display mode (3D display mode). Next, the information processing section 31 executes a process in step S10.
On the other hand, in step S9, the information processing section 31 updates the display mode data 75 to set the display mode of the upper LCD 22 to the planar display mode (2D display mode). The process of step S8 is executed when it is determined in step S5 that the internal imaging section 24 is selected, or when it is determined in step S6 that the display mode is set to the planar display mode by the 3D adjustment switch 25. Next, the information processing section 31 executes a process of step S10.
In step S10, the information processing section 31 executes a display process of the upper LCD 22. The display process in step S9 will be described in detail with reference to
In step S21, the information processing section 31 determines whether or not the display mode of the upper LCD 22 is the planar display mode (2D display mode). Specifically, the information processing section 31 determines whether or not the display mode of the upper LCD 22 is the planar display mode, with reference to the display mode data 75. When the result of the determination is negative, the information processing section 31 executes a process of step S22. On the other hand, when the result of the determination is positive, the information processing section 31 executes a process of step S24.
In step S22, since the display mode of the upper LCD 22 is the stereoscopic display mode, the information processing section 31 sets the parallax barrier of the upper LCD 22 to ON. Specifically, the information processing section 31 transmits, to the LCD controller of the upper LCD 22, a signal to set the parallax barrier to ON. Then, the information processing section 31 executes a process of step S23.
In step S23, the information processing section 31 outputs the left-eye image and the right-eye image to the upper LCD 22. Specifically, the information processing section 31 outputs, to the upper LCD 22, the left-eye image and the right-eye image, the positions of which have been adjusted in step S7. That is, with reference to the left-eye image position data 72 updated in step S7, the information processing section 31 outputs, to the upper LCD 22, the left-eye image so that the left-eye image is displayed in the position indicated by the left-eye image position data 72. Likewise, with reference to the right-eye image position data 73 updated in step S7, the information processing section 31 outputs, to the upper LCD 22, the right-eye image so that the right-eye image is displayed in the position indicated by the right-eye image position data 73. More specifically, the information processing section 31 alternately repeats a process of outputting pixel data of the right-eye image for one line in the vertical direction, and a process of outputting pixel data of the left-eye image for one line in the vertical direction. Thereby, rectangle-shaped images, which are obtained by dividing the right-eye image in the vertical direction, and rectangle-shaped images, which are obtained by dividing the left-eye image in the vertical direction, are alternately arranged to be displayed on the upper LCD 22. In this case, the rectangle-shaped images of the right-eye image and the rectangle-shaped images of the left-eye image are arranged in accordance with the amount of shift between the two images (the difference between the x-direction coordinate values of the respective images, which has been updated in step S7). Since the two images are thus outputted to the upper LCD 22, when the user views the screen of the upper LCD 22 from the front thereof, the right-eye image and the left-eye image are viewed by the user's right and left eyes, respectively, as if the images are displayed in the positions adjusted in step S7.
When the slider 25a of the 3D adjustment switch 25 is present at the first position (a position slightly higher than the lowermost position) shown in
Turning to
In step S24, the information processing section 31 sets the parallax barrier of the upper LCD 22 to OFF because the display mode of the upper LCD 22 set to the planar display mode. Specifically, the information processing section 31 transmits, to the LCD controller of the upper LCD 22, a signal to set the parallax barrier to OFF. Then, the information processing section 31 executes a process of step S25.
In step S25, the information processing section 31 outputs one image to the upper LCD 22. Specifically, when the internal imaging section 24 is selected, the information processing section 31 outputs, to the upper LCD 22, an image taken by the internal imaging section 24 (the image obtained in step S3). When the outer imaging section 23 is selected, the information processing section 31 outputs, to the upper LCD 22, one of two images taken by the outer imaging section 23. Thereby, the image taken by the outer imaging section 23 or the internal imaging section 24 is planarly displayed on the upper LCD 22. Then, the information processing section 31 executes a process of step S26.
In step S26, the information processing section 31 causes the upper LCD 22 to display a reference line 63 and a level line 64. Specifically, the information processing section 31 causes the upper LCD 22 to display a horizontal reference line 63a which extends through the center of the screen of the upper LCD 22 and is parallel to the horizontal direction of the screen of the upper LCD 22, and a vertical reference line 63b which extends through the center of the screen of the upper LCD 22 and is parallel to the vertical direction of the screen of the upper LCD 22. Further, the information processing section 31 causes the upper LCD 22 to display a level line 64 having an inclination according to the orientation of the game apparatus 10, which has been detected in step S5.
When the user views the reference line 63 and the level line 64, the position, in the depth direction, of the screen of the upper LCD 22 in which the reference line 63 and the level line 64 are displayed, is substantially equal to the position of the screen. That is, when the user views the upper LCD 22, the user feels that the reference line 63 and the level line 64 are displayed on the screen of the upper LCD 22 shown in
As is apparent from the flowchart shown in
Then, the information processing section 31 ends the display process of the upper LCD, and executes a process of step S11 shown in
Turning to
In step S12, the information processing section 31 stores the taken images obtained in step S3. For example, the information processing section 31 permanently stores, in the internal data storage memory 35 (nonvolatile memory), the two images taken by the outer imaging section 23 or the image taken by the internal imaging section 24. Alternatively, the information processing section 31 temporarily stores, as still images, in the main memory 32 (volatile memory), the two images taken by the outer imaging section 23 or the image taken by the internal imaging section 24. The temporarily stored images are stored in a nonvolatile memory such as the internal data storage memory 35 in accordance with a predetermined instruction from the user (for example, user's pressing a button indicating storage to the nonvolatile memory). When a stereoscopic image is stored, in addition to image data of a right-eye image and a left-eye image, the amount of shift between the positions of these images (the amount of adjustment in step 7) is also stored. When the outer imaging section 23 is selected and the display mode of the upper LCD 22 is set to the 2D display mode (Yes in step S6), only one of the left-eye image and the right-eye image is displayed on the upper LCD 22, but these two images are stored in step S12. When the outer imaging section 23 is selected and the display mode of the upper LCD 22 is set to the 2D display mode, one of the two images taken by the outer imaging section 23 may be stored. Next, the information processing section 31 executes a process of step S13.
In step S13, the information processing section 31 determines whether or not the shooting process is ended. For example, the information processing section 31 determines whether or not the user performs a predetermined operation (for example, whether or not the user presses the select button 14J or the like). When the result of the determination is negative, information processing section 31 executes a process of step S1. When the result of the determination is positive, the information processing section 31 ends the shooting process shown in
As described above, according to the present embodiment, an image in the direction of the line of sight of a user can be taken as a stereoscopic image by the outer imaging section 23, and the user's face (the direction reverse to the direction of the line of sight of the user) can be taken by the internal imaging section 24.
In the present embodiment, the upper LCD 22 for displaying a stereoscopically visible image (stereoscopic image) is provided on the inner side surface of the upper housing 21, and the lower LCD 12 for displaying an image in a planar manner is provided on the inner side surface of the lower housing 11. The operation buttons 14 and the touch panel 13, which are operated by the user, are provided on the lower housing 11. Thus, the lower housing 11 is configured so as to be held and operated by the user, which allows the user to easily operate the game apparatus 10 and easily view the upper LCD 22 provided on the upper housing 21. Even when the user is operating the game apparatus 10, the user's hand and fingers do not hang over the upper LCD 22 on which a stereoscopic image is displayed so as to be visible with naked eyes, the user can easily view the upper LCD 22. When a stereoscopic image is displayed on the upper LCD 22, the user focuses his/her eyes on a position in front of or deeper than the screen of the upper LCD 22, and thus the user recognizes the image displayed on the upper LCD 22 as a stereoscopic image. In this case, if the user's hand or fingers (including other objects) are present in front of or in the vicinity of the screen of the upper LCD 22, it is difficult for the user to focus his/her eyes on the screen. In the game apparatus 10 according to the present invention, however, the lower housing 11 which is held and operated by the user is provided. Therefore, the user's hand and fingers do not hang over the upper LCD 22 provided on the upper housing 21, and thus the user can easily focus his/her eyes on the screen.
Further, the outer imaging section 23 (stereo camera) is provided on the outer side surface reverse to the inner side surface of the upper housing 21, and the internal imaging section 24 is provided on the inner side surface of the upper housing 21. The two imaging sections (the outer imaging section (left) 23a and the outer imaging section (right) 23b) of the outer imaging section 23 are aligned so as to be parallel to the horizontal direction of the screen of the lower LCD 12. The internal imaging section 24 is provided at a position reverse of the middle position between the two imaging sections of the outer imaging section 23. The above-described configuration allows the user to take images in the direction of the line of sight of the user, by using the outer imaging section 23, while viewing the screen of the upper LCD 22 from the front thereof. A right-eye image and a left-eye image taken by the outer imaging section 23 are displayed on the upper LCD 22, and the right-eye image and the left-eye image are viewed by the user's right eye and left eye, respectively, through the parallax barrier. Thereby, a stereoscopically visible image is displayed on the upper LCD 22. On the upper LCD 22, the stereoscopic image taken by the outer imaging section 23 is displayed in real time, and the user, who is viewing the stereoscopic image displayed on the screen of the upper LCD 22, can store the stereoscopic image by pressing the shutter button (the L button 14G or the R button 14H). When the user, who is viewing the stereoscopic image displayed on the upper LCD 22 (who is viewing the screen of the upper LCD 22 from the front thereof), selects the internal imaging section 24, the user views the internal imaging section 24 from the front thereof. When the user, who is viewing the internal imaging section 24, presses the shutter button (the L button 14G or the R button 14H) of the imaging section, the user can shoot his/her face from the front thereof (the user can store the image). Accordingly, when the user switches the imaging section from the outer imaging section 23 to the internal imaging section 24, the user can shoot his/her face right from the front without the necessity of moving the position of his/her face to a correct position (a position in front of the internal imaging section 24). It should be noted that selection of the imaging section to be used (the outer imaging section 23 or the internal imaging section 24) may be performed by a selection operation using the respective operation buttons 14A to 14E provided on the lower housing 11.
The area of the screen of the upper LCD 22 is set to be greater than the area of the screen of the lower LCD 12. Specifically, the screen of the upper LCD 22 is horizontally longer than the screen of the lower LCD 12. That is, a rate of the horizontal width in the aspect ratio (horizontal to vertical ratio) of the screen of the upper LCD 22 is set to be greater than a rate of the horizontal width in the aspect ratio of the screen of the lower LCD 12. Thereby, the user's attention can be drawn to the screen of the upper LCD 22. Further, the screen of the lower LCD 12 being greater than the screen of the upper LCD 22 can cause the user to recognize that a manner of holding the game apparatus 10, in which the longitudinal direction of each housing coincides with the horizontal direction (the manner of holding shown in
Further, the resolution of the upper LCD 22 in the horizontal direction is set to be higher than the resolution of the lower LCD 12 in the horizontal direction. Thereby, even when a stereoscopic image is displayed on the upper LCD 22 and a planar image is displayed on the lower LCD 12, the user' attention can be drawn to the upper LCD 22. That is, when the upper LCD 22 and the lower LCD 12 have the same resolution in the horizontal direction, if a stereoscopic image is displayed on the upper LCD 22 while a planar image is displayed on the lower LCD 12, the resolutions of the right-eye image and the left-eye image in the horizontal direction, which are displayed on the upper LCD 22, are reduced as compared to that of the planar image displayed on the lower LCD 12. In this case, the image displayed on the lower LCD 12 becomes clearer than the image displayed on the upper LCD 22, which results in a difficulty in drawing the user's attention to the upper LCD 22.
For example, the resolution of the upper LCD 22 in the horizontal direction may be set to about 1.5 times or more (preferably, two times or more) of the resolution of the lower LCD 12 in the horizontal direction. When a stereoscopic image is displayed on the upper LCD 22, since a right-eye image and a left-eye image are displayed on one screen, the resolution, in the horizontal direction, of the stereoscopic image caused by the right-eye image and the left-eye image is substantially halved. When the upper LCD 22 and the lower LCD 12 have the same resolution in the horizontal direction, the screen of the lower LCD 12 becomes clearer than that of the upper LCD 22. In this case, the user's attention is drawn not to the upper LCD 22 but to the lower LCD 12. However, when the resolution of the upper LCD 22 in the horizontal direction is made higher than that of the lower LCD 12, the user's attention can be drawn to the upper LCD 22 on which a stereoscopic image is displayed.
Further, the lower housing 11 to be held by the user is provided, and the 3D adjustment switch 25 for switching the display mode of the upper LCD 22 is provided on the upper housing 21. Since the 3D adjustment switch 25 is provided on the upper housing 21, it is possible to avoid a situation that the user operates the 3D adjustment switch 25 accidentally while viewing the upper LCD 22 and thereby the display mode of the upper LCD 22 is changed. Since the 3D adjustment switch 25 also has a function of adjusting the distance between the two images in the horizontal direction, it is possible to avoid a situation that the amount of shift between the images is changed accidentally while the user is viewing the upper LCD 22 and thereby the manner in which the stereoscopic image is visible is changed.
Further, the 3D indicator 26 indicating whether or not the upper LCD 22 is in the stereoscopic display mode is provided on the upper housing 21. Thereby, the user can easily determine whether or not the upper LCD 22 is in the stereoscopic display mode. The indicators, other than the 3D indicator 26, for indicating the state of the game apparatus 10, such as the first LED 16A indicating ON/OFF of the power supply and the second LED 16B indicating presence/absence of wireless communication, are provided on the lower housing 11. The indicators other than the 3D indicator 26 may be provided on the outer side surface or the side surface of the upper housing 21, or on any of the inner side surface, the outer side surface, and the side surface of the lower housing 11. That is, the indicators other than the 3D indicator 26 may be provided on a surface different from the inner side surface of the upper housing 21. Thus, the indicators other than the 3D indicator 26 are provided in a position out of the sight of the user when the user views the upper LCD 22 from the front thereof, thereby preventing reduction in the viewability of the stereoscopic image displayed on the upper LCD 22.
(Alternative Embodiments)
Hereinafter, alternative embodiments of the game apparatus 10 will be described. In an another embodiment, while a stereoscopic image is displayed on the upper LCD 22, a user touches the lower LCD 12 (touch panel 13) to designate a position on the upper LCD 22. Specifically, if a touch operation is performed on the touch panel 13 while a stereoscopic image is displayed on the upper LCD 22, a position corresponding to the touch position is displayed on the upper LCD 22. A cursor may be displayed at the designated position on the upper LCD 22. Then, for example, an image is scrolled so that the designated position on the upper LCD 22 comes in the center of the screen. In another example, an amount of shift between a right-eye image and a left-eye image may be adjusted so that the position of an object in the depth direction, which object is displayed at the designated position on the upper LCD 22, is in the vicinity of the display screen.
x22=x12·A
y22=y12·B
where x22 is a coordinate value in the horizontal direction of the upper LCD 22, x12 is a coordinate value in the horizontal direction of the lower LCD 12, y22 is a coordinate value in the vertical direction of the upper LCD 22, y12 is a coordinate value in the vertical direction of the lower LCD 12, A is a ratio of the length of the lower LCD 12 in the horizontal direction to the length of the upper LCD 22 in the horizontal direction, and B is a ratio of the length of the lower LCD 12 in the vertical direction to the length of the upper LCD 22 in the vertical direction.
In this way, a position on the upper LCD 22 is indirectly designated by designating a position on the lower LCD 12 without directly designating the position on the upper LCD 22. Thereby, even when a stereoscopic image is displayed on the upper LCD 22, the user can easily designate the displayed stereoscopic image. That is, when the user tries to directly touch the stereoscopic image, since the user's eyes are not focused on the screen, the user might try to touch the space in front of the screen or a position deeper than the screen. Accordingly, it is difficult for the user to directly designate a position on the upper LCD 22. Further, when the user tries to directly touch the stereoscopic image with his/her finger or the touch pen 28, the finger or the touch pen 28 comes in the sight of the user, which interferes with the user's stereoscopic viewing. However, when the user designates a position on the lower LCD 12 to indirectly designate the corresponding position on the upper LCD 22, the user's stereoscopic viewing is not interfered with, and thus the user can easily designate the position on the upper LCD 22.
For example, when images taken by the outer imaging section 23 and stored are displayed as a stereoscopic image, the stereoscopic image may be scrolled so that a designated position on the upper LCD 22 comes in the center of the screen of the upper LCD 22.
When a stereoscopic image taken by the outer imaging section 23 is displayed in real time, the positions of a right-eye image and a left-eye image in the horizontal direction may be adjusted based on a designated position on the upper LCD 22. For example, in
In another embodiment, the game apparatus 10 may includes, as orientation detection means, an angular velocity sensor instead of (or in addition to) the acceleration sensor 39. The angular velocity sensor may detect angular velocities around three axes of xyz, or an angular velocity around the z axis. It is possible to calculate an orientation (inclination relative to the horizontal direction) of the game apparatus 10 by the angular velocity sensor. Specifically, an angle of rotation of the game apparatus 10 around the z axis can be calculated by integrating, with time, the angular velocity around the z axis, which is detected by the angular velocity sensor. In this case, initialization of the angular velocity sensor must be performed in a state where the game apparatus 10 is not inclined relative to the horizontal direction.
When the game apparatus 10 includes, as orientation detection means, an acceleration sensor 39 and an angular velocity sensor, an orientation of the game apparatus 10 can be detected more accurately by using an acceleration and an angular velocity. That is, when the game apparatus 10 is in a static state, an orientation of the game apparatus 10 can be accurately detected based on an acceleration detected by the acceleration sensor 39. When the game apparatus 10 is moving (when an acceleration other than a gravity acceleration is generated), an orientation of the game apparatus 10 is detected based on an angular velocity detected by the angular velocity sensor. Detection of an angular velocity by the angular velocity sensor has a certain amount of error, and an error in calculated rotation angle is accumulated with time. However, when the game apparatus 10 is in a static state, an orientation of the game apparatus 10 can be accurately detected based on an acceleration detected by the acceleration sensor 39. Therefore, the rotation angle calculated by the angular velocity sensor in the static state can be initialized. Thus, by using the acceleration sensor 39 and the angular velocity sensor, an orientation of the game apparatus 10 can be detected more accurately regardless of whether the game apparatus 10 is moving or not.
In another embodiment, the game apparatus 10 executes a predetermined game. For example, a predetermined game program is stored in the internal data storage memory 35 or the external memory 44, and a game is performed when the game apparatus 10 executes the program. For example, in this game, an object existing in a 3D virtual space is stereoscopically displayed on the upper LCD 22, and the object moves in the 3D virtual space in accordance with an instruction from the user.
In the game apparatus 10 according to the above-described embodiment, the cross button 14A is provided beneath the analog stick 15, and the lower LCD 12 in which an image is planarly displayed is provided beneath the upper LCD 22. When performing a game in which an image is planarly displayed on the lower LCD 12, not the analog stick 15 but the cross button 14A may be used. Since the game apparatus 10 is provided with the two operation means, the two operation means can be appropriately used between the case where an image is stereoscopically displayed and the case where an image is planarly displayed. Further, since the analog stick 15 is provided at the upper side while the cross button 14A is provided at the lower side, it is suggested to the user that the upper LCD 22 corresponds to the analog stick 15 while the lower LCD 12 corresponds to the cross button 14A.
In the present embodiment, an amount of shift between a left-eye image and a right-eye image is adjusted in accordance with the position of the slider of the 3D adjustment switch 25. In an another embodiment, the 3D adjustment switch 25 may be used for only ON/OFF of stereoscopic display, and an amount of shift between a left-eye image and a right-eye image may be adjusted by a touch operation to the lower LCD 12. For example, when the outer imaging section 23 is selected, if the slider of the 3D adjustment switch 25 is present between the first position and the second position, a left-eye image and a right-eye image, which are taken by the outer imaging section 23, are displayed on the upper LCD 22, and thereby a stereoscopic image is displayed. In this case, for example, an adjustment bar for adjusting the amount of shift between the left-eye image and the right-eye image may be displayed on the lower LCD 12. Then, the user touches a slider of the adjustment bar and slides the slider in a predetermined direction (for example, horizontal direction) to adjust the amount of shift between the left-eye image and the right-eye image. Further, semi-transparent left-eye image and right-eye image may be superimposed on one another and displayed on the lower LCD 12. In this case, since the user can view the left-eye image and the right-eye image with both eyes, the user can easily recognize the positional relationship between the two images. Then, for example, the user may touch and move the left-eye image or the right-eye image to adjust the amount of shift between the two images. Further, in another embodiment, an amount of shift between a left-eye image and a right-eye image may be adjusted in accordance with an operation using the operation buttons 14A to 14E. For example, an amount of shift between a left-eye image and a right-eye image in the horizontal direction may be adjusted when the rightward or leftward direction button of the cross button 14A is pressed.
In the present embodiment, the upper LCD 22 is provided on the inner side surface of the upper housing 21, and the lower LCD 12 is provided on the inner side surface of the lower housing 11. In another embodiment, the upper LCD 22 and the lower LCD 12 may be provided on the outer side surfaces of the respective housings. That is, the upper and lower LCDs 22 and 12 may be configured so that, when the two housings are in their opened states, the surfaces of the housings on which the LCDs are provided face in the same direction.
In the present embodiment, the upper housing 21 and the lower housing 11 are foldable. In another embodiment, the upper housing 21 or the lower housing 11 may be slidable in the vertical direction, and the upper LCD 22 and the lower LCD 12 may be connected to each other in the vertical direction when the housing is slid.
The positions of the operation buttons 14 in the above-described embodiment are merely example, and the operation buttons 14 may be arranged in any way. Further, in the above-described embodiment, since the 3D adjustment switch 25 is provided on the inner side surface of the upper housing 21, the user can visually recognize the 3D adjustment switch 25 even when the user views the upper LCD 22. In another embodiment, the 3D adjustment switch 25 may be provided on the outer side surface or the side surface of the upper housing 21, or it may be provided on the lower housing 11.
In the above-described embodiment, an orientation of the game apparatus 10 is detected by using the acceleration sensor 39 and/or the angular velocity sensor. Then, as shown in
In the above-described embodiment, the upper LCD 22 is a parallax barrier type liquid crystal display device, and the parallax barrier is ON/OFF controlled to switch the display mode between the stereoscopic display mode and the planar display mode. In another embodiment, a stereoscopic image and a planar image may be displayed by using, for example, a lenticular type liquid crystal display device. Also in the case of using the lenticular type display device, an image is stereoscopically displayed by vertically dividing two images taken by the outer imaging section 23 into rectangle-shaped images and then alternately arranging the rectangle-shaped images. Also in the case of using the lenticular type display device, an image taken by the internal imaging section 24 is planarly displayed by causing the user to view this image with both eyes. That is, also in the lenticular type liquid crystal display device, the user is caused by view the same image with right and left eyes by vertically dividing the same image into rectangle-shaped images and alternately arranging the rectangle-shaped images. Thus, the image taken by the internal imaging section 24 can be displayed as a planar image.
In another embodiment, the present invention is not limited to a game apparatus, but may be applied to any hand-held electronic device, such as a PDA (Personal Digital Assistant), a mobile phone, a personal computer, a camera, or the like. For example, a mobile phone may be provided with: a display section for displaying a stereoscopically visible image, which is provided on a main surface of a housing; an imaging section provided on the main surface; and a stereo camera provided on a rear surface of the housing, which is reverse to the main surface.
In the above-described embodiment, the information processing section 31 of the game apparatus 10 executes a predetermined program to perform the process steps in the above-described flowcharts. In another embodiment, some or all of the above-described processes may be performed by a dedicated circuit provided to the game apparatus 10.
The present invention can be used as a game apparatus, for example.
Number | Date | Country | Kind |
---|---|---|---|
2010122077 | May 2010 | JP | national |
2010 135535 | Jun 2010 | JP | national |
2010 185333 | Aug 2010 | JP | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/JP2011/001186 | 3/1/2011 | WO | 00 | 4/3/2012 |