Video game system for capturing images and applying the captured images to animated game play characters

Information

  • Patent Grant
  • 6677967
  • Patent Number
    6,677,967
  • Date Filed
    Wednesday, February 21, 2001
    23 years ago
  • Date Issued
    Tuesday, January 13, 2004
    20 years ago
Abstract
A frame of still picture data is captured at an instant specified by a user from video signals supplied from a given video source, such as a television receiver, a video camera, etc., and the image data is displayed. When the user specifies an area of image to be cut out from the displayed still picture, the image data in the specified area is cut out and recorded as a cutout image. Each cutout image recorded is displayed in the form of an icon. When any of the icons is selected by the user, the corresponding cutout image data is read and pasted in a part to be changed in the original image data. Thus an image can be easily created by user's choice.
Description




BACKGROUND OF THE INVENTION




1. Field of the Invention




The present invention relates to image display devices such as home videogame systems, and more particularly to an image display device which captures image data supplied from a video source, such as a television receiver, a video camera, etc., and combines the captured image data with predetermined image data to create and display a combined image. The invention also relates to such a system which processes an image obtained by the capture process before applying the processed image to the predetermined image.




2. Description of the Background Art




Conventional image processing software for personal computers allows a user to cut out a partial image from a still picture and superpose it onto another still picture to create a desired image. Such conventional image processing software, however, generally requires the user to perform complicated operations to accomplish this result. For example, it was generally necessary in one such prior arrangement for a user skilled in operating personal computers to perform a sequence of complex image creating process steps by, for example, specifying each action corresponding to the processing procedure through operating system functions. In such conventional image processing techniques, for example, a user might be required to perform the operations of:




reading two still pictures and storing the information into a given storage location on hard disk;




specifying a cutout area on one of the still pictures;




temporarily saving the cutout image data in a memory;




specifying a display area on the other still picture for superposing the cutout image data;




writing the cutout image data into the storage region corresponding to that display area;




discarding the unwanted image after cut out; and




displaying the superposed image.




It was generally difficult for users untrained in operating personal computers to create desired images using such steps. Further, even for an operator skilled in operating personal computers, the operation was generally so complicated that creating desired composite images took a long time. Accordingly, it was generally quite difficult for children to create such computer graphic images in the course of playing a game, for example. Moreover, generally only a limited number of images could be combined at a time, and displaying the combined image as a moving or animated picture was difficult.




SUMMARY OF THE INVENTION




The present invention solves this problem by providing an image creating arrangement capable of creating a desired composite image by very easy and simple operations.




The present invention also provides an image creating device capable of displaying a created image and animating it according to a predetermined operating program.




One aspect of the present invention is directed to an image creating arrangement for capturing image data supplied from a video source and combining the captured image data with predetermined image data to create a desired image, and displaying the desired image on a display device. The desired image can be animated in response to user input controls and/or program instructions. An example image creating device implementation comprises:




an operation input arrangement used to allow a user to input instructions for image creation;




an initial image data store temporarily storing initial image data;




a captured image store temporarily storing captured image data;




a cut out image store temporarily storing image data cut out from the captured image data;




a display image store storing image data for the desired image to be displayed;




a program store storing a program for image creation; and




a processor performing an image creation operation on the basis of the stored program;




wherein the stored program includes,




a first write program instruction sequence responsive to operation of the operation input arrangement, for capturing the image data supplied from a video source and writing the image data into the captured image store as captured image data,




a first read/display program instruction sequence for reading the stored captured image data and displaying the captured image on a display device,




a cutout program instruction sequence for, when the operation input arrangement is operated to specify an image area to be cut out from the captured image displayed on the display device, temporarily storing the image data corresponding to the specified area into the cutout image store as cutout image data,




a second read/display program instruction sequence for reading the cutout image data stored in the cutout image store and displaying the cutout image data as an icon or other small image in part of screen of the display device,




a second write program instruction sequence for reading the initial image data stored in the initial image data store and writing the initial image data into the display image store,




a third write program instruction sequence for, when the operation input arrangement is operated to select the displayed cutout image (e.g., in order to combine and display the cutout image in an image display position corresponding to a part of the displayed initial image), writing the corresponding cutout image data stored in the cut out image store into a corresponding region of the display image store, and




a third read/display program instruction sequence for reading the combined image data stored in the display image store in which the cutout image data is written and displaying the combined image data on the display device.




The preferred implementation of the invention provides many advantages. For example, it is not necessary to newly create an initial image, since an initial image for use as a base for image creation can be prepared in advance. The process of pasting a cutout image into a specified position can be accomplished by the easy operation of specifying the position on the displayed original image. Moreover, as it is possible to capture the image while viewing the displayed object produced from video signal supplied from a video source, an image can be captured not only from a still picture but also from a moving picture. Further, an image in a desired area can be cut out by an easy operation of specifying the cutout area on the captured image. Moreover, an image to be pasted can be easily selected since cutout images can be recorded as icons. In this way, a desired image can be created from previously prepared initial original image(s) by very easy operations, so that even users unskilled in operating personal computers and children can easily create images.




According to a further aspect of the invention, an image creating device may provide a program store further containing a correcting program for applying correction and/or editing processing to the initial image data in a peripheral part in the cutout area specified by the operation input portion before the cutout image data is written into the second cutout image store. Thus, when an image in a desired area is cut out from the captured image, its peripheral part can be corrected. Accordingly, the original image and the cutout image can be combined without causing unnatural image certificates or other impressions at the boundary between them.




According to a further aspect provided by the invention, an image creating device cutout program instruction sequence cuts out the image data in a rectangular area specified by the operation input arrangement, and the correcting program corrects the cutout image data in the specified rectangular area by shading off a peripheral part around a shape defined by excluding the four corners of the specified rectangular area.




According to a still further aspect, the program store further contains a motion giving (e.g., animation) program for applying motion to individual parts of the initial image. When reading the combined image data stored in the display image store (e.g., on the basis of the third read/display program instruction sequence), the processor processes the combined image data on the basis of the motion program to change position and motion of the individual parts of the combined image. It is thus possible to animate the image created by combining the captured image with the original image according to a predetermined operating program, which enables more enjoyable image creation.




According to a further aspect, the cutout program temporarily stores the cutout image data corresponding to the specified area into the cutout image store each time the operation input portion is operated to specify an area of image to be cut out. The second read/display program instruction sequence reads the plurality of cutout images stored in the cutout image store and separately displays the plurality of cutout image data parts of the screen of the display device. As stated above, according to this aspect, a plurality of cutout images stored in the second temporarily storing portion are read and displayed in the display device. This allows the user to see already recorded cutout images all together to easily select a cutout image to be pasted on the original image.




According to a further aspect, the initial image store stores the initial image data in the form of 3D polygon data composed of a group of a plurality of polygons to display an image of a moving object as the initial image. The program store further contains a motion applying program instruction sequence for applying motion to individual parts of the object.




The second write program instruction sequence may create display image data on the basis of the plurality of polygon data stored in the original image data storing portion, and may write the display image data into the display image store. The third write program instruction sequence writes the cutout image data stored in the cutout image store into a storage region in the display image store corresponding to a polygon position of the part to be changed which was specified by operation of the operation input arrangement. The processor reads the combined image data stored in the display image store to display an image obtained by combining the image based on the polygon data with the cutout image in the display device, and writes image data changed so that the individual parts based on the polygon data can be moved at a predetermined cycle or rate on the basis of the motion applying program into the third temporary store.




As stated above, according to this aspect, a three-dimensional image can be created quite easily. Further, the created image can be displayed as an animated image moving according to a predetermined operating program.




According to a further aspect, the program store further contains a display control program instruction sequence for controlling display of the moving object. The display control program includes a program for displaying a background image for the moving object, and a program responsive to operation of the operation input arrangement, for applying changes different from original motion to the motion of the moving object. A background image can be automatically created when creating three-dimensional images.











BRIEF DESCRIPTION OF THE DRAWINGS




These and other objects, features, aspects and advantages of the present invention will become more apparent from the following detailed description of the present invention when taken in conjunction with the accompanying drawings.





FIG. 1

is a perspective view showing an example external appearance of an example image display (e.g., video game) system according to an embodiment of the present invention;





FIG. 2

is a block diagram showing an example electrical block diagram of the video game system shown in

FIG. 1

;





FIG. 3

is a diagram showing an example memory map of a cartridge


30


and/or a disk


45


of an example of an external storage medium;





FIG. 4

is a diagram showing an example memory map of a RAM


15


provided in the game system


10


;





FIG. 5

is a flowchart showing an example main routine used to control operation of an image display device of an example embodiment of the present invention;





FIG. 6

is a subroutine flowchart showing an example image capturing/image selecting process;





FIG. 7

is a subroutine flowchart showing an example image confirming process;





FIG. 8

is a subroutine flowchart showing an example cutout area selecting process;





FIG. 9

is a subroutine flowchart showing an example cutout image confirming process;





FIG. 10

is a subroutine flowchart showing a cutout image saving process.





FIG. 11

is a diagram showing an example display provided when a character is selected;





FIG. 12

is a diagram showing an example display used in the image selecting process;





FIG. 13

is a diagram showing an example displayed used in the image confirming process;





FIG. 14

is a diagram showing an example display used in the cutout image area selecting process;





FIG. 15

is a diagram showing an example display used in the cutout image saving process;





FIG. 16

is a diagram showing an example display used in a cutout image pasting process;





FIG. 17

is a diagram showing an example display shown while a desired image is being created (e.g., edited) by using cutout image; and





FIG. 18

is a diagram showing an example of a final combined created image providing a character that can be animated.











DETAILED DESCRIPTION OF THE PREFERRED EXAMPLE EMBODIMENTS





FIG. 1

is a perspective view showing the external appearance of an example image creating (display) system


5


according to an embodiment of the present invention. In this example, system


5


comprises a home video game machine system. In

FIG. 1

, the image creating system


5


of this embodiment captures video signals supplied from a given video source (a television receiver, a video camera, a video cassette recorder, etc.) and processes data by using the image based on the captured video signals to create an image suited to user's preference.




In this example, system


20


includes a main unit of the video game system (hereinafter referred to as a game machine)


10


. Formed on the top surface of the housing


11


of the game machine


10


are an insertion slot


111


for inserting a cartridge


30


, and a power switch


112


and a reset switch


113


on the side. The cartridge insertion slot


111


has a cartridge connector


12




c


(see

FIG. 2

) provided on its inner side. Provided on the front panel of the game machine


10


are controller connectors


121


,


122


,


123


and


124


for connection with hand-controllers


20


. An extension connector


12




d


for connection with a disk drive device


40


is provided on the bottom of the game machine


10


.




The controller


20


has a housing


21


with a plurality of (three in the drawing) handles. Provided on the top of the controller housing


21


are:




a direction specifying switch


22


for specifying direction of movements of characters or objects operated by a player,




an analog stick (also called a 3D stick or a joy stick)


23


for analogously specifying the moving direction, and




a group of operating switches


24


including a reset/start button


25


.




The group of operating switches


24


includes switches


24




a


to


24




f


, which are used to specify actions of the player characters or player objects in a game mode, and to select various kinds of commands in an image creating input mode. Side switches


26


L and


26


R for requesting other actions are provided on the upper side of the housing


21


.




Provided on the top of the housing


41


of the disk drive


40


is a connector


42


to be connected to the extension connector


12




d


. A disk insertion slot


43


is formed on the front side of the housing


41


. Inserted into the disk insertion slot


43


is a writable/readable disk-like recording (or storing) medium (hereinafter referred to as a disk)


45


, such as a magnetic disk, magneto-optic disk, etc.




A display device


50


like a household television receiver is connected to the game machine


10


.




Game System Example Block Diagram





FIG. 2

is a block diagram showing the electric


11


configuration of the video game machine system


10


shown in FIG.


1


. In

FIG. 2

, the game machine


10


(which may be, for example, a Nintendo 64 video game system manufactured by Nintendo since 1996) includes an information processing unit (hereinafter referred to as a CPU)


13


. The CPU


13


performs processing for image creation or game play on the basis of program instructions stored in the memory cartridge


30


or the disk-like storage medium


45


. A bus control circuit


141


included in a coprocessor (hereinafter referred to as an RCP)


14


is connected to the CPU


13


. The RCP


14


includes a first image processing unit (a reality signal processor, hereinafter referred to as RSP)


142


and a second image processing unit (reality display processor, hereinafter referred to as RDP)


143


. The RSP


142


is a processor for executing signal processing such as polygon coordinate transformation, light source processing, etc. The RDP


143


is a processor for converting polygon data into image data that can be displayed and stored in a data format within frame memory included in the RAM


15


. The RSP


142


, RDP


143


, and RAM


15


are connected to the CPU


13


through the bus control circuit


141


. The RAM


15


will be described later in greater detail referring to FIG.


4


.




Also connected to the bus control circuit


141


are connectors


12




c


,


12




d


, and a controller control circuit


16


. The memory cartridge (hereinafter referred to as a cartridge)


30


is detachably connected to the connector


12




c


. Connected to the connector


12




d


is the disk drive


40


. The connectors


121


,


122


,


123


and


124


are connected to the controller control circuit


16


. These connectors


121


to


124


are connected to controllers


20


of the respective users (or players). Further, an audio signal generating circuit


16




a


and a video signal generating circuit


16




v


are connected to the bus control circuit


141


. The audio signal generating circuit


16




a


is connected to a speaker


51


in the display device


50


through an output connector


17




a


. The video signal generating circuit


16




v


is connected to a CRT display


52


in the display device


50


through an output connector


17




v.






Example Memory Maps





FIG. 3

shows a storage region (memory) map of the cartridge


30


and/or disk


45


as an example of an external storage medium. When the cartridge


30


is used as the external storage medium, it includes a non-volatile memory (hereinafter referred to as ROM)


31


for fixedly storing program data and a writable/readable memory (hereinafter referred to as RAM)


32


. The ROM


31


includes




an identification (ID) information storage region


311


,




an OS information storage region


312


,




an image creating program storage region


313


,




an animation display program (game program) storage region


314


,




an audio generating program, including sound data program storage region


315


, and




a graphic data storage region


316


.




Stored in the ID information storage region


311


in the ROM


31


are the serial No. of the external storage medium, game version data, etc.




Stored in the OS (operating system) storage region


312


are information such as graphic microcode, sound microcode, CPU library, etc.




Stored in the image creating program storage region


313


are programs for realizing the operation shown in the flowcharts in

FIG. 5

to

FIG. 10

described later, such as:




write/read/display control program instructions,




polygon coordinate calculating program instructions,




icon display program instructions,




parts select/display program instructions,




color/brightness control program instructions, etc.




Stored in the animation storage region


314


are programs for giving predetermined action (or motion) by using a user-created image (i.e., game program).




Stored in the audio program storage region


315


are audio programs for BGM, sound effects in game, etc.




Stored in the graphic data storage region


316


are original picture (or original image) data, polygon data, texture data, etc.




The RAM


32


is used to store image data for creation captured from video signal by the user, or game backup data.




When the disk


45


is used as the external storage medium, the storage region is divided into a read-only region


45


A and a writable region


45


B. The read-only region


45


A corresponds to the ROM


31


in the cartridge


30


, which is used to store the ID information storage region


451


, OS information storage region


452


, image creating program storage region


453


, animation display program storage region


454


, audio generating program storage region


455


, and graphic data storage region


456


. The writable region


45


B corresponds to the RAM


32


, which is used to store creation image data storage region.




Whether to use the cartridge


30


or the disk


45


as the external storage medium depends on the required storage capacity (especially the capacity required in the writable region). This choice can be freely selected by the software developer.





FIG. 4

is an example storage region memory map of the RAM


15


provide in the game machine


10


. In

FIG. 4

, the RAM


15


provides a program data transfer and storage region


15


A and an image creating data storage region


15


B. In one example, part of the program data stored in the ROM


31


or the read-only region


45


A in the disk


45


shown in

FIG. 3

are transferred and stored in the program data transfer and storage region


15


A. The image creating data storage region


15


B is used to temporarily store image data being processed in respective process steps for creating a desired image by using a captured image.




For example, the image creating data storage region


15


B includes:




an original or initial image data storage region


151


,




a monochrome captured image data storage region


152


,




a color captured image data storage region


153


,




a cutout image data storage region


154


, and




a display image data storage region


155


.




The original image data storage region


151


is used to store original image data (e.g., an initial picture) for image creation by the user. In one example, such original image data comprises data representing a plurality of polygons defining images for display, e.g., an image of the character shown in the frame in

FIG. 11

(described later), and texture (pattern tiles) data to be pasted on to certain parts of such an image formed with a plurality of polygons, such as hands, legs, body, face, etc.




The monochrome captured image data storage region


152


is used to temporarily store video luminance data that may vary frame by frame (i.e., moving pictures) when a video signal is supplied.




The color captured image data storage region


153


is used to temporarily store color image data of an image to be captured from the video (e.g., moving pictures) displayed in monochrome.




The cutout image data storage region


154


is used to temporarily store images to be cut out from the captured color image data as texture data. This cutout image data storage region


154


stores color data for a plurality of images (e.g., a face of a person) cut out from the image captured from the video signal, for example.




The display image data storage region


155


is a so-called frame memory. It defines a storage region corresponding to the number of dots on the display screen of the CRT display


52


. It is used to temporarily store color data displayed on the screen.




Example Steps for Creating an Animation from Captured Images





FIG. 5

is a flowchart showing an example main routine used to define the operations of the image creating/display device of an example embodiment of the present invention.

FIGS. 6

to


10


are flowcharts of subroutines showing details of certain steps in the main routine. In

FIG. 5

, the numbers of diagrams enclosed in parentheses within the step boxes indicate those flowcharts of corresponding subroutines, and the Figure numbers on the right-hand side of the step numbers indicate corresponding drawings showing examples of images displayed in the respective steps.




The operation of the image creating device of this embodiment will be described referring to the main routine shown in FIG.


5


and the examples of screen displays corresponding to the individual steps. First, in step S


1


, a plurality of objects such as, for example, characters such as dolls, animals, and the like, are displayed in the CRT display


52


as starting or initial objects. Such initial objects are candidates the user may select for changing the face, clothes, etc. The user selects a desired character such as a doll or animal.

FIG. 11

shows an example of an image displayed when a character is selected. At this time, icons showing functions selectable by the user are displayed around the display frame showing the original initial (unmodified) image of the selected character.




Next, in step S


2


, when the user selects the “Capture Image” icon indicating capturing of image, the image creating device enters the capture image mode. In response, input video supplied from a predetermined video source (e.g., a television receiver, a video camera, a videocassette recorder, etc.) is displayed in an input video display frame on the CRT display


52


. Further, commands indicating “Capture Image” and “Return” for returning to the display in

FIG. 11

are displayed as the next selectable functions below the input video display frame (see FIG.


12


). With these commands displayed, the display also indicates which of the operating switches


24




a


and


24




b


is to be used (A, B), and which of the four points (up, down, left, right) of the direction—specifying switch


22


is to be pressed (left, right) to select the commands. At this time, a live moving picture which varies frame by frame is displayed in monochrome in the input video display frame.




When the user presses the operating switch


24




a


or the left side of the direction—specifying switch


22


to specify “Capture Image,” the screen shows a display for a process of confirming the image (see

FIG. 13

) in the next step S


4


. In this image confirming process, the captured image is displayed in color in a large area in the center of the screen, and the live video being inputted is also displayed in monochrome in a small area in the lower right part of the screen. Also, commands are displayed asking the user whether to confirm selection of the capture the color image displayed in the upper center of the screen or whether to cancel it are displayed.




When the user selects a decision command, a process of selecting a cutout area as texture is performed in step S


5


(see FIG.


14


). For example, as shown in

FIG. 14

, the captured color image confirmed in step S


4


is displayed in a large area, with a box (e.g., or a rectangular box) for specifying the area to be cut out from it. The user can change the size and dimensions of this displayed box in arbitrary ways by operating a cursor shaped like a hand. Controls for adjusting brightness, contrast, tone, etc. of the captured image data are displayed below the display frame showing the captured color image. The levels of brightness, contrast, and tone corresponding to the controls can be changed by operating the cursor. When the user has finished selecting a cutout area, an icon indicating the decision (e.g., “thumbs up”) is selected.




Next, in step S


6


, a process for confirming the cutout image as texture is performed. Next, in step S


7


, the cutout image is saved as texture data in a free area in the cutout image data storage region


154


. Then, as shown in

FIG. 15

, the selected cutout image texture data is displayed in the center of the screen. Previously selected cutout image texture data are also saved in some area in the image data storage region


154


, and displayed as icons or thumbnails on the right side of the screen in a reduced size. At step S


8


, the image capturing processing is finished.




Next, in step S


9


, the saved selected cutout image texture data (e.g., of a face cut out from the captured image) is displayed in an enlarged manner as shown in FIG.


16


. At this time, the user can edit the face of the cutout image as desired. For example, icons of a wig, nose, glasses, etc. can be selected, to cover the head with a chosen wig, change the shape of the nose, or to put glasses on the face.

FIG. 17

shows an example image obtained by editing the face in the cutout image through user's choice of editing commands and operations. In the next step S


10


, an image creating process for editing other parts of the final image (e.g., editing based on choice, for changing the clothes, shapes of other parts, etc.) is performed, to complete the image of a character suited to the user's taste, as shown in FIG.


18


.




Next, details (subroutines) of steps S


2


, S


3


, S


4


, S


5


, S


6


and S


7


in the main routine in

FIG. 5

will be described.




The processing in steps S


2


and S


3


are accomplished by executing the image capturing/image selecting subroutine shown in FIG.


6


. In step S


11


, with the image of the character as a base of the image creation displayed as shown for example in

FIG. 11

, the user selects a command indicating image capturing, and then the routine moves to step S


12


. In step S


12


, in order to display a monochrome image based on the video signal, luminance data contained in the color video signal is extracted and converted into monochrome texture data, and the luminance data of monochrome captured image is written into the storage region


152


. Next, in step S


13


, the monochrome captured image data stored in the storage region


152


is written into the storage area corresponding to the monochrome image display frame shown in

FIG. 12

in the display image data storage region


155


, and font data for displaying the command of “Capture Image” or “Return” are displayed below it on the screen. Subsequently, in step S


14


, it is determined whether the operating switch


24




a


is pressed (or the left side of the direction specifying switch


22


is pressed)—in other words, whether the “Decision” command is selected. When the decision command is not selected, it is determined in the next step S


15


whether the operating switch


24




b


(or the right side of the direction specifying switch


22


) for selecting the “Return” command is pressed. When the “Return” command is not selected, the routine returns to step S


11


and the operations in steps S


11


to S


15


are repeated. The image creating device stays in a waiting state until some command is selected. If it is determined that the “Return” command is selected in this waiting state, the routine returns to the operation in step S


2


.




If it is determined in step S


14


that the switch


24




a


(or the left side of the direction specifying switch


22


) is pressed, or that the “Decision” command is selected, the routine moves to the next Step S


4


.




Next, referring to

FIG. 7

, the image confirming process (step S


4


) will be described in detail. First, at the instant of image capture determined in the above-described step S


14


, a process of converting from the video signal of the input video into color data is performed (step S


21


). Next, in Step S


22


, the color data of the capture color image is written into the storage region


153


. Next, in step S


23


, the video signal is converted into monochrome image data (luminance signal) to be displayed as a live video moving picture, and written into the storage region


152


. Next, in step S


24


, the color captured image data stored in the storage region


153


is displayed as a still picture color image as shown in the upper frame in FIG.


13


. Next, in Step S


25


, the monochrome captured image stored in the storage region


152


is displayed in the area shown by the lower frame in

FIG. 13

as a live video moving picture. At this time, the system also displays in the lower left part of the screen that a “Decision” or “Cancel” command should be selected, to prompt the user to select one of them. Then, in step S


26


, if it is determined that “Decision” has been selected, the routine moves to the next Step S


5


. When “Decision” is not selected and “Cancel” is not selected either, the operation moves back to step S


22


. When “Cancel” is selected, the routine returns to the image selecting process in step S


3


.




Next, referring to

FIG. 8

, the processing (step S


5


) for selecting the cutout area (in other words, the size of the required texture) will be described in detail. First, in step S


31


, the color captured image stored in the storage region


153


is displayed in a large area on the screen, and the box for the particular image to be cut out is selected by using the cursor (see the rectangular box around the face in FIG.


14


). Next, in step S


32


, it is determined whether the icon indicating that the cutout image area has been determined was selected. If it wasn't selected, the routine moves to step S


34


. In step S


34


, it is determined whether the icon indicating cancellation was selected. If it is determined that it wasn't selected, the routine moves to step S


35


. In step S


35


, it is determined whether any of the color control bars for brightness, contrast, and tone has been selected. If the color control bars have been selected, the color data of the color captured image stored in the storage region


153


is corrected or adjusted as selected by the user in the next step S


36


. Subsequently, the operation moves back to step S


31


.




If it is determined in step S


35


that the color control bars have not been selected, then it is determined in step S


37


whether the box for selecting a cutout area has been changed. If it has not been changed, the routine returns to the operation in step S


31


. If it has been selected, a process of changing the area selecting box is performed in step S


38


.




Subsequently, when the color control and the cutout area selecting box have been specified and the icon indicating the determination is selected, this selection is detected in step S


32


and the routine moves to the next step S


6


. When it is determined in step S


34


that the icon indicating cancellation is selected, the routine returns to the image selecting process in step S


3


.




Next, referring to

FIG. 9

, the process of confirming the cutout image (step S


6


) will be described in greater detail. When the cutout area has been selected in the above-described cutout area selecting process (step S


5


), the color image in the selected area is cut out in step S


41


. When it is required that the size of the cutout image be changed, a process of changing the size (a process of enlarging or contracting the section of the image to be cut out) is performed in step S


42


. Next, in step S


43


, the color data of the cutout image is written into a free area in the storage region


154


. Next, in step S


44


, the cutout image data stored in the respective areas in the storage region


154


are displayed as icons on the right side of the screen as shown in

FIG. 15

, with the latest cutout image being displayed in the large area in the center of the screen. At the same time, icons for selecting saving or canceling of the latest cutout image are displayed on the left side of the screen. Next, in step S


45


, it is determined whether the icon for specifying “Save” is selected. When “Save” is not selected, it is determined in step S


46


whether the icon indicating “Cancel” is selected. When neither of the icons is selected, the routine returns to the operation in step S


44


and the operations in steps S


44


, S


45


and S


46


are repeated.




When the icon indicating “Save” is selected, it is decided in step S


45


and the routine moves to the next step S


7


. When the icon indicating “Cancel” is selected, it is detected in step S


46


and the routine returns to the cutout area selecting process in step S


5


.




Next, referring to

FIG. 10

, the process of saving the cutout image (step S


7


) will be described in detail. First, in step S


51


, the process of correcting the color captured image stored in the storage region


153


is performed, and then the color captured image data is transferred and stored into the storage region


155


. In the process of correcting color captured image, when the captured image is a face of a person, the face is represented in an egg-shaped area inside the rectangular box (see FIG.


15


). Accordingly, the captured image data in the four corners of the rectangular box is shaded or masked off. Or, alternatively, color data in two separate, distinct points in the face (e.g., forehead and chin) can be added and divided by two to obtain the average, and the average color data is written in the part of the captured image data outside the face outline. Another method of correcting the captured face image involves writing color data as if a filter or mask is superimposed on the cutout color captured image, wherein the filter or mask is transparent in the egg-shaped part surrounded by the rectangular box and the transparency decreases (i.e., the mask becomes more opaque) in areas outward from the periphery of the egg-shaped part. The color captured image thus corrected is enlarged and displayed in a large area in the center of the CRT display


52


(see

FIG. 16

) and is superimposed onto the shoulders of the character to receive the captured facial image.




At the same time, a plurality of icons of textures showing previously captured cutout images are displayed around the display frame, and various commands are also displayed. Next, in step S


52


, it is determined whether the cutout color captured image should be saved in the external storage medium, e.g., in the storage region in the RAM


32


(or in the writable region


45


B in the magnetic disk


45


). When the user does not select the icon indicating saving, this is detected in step S


52


and the routine moves to step S


53


. The program detects in step S


53


whether the icon indicating canceling is selected, and if it is determined that the icon is not selected, the routine moves to the operation in step S


51


. In this way, when none of the commands is selected, the operations in steps S


51


, S


52


and S


53


are repeated. At this time, if the user selects the command indicating saving (or backup storage), this is detected in step S


52


and color data of the plurality of cutout images stored in the storage region


154


are written into the RAM


32


(or the writable region


45


B in the magnetic disk


45


) and the routine moves to the operation in step S


8


.




When the icon indicating canceling is selected, this is detected in step S


53


and the routine returns to the cutout image area selecting process in step S


5


.




Next, with the character having the face cut out and pasted onto it as described above being displayed as shown in

FIG. 18

, the operation for causing the character to move (i.e., be animated) in a desired way will be described. In this case, as shown in

FIG. 18

, the face part of the character is changed to the cutout color captured image, and textures of clothes chosen by the user are pasted thereon.




Then, as shown in the left side of the display in

FIG. 18

, icons for selecting types of motion of the character are displayed. When one of the icons is selected, an animation display program (a program for realizing display of motion of the character) corresponding to the selected icon is read from the storage region


314


in the ROM


31


or from the storage region


454


in the magnetic disk


45


, and written into the transfer and storage region in the RAM


15


. Then the CPU


13


executes the transferred animation display program to calculate coordinate data of the polygons of the character, frame by frame, according to previously prepared movement patterns and animations. Further, textures (pattern tiles) of the hands, legs, body, clothes, etc. are pasted on the individual parts of the character on the basis of the calculation about a plurality of polygons, and they are written into the storage region


155


. The texture data written in the storage region


155


is read and displayed, to give motion to the individual parts of the body of the character as a still object shown in FIG.


18


.




The user can play games by using the image created in the above-described embodiment. In this case, the user operates the direction specifying switch


22


or the analog stick


23


to specify movement of the created character, and the background image is changed by program according to the movement of the character.




While the above-described embodiment provides an example in which image of a part of the body (e.g., the face) of an initial image, e.g., an image of a character, is changed to a separately captured color image, it is noted here that the present invention can be applied with a variety of modifications and changes. For example, the face part can be changed to a face of an animal, or an original picture of an animal can be used and the face of the original picture can be changed to a face of another animal.




While the present invention has been described in detail, the foregoing description is in all aspects illustrative and not restrictive. It is understood that numerous other modifications and variations can be devised without departing from the scope of the invention.



Claims
  • 1. In a software-controlled home video game machine system specifically designed for interactive 3D video game play, including 3D animated graphics and associated sound generation, said home video game machine system including a user-operable hand-held controller having a housing and a joystick provided thereon communicating with (i) a main processor, (ii) a 3D graphics coprocessor connected to the main processor for providing at least polygon coordinate transformation and light source processing, and (iii) at least one memory including a frame buffer communicating with the 3D graphics coprocessor, said at least one memory storing plural polygon coordinates defining surfaces of 3D animated video game characters, the home video game machine system playing interactive games based on software loaded therein, a method for allowing a video game player to create animated 3D images from captured 2D image data and interact with the animated 3D images to provide interactive game play, comprising:(a) providing a captured 2D color image to the home video game machine system for storage into said at least one memory; (b) allowing the video game player to select a portion of the captured 2D color image by operating the user-operable hand-held controller of the home video game machine system, the selected portion being at least temporarily stored in said at least one memory communicating with the 3D graphics coprocessor of the home video game machine system; (c) processing, with at least one of the home video game machine system main processor or the home video game machine system 3D graphics coprocessor, the selected 2D color image portion to convert the selected 2D color image portion into a color texture; (d) texture mapping, using the home video game machine system 3D graphics coprocessor, the color texture obtained from the processed selected 2D color image portion onto a predefined surface of a 3D video game character defined by plural polygon coordinates stored in the at least one memory communicating with the 3D graphics coprocessor; and (e) animating and displaying, using at least the main processor and 3D graphics coprocessor, the 3D video game character having the applied color texture to provide interactive video game play in response to manipulation of the user-operable hand-held controller.
  • 2. The method of claim 1 wherein step (a) comprises capturing live video from a camera and generating a 2D still image from said captured video and storing the 2D still image in the at least one memory communicating with the home video game machine system 3D graphics coprocessor.
  • 3. The method of claim 1 wherein step (b) includes displaying a cutout window overlaid onto the captured image and allowing the game player operating the user-operable hand-held controller to change the position and dimensions of the cutout window so as to select a cutout portion of the captured 2D image and store the selected cutout portion in the at least one memory for application to the 3D character using the home video game machine system 3D graphics coprocessor.
  • 4. The method of claim 1 wherein step (b) includes allowing the game player operating the user-operable hand-held controller to adjust brightness, contrast and/or tone of the captured 2D image.
  • 5. The method of claim 1 wherein step (b) includes allowing the game player operating the user-operable hand-held controller to add or change portions of the captured 2D image.
  • 6. The method of claim 5 wherein step (b) includes allowing the game player operating the user-operable hand-held controller to modify the nose portion of a captured 2D facial image.
  • 7. The method of claim 5 wherein step (b) includes allowing the game player operating the user-operable hand-held controller to modify the ear portion of a captured 2D facial image.
  • 8. The method of claim 5 wherein step (b) includes allowing the game player operating the user-operable hand-held controller to modify the hair portion of a captured 2D facial image.
  • 9. The method of claim 1 wherein step (a) includes automatically masking off all but a facial portion of a captured 2D image so that the facial portion can be selected in step (b) and stored in the at least one memory.
  • 10. The method of claim 1 wherein step (c) includes converting the captured 2D image to texture data and step (d) includes using the home video game machine system 3D graphics coprocessor to utilize said texture data to texture a facial portion of the 3D video game character defined by polygon data stored in the at least one memory.
  • 11. The method of claim 1 further comprising allowing the video game player to edit to a surface of the 3D video game character other than the predetermined surface by manipulation of the user operable hand-held controller.
  • 12. The method of claim 1 wherein the video game player is allowed to edit the predetermined surface of the 3D video game character through operation of the user-operable hand-held controller after the color texture is texture mapped onto the predefined surface of the 3D video game character.
Priority Claims (1)
Number Date Country Kind
9-337655 Nov 1997 JP
Parent Case Info

This application is the Division of application Ser. No. 09/190,601 filed Nov. 12, 1998 , now U.S. Pat. No. 6,285,381.

US Referenced Citations (136)
Number Name Date Kind
4104625 Bristow et al. Aug 1978 A
4261012 Maloomian Apr 1981 A
4286849 Uchidoi et al. Sep 1981 A
4359223 Baer et al. Nov 1982 A
4467349 Maloomian Aug 1984 A
4486774 Maloomian Dec 1984 A
4517658 Iida May 1985 A
4521014 Sitrick Jun 1985 A
4539585 Spackova et al. Sep 1985 A
4541010 Alston Sep 1985 A
4566263 Miyazaki et al. Jan 1986 A
4591897 Edelson May 1986 A
4662635 Enokian May 1987 A
4684995 Baumeister Aug 1987 A
4688105 Bloch et al. Aug 1987 A
4710873 Breslow et al. Dec 1987 A
4731743 Blancato Mar 1988 A
4803554 Pape Feb 1989 A
4823285 Blancato Apr 1989 A
4827347 Bell May 1989 A
4885702 Ohba Dec 1989 A
5016107 Sasson et al. May 1991 A
5093731 Watanabe et al. Mar 1992 A
5095798 Okada et al. Mar 1992 A
5111283 Nagasawa et al. May 1992 A
5111409 Gasper et al. May 1992 A
5153729 Saito Oct 1992 A
5182647 Chang Jan 1993 A
5184830 Okada et al. Feb 1993 A
5185665 Okura et al. Feb 1993 A
5185818 Warnock Feb 1993 A
5191645 Carlucci et al. Mar 1993 A
5200863 Orii Apr 1993 A
5237648 Mills et al. Aug 1993 A
5239419 Kim Aug 1993 A
5249242 Hanson et al. Sep 1993 A
5255357 Byron et al. Oct 1993 A
5262867 Kojima Nov 1993 A
5293236 Adachi et al. Mar 1994 A
5301026 Lee Apr 1994 A
5301267 Hassett et al. Apr 1994 A
5303334 Snyder et al. Apr 1994 A
5305118 Schiller et al. Apr 1994 A
5343243 Maeda Aug 1994 A
5363119 Snyder et al. Nov 1994 A
5369736 Kato et al. Nov 1994 A
5371540 Tamura et al. Dec 1994 A
5396225 Okada et al. Mar 1995 A
5414444 Britz May 1995 A
5423554 Davis Jun 1995 A
5430496 Silverbrook Jul 1995 A
5437008 Gay et al. Jul 1995 A
5438359 Aoki Aug 1995 A
5440401 Parulski et al. Aug 1995 A
5442543 Tresp Aug 1995 A
5459819 Watkins et al. Oct 1995 A
5469536 Blank Nov 1995 A
5475441 Parulski et al. Dec 1995 A
5475539 Orii Dec 1995 A
5477264 Sarbadhikari et al. Dec 1995 A
5493409 Maeda et al. Feb 1996 A
5504842 Gentile Apr 1996 A
5506944 Gentile Apr 1996 A
5519826 Harper et al. May 1996 A
5524194 Chida et al. Jun 1996 A
5533181 Bergsneider Jul 1996 A
5539459 Bullitt et al. Jul 1996 A
5539865 Gentile Jul 1996 A
5543835 Mumura Aug 1996 A
5544290 Gentile Aug 1996 A
5546528 Johnston Aug 1996 A
5553864 Sitrick Sep 1996 A
5577189 Gay et al. Nov 1996 A
5581299 Raney Dec 1996 A
5581311 Kuroiwa Dec 1996 A
5586238 Murata Dec 1996 A
5592609 Suzuki et al. Jan 1997 A
5595389 Parulski et al. Jan 1997 A
5600767 Kakiyama et al. Feb 1997 A
5601487 Oshima et al. Feb 1997 A
5602976 Cooper et al. Feb 1997 A
5612716 Chida et al. Mar 1997 A
5612732 Yuyama et al. Mar 1997 A
5621431 Harper et al. Apr 1997 A
5625711 Nicholson et al. Apr 1997 A
5625716 Borg Apr 1997 A
5631701 Miyake May 1997 A
5633678 Parulski et al. May 1997 A
5633733 Miyazawa May 1997 A
5633985 Severson et al. May 1997 A
5634064 Warnock et al. May 1997 A
5638498 Tyler et al. Jun 1997 A
5638502 Murata Jun 1997 A
5638503 Hoel Jun 1997 A
5659490 Imamura Aug 1997 A
5666159 Parulski et al. Sep 1997 A
5680533 Yamato et al. Oct 1997 A
5681223 Weinreich Oct 1997 A
5682197 Moghadam et al. Oct 1997 A
5689611 Ohta et al. Nov 1997 A
5696850 Parulski et al. Dec 1997 A
5706097 Schelling et al. Jan 1998 A
5706457 Dwyer et al. Jan 1998 A
5708883 Segan et al. Jan 1998 A
5715486 Kim et al. Feb 1998 A
5719799 Isashi Feb 1998 A
5729637 Nicholson et al. Mar 1998 A
5737452 Schiller Apr 1998 A
5737491 Allen et al. Apr 1998 A
5742291 Palm Apr 1998 A
5745122 Gay et al. Apr 1998 A
5748196 Coelho et al. May 1998 A
5748326 Thompson-Bell et al. May 1998 A
5754227 Fukuoka May 1998 A
5781198 Korn Jul 1998 A
5784525 Bell Jul 1998 A
5796429 Suzuki et al. Aug 1998 A
5830065 Sitrick Nov 1998 A
5835101 Otsuka Nov 1998 A
5880740 Halliday et al. Mar 1999 A
5937081 O'Brill et al. Aug 1999 A
5984780 Takemoto et al. Nov 1999 A
6026215 Fantone et al. Feb 2000 A
6061532 Bell May 2000 A
6064393 Lengyel et al. May 2000 A
6115036 Yamato et al. Sep 2000 A
6120379 Tanaka et al. Sep 2000 A
6139432 Watanabe et al. Oct 2000 A
6208347 Migdal et al. Mar 2001 B1
6227974 Eilat et al. May 2001 B1
6229904 Huang et al. May 2001 B1
6256047 Isobe et al. Jul 2001 B1
6280323 Yamazaki et al. Aug 2001 B1
6282362 Murphy et al. Aug 2001 B1
6283858 Hayes, Jr. et al. Sep 2001 B1
6343987 Hayama et al. Feb 2002 B2
Foreign Referenced Citations (12)
Number Date Country
2130842 Jun 1984 GB
2323738 Sep 1998 GB
54-179710 Jun 1953 JP
53-127524 Nov 1978 JP
54-319 Jan 1979 JP
54-16320 Feb 1979 JP
57-41360 Mar 1982 JP
57-72666 May 1982 JP
58-24270 Feb 1983 JP
11-154240 Nov 1997 JP
WO9743020 Nov 1997 WO
WO 01635560 Aug 2001 WO
Non-Patent Literature Citations (43)
Entry
Public Patent Bulletin HEI Publication No. 08-018839 (Application No. 06-144326), Akihito et al., “Image Pickup Device”, Jan. 19, 1996, pp. 1-7 plus English Abstract JP 8018839.
Public Patent Bulletin HEI Publication No. 08-223524 (Application No. 07-020698), Kunio et al., “Portable Video Camera”, Aug. 30, 1996, pp. 1-14 plus English Abstract JP8223524.
Public Patent Bulletin SHO Publication No. 58-024270 (Application No. 56-121786), Shinichi, “Electrophotographing Device”, Feb. 14, 1983, 8 pages plus English Abstract JP 58024270.
Public Patent Bulletin SHO Publication No. 62-00167 (Application No. 60-139352), Hitoshi, “Half Tone Recording Device”, Jan. 6, 1987, pp. 1-6 plus English Abstract JP 62000167.
Public Patent Bulletin SHO Publication No. 63-272261 (Application No. 62-107832), Tomokage, “Picture Information Reader”, Nov. 9, 1988, pp. 1-6 plus English Abstract JP 63272261.
Public Patent Bulletin HEI Publication No. 02-039195 (Application No. 63-188421), Koji et al., “Display System for Multi-Valued Image”, Feb. 8, 1990, pp. 1-7 plus English Abstract JP 2039195.
Public Patent Bulletin SHO Publication No. 58-036076 (Application No. 56-133863), Masao, “Video Synthesizing Device”, Mar. 2, 1983, pp. 1-7 plus English Abstract JP 58036076.
Public Patent Bulletin SHO Publication No. 63-213078 (Application No. 62-045991), Kenji et al., “Picture Synthesizing Method”, Sep. 5, 1988, pp. 1-5 plus English Abstract JP 63213078.
Public Patent Bulletin HEI Publication No. 03-139064 (Application No. 01-275816), Kauhiro et al., “Image Synthesizer”, Jun. 13, 1991, pp. 1-5 plus English Abstract JP 3139064.
Public Patent Bulletin HEI Publication No. 03-275092 (Application No. 02-074888), Hirobumi et al., “TV Game Device”, Dec. 5, 1991, pp. 1-5 plus English Abstract JP 3275092.
Public Patent Bulletin HEI Publication No. 04-329758 (Application No. 03-128432), Tomoaki et al., “Picture Recorder”, Nov. 18, 1992, pp. 1-4 plus English Abstract JP 4329758.
Public Patent Bulletin HEI Publication No. 09-140936 (Application No. 07-303965), Isamu et al., “Game Machine”, Jun. 3, 1997, pp. 1-8 plus English Abstract JP 9140936.
Public Patent Bulletin Publication No. 6-96180 (Application No. EP 19930304530 19930610), Gay et al., “Method of Establishing Constraints Between Graphical Elements”, Dec. 29, 1993, pp. 1-31 plus English Abstract (EP0576178 A3) JP 6096180 A.
Public Patent Bulletin Publication No. 5-110835 (Application No. EP19910309007 19911002), Schiller, “Method of Assigning Pixels to Cells of a Halftone Grid”, Aug. 12, 1992, pp. 1-23 plus English Abstract JP 5110835.
Public Patent Bulletin Publication No. 7-46462 (Application No. EP 19930302488 19930330), Ubillos, “Method and Apparatus for Video Editing”, Oct. 6, 1993, pp. 1-13 plus English Abstract JP 7046462A.
Public Patent Bulletin Publication No. 8-190504 (Application No. EP19950304792 19950710), Paknad et al., “Method and Apparatus for Finding File Paths on a Computer System”, Apr. 10, 1996, pp. 1-28 plus English Abstract JP 8190504.
Public Patent Bulletin Publication No. 8-194697 (Application No. EP19950305330 19950731), Ayers, “Method and Apparatus for Identifying Words Described in a Page Description Language File”, Mar. 13, 1996, pp. 1-26 plus English Abstract JP8194697A.
Public Patent Bulletin Publication No. 8-202337 (Application No. EP19950304971 19950717), Dowling et al., “Method of Generating Multiple-Master Typefaces”, Mar. 13, 1996, pp. 1-6 plus English Abstract JP8202337.
Public Patent Bulletin Publication No. 8-249329 (Application No. EP 19950303939 19950608), Ayers et al., “Method and Apparatus for Identifying Words Described in a Portable Electronic Document”, Mar. 20, 1996, pp. 1-26 plus English Abstract JP 8249329A.
Public Patent Bulletin Publication No. 9-282474 (Application No. JP19960115615 19960412), Fuminori, “Gradational Image Generating Method”, Oct. 31, 1997, pp. 1-8 plus English Abstract JP9282474A.
Public Util. Mod. Bull. HEI Publication No. 6-14215 (Application No. JP19910305002 19911120), Koike, “Digital Signal Processor”, Jan. 21, 1994, pp. 1-5 plus English Abstract JP6014215.
Campbell, Colin, “Digimask Puts Face on Gaming's Future,” Games Business (May 15, 2000).
“ECTS: Digimask Faces Gaming Partners,” The Fastest Game News Online (Sep. 5, 2000).
“ARIS Builds Digimast Future,” www.uk.aris.com (Aug. 10, 2000).
Sturgeon, Hannah, 3-d,netimperative.com (Jul. 30, 2000).
Wapshott, Tim, “Face up to reality, ” The Times Interface 2000, www.Times-archive.co.uk (Oct. 24, 2000).
Fudge, James, “Digimask Demonstration Released,” (Jun. 30, 2000).
Owen, Steve, “Old Telstar Crew Shows Its Face,” www.gamespot.co.uk (May 26, 2000).
“Facing the Future, Industry: A new technology called Digimask allows web users to create 3D animations of their own faces,” www.fgonline.com, (Mar. 29, 2000).
“Didgmask puts you in picture,” www.Sunday-times.co.uk (Mar. 26, 2000).
Fudge, James, “Digimask at E3Expo 2000,” www.cdmag.com (May 3, 2000).
Press release, “Tomorrow's Not Just Another Day For Digimask,” www.digimask.com (Oct. 6, 2000).
Press release, “Digimask Puts Face to Big Game Names,” www.digimask.com (Sep. 4, 2000).
Press release, “Digimask Prepares to Face its Public,” www.digimask.com (Jul. 21, 2000).
Press release, “Digimask Appoint Foresight for Web Relaunch,” www.digimask.com (Jun. 21, 2000).
Press release, “Pearce Joins Digimask for New Global Role,” www.digimask.com (Sep. 4, 2000).
Press release, “Hetherington Secures £5 Million Funding for Digimask,” www.digimask.com (Mar. 31, 2000).
Web site information, ign64-ign.com, “Rare Puts Gamers in the Game” (Mar. 12, 1999).
Web site information, ign64-ign.com, “Perfect Dark Ditches Face Mapping” (Feb. 2, 2000).
Web site information, ign64-ign.com, “IGN's Best of E3 Awards” Mar. 17, 1999).
Gelmis, “Nintendo Users Putting on a Game Face,” Newsday (Oct. 13, 1999).
Web site information, digimask.com.
GAME BOY® Camera Instruction Booklet.