This application claims priority to an application entitled “Apparatus and Method for Controlling Music Play in Mobile Communication Terminal” filed in the Korean Intellectual Property Office on Aug. 27, 2004 and assigned Serial No. 2004-67862, on Feb. 26, 2005 and assigned Serial No. 2005-16294, and on Apr. 12, 2005 and assigned Serial No. 2005-30528 the contents of which are incorporated herein by reference.
1. Field of the Invention
The present invention relates to an apparatus and a method for utilizing a mobile communication terminal for entertainment, and more particularly to an apparatus and a method for controlling music play in a mobile communication terminal recognizing continuous motion through a motion recognition sensor unit.
2. Description of the Related Art
Recently, portable mobile communication terminals such as cell phones and Personal Digital Assistants (PDAs) have achieved wide spread use. With the development of communication technology, these mobile communication terminals are providing additional functions such as game, alarm and MP3 player functions using various images and sounds in addition to communication functions such as telephone and Short Message Service (SMS) functions. Accordingly, mobile communication terminal users enjoy various conveniences in using the terminals as well as the original communication function.
However, users require that mobile communication terminals always carried by the users also have more interesting functions. Accordingly, mobile communication terminal manufacturers have tried to develop other interesting functions by means of an existing sound output function, display function, etc.
Typically, a beatbox (electronic percussion on drum machine) through a motion recognition sensor unit (e.g., an inertia sensor) allows a corresponding musical performance to be played by means of instrumental sounds previously stored in a mobile communication terminal. Further, a beatbox outputs sounds from preset sound sources whenever a user shakes a mobile communication terminal. While sounds are outputted, a background scene is fixed to one image.
Types of sound sources used in a sound function through a motion recognition sensor unit may be restricted. When these restricted instrument sound sources are used, enjoyment obtained by utilizing the sound function by users may also be restricted. Further, it may be impossible for users unskilled with musical beats to enjoy a beatbox play function. Furthermore, when a background scene is fixed to one image during the sound function, it may be visually unimpressive as compared with the output sound.
Accordingly, the present invention has been made to solve the above-mentioned problems occurring in the prior art, and it is an object of the present invention to provide an apparatus, in which a user can play an interesting game using a mobile communication terminal, and a control method thereof.
It is another object of the present invention to provide an apparatus, in which a mobile communication terminal can output sounds in response to a user's motion, and a control method thereof.
It is further another object of the present invention to provide an apparatus and a method which can improve the interest a user experiences by increasing the degree of freedom in using a sound function [beatbox, music box (electronic sound synthesizer), etc] of a mobile communication terminal having a motion recognition sensor unit.
It is still another object of the present invention to provide an apparatus and a method which can utilize a sound function as various sound sources in addition to sound sources included in a mobile communication terminal.
It is yet another object of the present invention to provide an apparatus and a method, in which a user follows and learns beats according to types of music in learning a beatbox, thereby increasing a users interest in learning the beatbox.
It is yet another object of the present invention to provide an apparatus and a method which can provide an enjoyable experience from a visual standpoint as well as an auditory standpoint when a sound function is used in displaying a background scene.
In order to accomplish the aforementioned object, according to one aspect of the present, there is provided an apparatus for controlling music play in a mobile communication terminal, the apparatus including a motion recognition sensor unit for detecting a motion of the mobile communication terminal and outputting detection signals; a sound source chip for outputting sound; and a controller for receiving the detection signals from the motion recognition sensor unit, calculating motion values of the mobile communication terminal, and controlling the sound source chip to output sounds dependent on the calculated motion values.
In order to accomplish the aforementioned object, according to another aspect of the present, there is provided a method for controlling music play in a mobile communication terminal, the method including determining if a motion of the mobile communication terminal exists; calculating motion values of the mobile communication terminal when the motion of the mobile communication terminal exists; and outputting sounds dependent on the calculated motion values.
In order to accomplish the aforementioned object, according to further another aspect of the present, there is provided an apparatus for controlling music play in a mobile communication terminal, the apparatus including a user interface for user input required for the music play; a display unit for displaying an information screen for music to be played; a motion recognition sensor unit for instantaneously detecting a motion of the mobile communication terminal; a sound file storage unit including an area for storing at least one music information; a controller for controlling corresponding music to be played according to a motion of the mobile communication terminal detected by the motion recognition sensor unit; and a speaker for outputting sounds of the played music.
In order to accomplish the aforementioned object, according to still another aspect of the present invention, there is provided a method for controlling music play in a mobile communication terminal, the method including displaying a screen for changing beatbox setup; and displaying a screen for guiding a user to learn a beatbox while outputting the beatbox according to a corresponding state when a play command is input in a state where the beatbox has been set or in an initial setting state.
In order to accomplish the aforementioned object, according to yet another aspect of the present, there is provided a method for controlling music play in a mobile communication terminal, the method including displaying a screen for selecting a tune to be played by a music box; displaying an information screen for the selected tune; and outputting sounds corresponding to musical notes of the tune whenever a motion of the mobile communication terminal is detected.
The above and other objects, features and advantages of the present invention will be more apparent from the following detailed description taken in construction with the accompanying drawings in which:
a and 14b are flow diagrams illustrating a beatbox-learning-play method in a mobile communication terminal according to a preferred embodiment of the present invention;
Hereinafter, preferred embodiments according to the present invention will be described with reference to the accompanying drawings. In the following description, particular items such as a detailed display (design) of a menu screen are shown, but these are provided for helping the general understanding of the present invention. It will be understood by those skilled in the art that these elements may be modified within the scope of the present invention. In the following description of the present invention, a detailed description of known functions and configuration incorporated herein will be omitted when it may make the subject matter of the present invention unclear.
The present invention provides a function for outputting sounds dependent on the motion of a mobile communication terminal. For this purpose, the present invention detects the motion of a mobile communication terminal, calculates motion values of the detected motion, and outputs sounds dependent on a corresponding motion value. The sounds may be output in a way preset by a user.
Accordingly, a user controls the motion of a mobile communication terminal, thereby enabling sounds to be output as the user wants. For example, interesting sounds may be output in different settings such as a concert hall or a karaoke.
A function of learning a beatbox according to a preferred embodiment of the present invention includes a screen for setting types, sound sources and volume of the beatbox to be learned, and a screen for imitating and learning the beatbox. The learning the beatbox is carried out in such a manner that beats selected by a user are output bar by bar and the user imitates the corresponding bars. Further, in order to learn the beatbox, the user selects already stored beats, instruments and volume. When the learning starts, the user can learn the beat while imitating the beat bar by bar.
It is preferred that the motion recognition sensor unit 50 may include a gyro sensor. The gyro sensor is a sensor using an angular momentum in order to detect an angular motion with respect to an inertia space around one or more axes orthogonal to a spin axis.
The motion recognition sensor unit 50 instantaneously detects the motion of the mobile communication terminal. That is, when the mobile communication terminal moves, the motion recognition sensor unit 50 outputs detection signals representing the motion of the mobile communication terminal in a horizontal and a vertical (back and forth or right and left) direction. The motion recognition sensor unit 50 may be embodied by a sensor outputting the detection signals based on the motion of the mobile communication terminal.
The controller 10 recognizes the reciprocating motion of the mobile communication terminal by means of the detection signals output from the motion recognition sensor unit 50. Further, the controller 10 calculates motion values of the mobile communication terminal by means of the detection signals output from the motion recognition sensor unit 50. Furthermore, the controller 10 controls sounds to be output through a sound source chip 80 according to the motion values of the mobile communication terminal.
When sound data of a sound file such as a musical instrument digital interface (MIDI) are input from the controller 10, the sound source chip 80 converts the input sound data into analog signals audible to a user and outputs the analog signals. According to the embodiment of the present invention, it is preferred that the sound source chip 80 is a high performance sound source chip for outputting the tone of a piano, a guitar, a drum, etc., as if its original tone is output.
When the detection signals based on the motion of the mobile communication terminal are received from the motion recognition sensor unit 50 after a predetermined instrument is selected by a user, the controller 10 recognizes the reciprocating motion of the mobile communication terminal and controls the sounds of the instrument selected by the user to be output through the sound source chip 80 according to the reciprocating motion.
For example, when the user shakes the mobile communication terminal after selecting a tambourine from a karaoke, the controller 10 detects the motion of the mobile communication terminal and outputs tambourine sounds dependent on the motion of the mobile communication terminal. Accordingly, the user can hear sounds similar to that made when a tambourine is actually shaken.
Further, when the detection signals based on the motion of the mobile communication terminal are received from the motion recognition sensor unit 50 after a predetermined sound file is selected by a user, the controller 10 recognizes the motion of the mobile communication terminal and controls sounds, which correspond to each note of music in a sound file selected by the user, to be output through the sound source chip 80 in sequence according to the motion.
For example, when the user shakes the mobile communication terminal after selecting a sound file ” twinkle twinkle little star”, the controller 10 outputs notes corresponding to the motion of the mobile communication terminal, i.e., sounds corresponding to ” do-do-sol-sol-la-la-sol”. The output sounds may be sounds from the instrument set by the user.
Further, the controller 10 may control an image proper for the output sounds to be displayed through a display unit 30. For example, the controller 10 may display a dynamic image or an image, of each instrument being played, on the display unit 30. Further, the controller 10 may display a musical note representing music in a sound file and indicate notes corresponding to currently output sounds for discernment on the display unit 30.
The display unit 30 displays various messages, etc., under the control of the controller 10, and may include a Liquid Crystal Display (LCD), a Thin Film Transistor (TFT), an Organic Electroluminescence (EL), etc. A user interface unit 40 includes a plurality of number keys and function keys and outputs key input data corresponding to keys pressed by the user to the controller 10.
Further, a memory unit 20 connected to the controller 10 includes a Read Only Memory (ROM) and a Random Access Memory (RAM) for storing a plurality of programs and information required for controlling operations of the mobile communication terminal, a voice memory, etc. A sound file storage unit 22 stores a plurality of sound files, and these sound files may be downloaded through the Internet, etc., by the user.
Herein, the sound file may include data regarding predetermined percussion instrument sounds, or data regarding specific music. When the sound file includes the data regarding specific music, the sound file may include a predetermined indicator allowing the controller 10 to discern notes constituting the music according to the embodiment of the present invention. Further, the sound file may include a predetermined indicator allowing the controller 10 to discern beats corresponding to the music according to another embodiment of the present invention. Accordingly, when the sound file includes the indicator required for discerning the beats, the controller 10 may output sounds corresponding to each beat based on the motion of the mobile communication terminal.
The mobile communication terminal typically performs wireless communication with a base station. For this, a Radio Frequency (RF) module 90 transmits/receives RF signals with the base station (not shown) through an antenna. In detail, the RF module 90 converts the received RF signals to Intermediated Frequency (IF) signals and outputs the IF signals to a baseband processor 60. Further, the RF module 90 converts IF signals input from the baseband processor 60 to RF signals, and transmits the RF signals. The baseband processor 60, a Baseband Analog ASIC (BAA) providing an interface between the controller 10 and the RF module 90, converts baseband digital signals applied from the controller 10 to analog IF signals, and applies the analog IF signals to the RF module 90. Further, the baseband processor 60 converts analog IF signals applied from the RF module 90 to baseband digital signals, and applies the baseband digital signals to the controller 10. A voiceband signal processor 70 connected to the controller 10 is connected to a microphone 72 and a speaker 74. The microphone 72 converts sounds input for recording, communication, etc., to electrical signals. The speaker 74 converts electrical signals to sounds and outputs the sounds, thereby enabling the user to hear the music play. The voiceband signal processor 70 digitally processes the electrical signals input from the microphone 72, and converts the digital signals into analog signals for output to the speaker 74.
When the user has selected a predetermined instrument, the controller 10 displays an image of the selected instrument on the display unit 30 in step 130. Step 130 is a selective step which may be omitted from the flow diagram of
In step 150, the controller 10 calculates motion values based on the motion of the mobile communication terminal. When back and forth or right and left motion of the mobile communication terminal is detected through the motion recognition sensor unit 50, e.g., a gyro sensor, and signals from the detected motion are transferred to the controller 10, the controller 10 calculates the motion values according to the signals from the motion recognition sensor unit 50.
In step 160, the controller 10 outputs sounds dependent on the motion values and displays a dynamic image of the instrument selected by the user. The display of the dynamic image of the instrument is a selective step which may be omitted from the flow diagram of
In this way, the mobile communication terminal user can hear the sounds as if an instrument is actually being played.
Referring to
When the user has selected a predetermined percussion instrument, the controller 10 determines if the mobile communication terminal is in motion through the motion recognition sensor unit 50 in step 230. Then, the controller 10 determines if a reciprocating motion of the mobile communication terminal has occurred once in step 240. That is, the controller 10 determines if the mobile communication terminal has reciprocated once more than a predetermined speed by means of the detection signals from the motion recognition sensor unit 50.
When the mobile communication terminal has reciprocated once, the controller 10 outputs sounds of the percussion instrument selected by the user, in step 250. The output sounds of the percussion instrument may be sounds which are output when hitting the percussion instrument once. In addition, the output sounds of the percussion instrument may be sounds which are output when hitting the percussion instrument twice. In step 260, the controller 10 determines if the play mode has been ended. As described above, the play mode may be ended by the user, e.g., the play mode may be ended when a call is terminated. When the play mode has not been ended, the control flow returns to step 240.
In the meantime, when the sounds according to the reciprocating motion of the mobile communication terminal are output, an image or a dynamic image as illustrated in
As described above, because the sounds are output when the mobile communication terminal user shakes the mobile communication terminal after selecting the tambourine from a karaoke, etc., the user can hear the sounds similar to those when the tambourine is actually shaken.
When a user selects a play mode, the controller 10 determines if the mobile communication terminal is in motion through the motion recognition sensor unit 50 in step 330. Then, the controller 10 determines if a reciprocating motion of the mobile communication terminal has occurred in step 340. That is, the controller 10 determines if the mobile communication terminal has reciprocated more than a predetermined speed by means of the detection signals from the motion recognition sensor unit 50.
When the mobile communication terminal has reciprocated once, the controller 10 displays each note corresponding to the reciprocating motion and simultaneously outputs sounds in step 350.
One embodiment of the present invention may include a predetermined indicator allowing notes constituting music to be discerned.
In step 360, the controller 10 determines if the play mode has been ended. When the play mode has not been ended, the control flow returns to step 340.
When the sounds are output according to the reciprocating motion of the mobile communication terminal, images of those notes as illustrated in
When the user wants to change the instrument types, the user presses the down direction key in the state of
If the user wants to change the volume value, the user presses the down direction key in the state of
After the beatbox-learning screen is displayed as illustrated in
The beatbox-learning-play is performed through listening and imitation performed in turn, and the user may learn a selected beat through this process. A description will be given on an assumption that the user learns a beatbox of two bars.
The first bar of the tune performed during “Goodgery Rhythm” is “dung-gi duck dung du-ru-ru-ru ” and the second bar is “kung-gi duck kung du-ru-ru-ru”.
Then, when the user presses the confirmation key, the screen for listening the second bar is displayed as illustrated in
In addition to the basic instruments, a user may customize “My Instrument” for use. The “My Instrument” item includes a customizing sub-item for recording sound sources and a storage space sub-item for reproducing or deleting the recorded sound sources. In order to record sound sources to be used as an instrument, recording time is first set. After the time is set, a bar graph reporting a time point at which the sound sources are recorded is shown, and the user makes the sounds of an instrument or voice to be used as the sound sources during the recording time. The sound sources recorded in this way may be used in the beatbox play and learning.
When the user presses the cancel key in each state of FIGS. 9(3b) to (3e), returns to the previous state.
When the user presses the cancel key in each state of FIGS. 10(4b) and (4c), returns to the previous state.
As compared with the construction of
The sound file storage unit 23 for storing music information relating to the play may include an area (beatbox storage space) for storing types of a beatbox, an area (instrument storage space) for storing the original instrument and “My Instrument” customized by a user, an area (background storage space) for storing preset information (e.g., musical notes or color based on progress of the play, avatars, etc) according to predetermined conditions for display of a background scene, an area (music box storage space) for storing tune information (musical notes and words) for play of a music box, etc.
The display unit 30 displays a screen (menu screen) for changing beatbox type, instrument type and volume, a beatbox-listening screen, and a beatbox-imitation screen. User input required for using functions for learning the beatbox is accomplished through a user interface unit 40. A controller 10 controls a corresponding menu screen to be displayed on the display unit 30 in response to the input through the user interface unit 40. The controller 10 controls the beatbox-learning-setting, the beatbox-learning-play, the “My Instrument” customization, the “My Instrument” check, the beatbox-play, etc. Further, the controller 10 controls the played beats to be varied based on the motion of the mobile communication terminal detected by the motion recognition sensor unit 50 in a state where a beatbox-imitation screen has been displayed.
Although the present embodiment includes music information relating to the beatbox or the music box for illustrative purposes, the present embodiment may include only some of the music information relating to the beatbox and the music information relating to the music box, or may include other music information.
When predetermined sounds are output from the beatbox or the music box, background colors may be varied.
When the user commands performance of a music box function through the user interface unit 40, the controller 10 detects the user's command and controls a list screen of tunes to be played by the music box to be displayed on the display unit 30 so that the user may select predetermined tunes. Herein, when the tune selected by the user is a tune (e.g., “Voices Of Spring, Waltz” by Johann Strauss) having only musical notes without words, a background scene displayed on the display unit 30 may have a color corresponding to the musical notes of the tune as the corresponding tune continues to be played. Information on the color may be retrieved from the background scene storage space. Herein, the corresponding tune is played by the musical note. Whether to play the next musical notes is determined if the motion recognition sensor unit 50 has detected the motion of the mobile communication terminal. In other words, the musical notes of a tune to be played by the music box are predetermined, but the tune is played and the beat changes only when the user shakes the terminal, similarly to a beatbox as described later. That is, a motion detection interval corresponds to a play interval and a beat. In other words, whenever the motion of the terminal is detected, sounds corresponding to the musical notes of a played tune are output in sequence one by one. Further, because the selected tune does not have words, a specific avatar may be displayed on a background scene. For example, when the musical note includes “sol-sol-la-la-sol-sol-mi”, it is possible to change the color of the background scene in the sequence of red, red, blue, blue, red, red and yellow.
When a tune with words is selected, the controller 10 controls the display unit 30 to display a predetermined quantity of words of the tune to be displayed on the background scene. In this state, the controller 10 controls the tune to be played by the musical note and simultaneously controls display types of letters (letters of the words) corresponding to each musical note to be varied whenever the motion of the mobile communication terminal is detected by the motion recognition sensor unit 50. Herein, the display type may be variously embodied by flickering, colors, etc.
In a detailed example for the latter case, when it is assumed that a musical note “mi” is set as a yellow color and a musical note “pa” is set as a blue color, in a state where the words “Are You Sleeping?” are displayed with a black color on the screen and electronic sounds of the musical note “pa” corresponding to the words “You” is currently output, the words “Are You” have colors of yellow and blue because musical notes of the words “Are You” correspond to “mi pa” respectively, and the next word “Sleeping” maintains the black color. In another example, it is also possible to allow the words “Are You” to have a red color and the next word “Sleeping” to maintain a black color. This is a simple case where the colors of the words change according to whether the play is performed.
In the beatbox, different colors are output to the background scene when the terminal is shaken and thus predetermined sounds are output from sound sources and when the terminal is not shaken and thus the predetermined sounds are not output (no sound sources). For example, when the predetermined sounds are output, a red color is output. However, when the predetermined sounds are not output, it is possible to change a color in a predetermined sequence or maintain one color other than the red color. Further, in the beatbox, when the terminal is not shaken, beatbox play is not accomplished because the predetermined sounds are not output. However, when the terminal is shaken, the beatbox play is accomplished because the predetermined sounds are output according to the shake.
The motion recognition sensor unit 50 is a sensor for measuring information about motion of an object. The motion recognition sensor unit 50 incorporates an acceleration sensor for measuring acceleration in order to calculate a change in positions of the object, and an angular velocity sensor (also known as a gyroscope) for measuring angular velocity in order to calculate a change in a rotation angle of the object. In realizing the preferred embodiment of the present invention, the acceleration sensor or the angular velocity sensor may be used as described below but the subject matter of the present invention is not limited to the sensor itself.
There are various methods for detecting time points at which the terminal is shaken by means of the acceleration sensor. The shaking motion may cause a rapid change in acceleration. The rapid change in acceleration may be expressed by an acceleration having a large slope on a time axis. When a predetermined threshold value has been set for a slope and a slope of a measured acceleration has exceeded this threshold value, it may be determined that the shake has been detected. In another method, the shaking motion greatly increases an acceleration value. When a predetermined threshold value has been set for the acceleration value and a measured acceleration value has exceeded this threshold value, it may be determined that a shake has been detected. In yet another method, it may be determined that the shake has been detected by applying the above two methods.
There are various methods for detecting time points at which the terminal is shaken by means of the angular velocity sensor. The shaking motion causes a rapid change in angular velocity. The rapid change in the angular velocity may be shown as a large slope of the angular velocity on a time axis. When a predetermined threshold value for a slope has been set and a slope of a measured angular velocity has exceeded this threshold value, it may be determined that the shake has been detected. In another method, the shaking motion greatly increases an angular velocity value. When a predetermined threshold value for the angular velocity value has been set and a measured angular velocity value has exceeded this threshold value, it may be determined that the shake has been detected. In further another method, it may be determined that the shake has been detected by applying the above two methods.
In step 611, the controller 10 controls a beatbox-setting screen to be displayed. In step 612, the controller 10 determines if a menu key is input. As a result of the determination, when the menu key input is detected, the controller 10 controls a beatbox-menu screen to be displayed in step 613. In step 614, the controller 10 determines if a user has selected one of beatbox items displayed on the screen. The user may select the item by inputting the up and the down direction key and the confirmation key as illustrated in
However, when the selection has not been ended, the controller 10 controls an instrument setting screen to be displayed in step 617. In step 619, the controller 10 determines if the user has selected one of the instruments displayed on the screen. In the afore described
However, when the selection has not been ended, the controller 10 controls a volume setting screen to be displayed in step 622. In step 623, the controller 10 determines if the user has selected one of the instruments displayed on the screen. In the afore described
In step 711 of
However, when the shake of the terminal is not detected, the controller 10 determines if the confirmation key input is detected in step 720 of
However, when the menu key input is not detected in step 726, the controller 10 determines if the confirmation key is input in step 727. When the confirmation key input is detected, the controller 10 displays the motion-beatbox screen and ends the procedure.
In step 811, the controller 10 controls a motion-beatbox menu screen to be displayed. In step 812, the controller 10 determines if a “My Instrument” item has been selected. When it is determined that the “My Instrument” item has been selected, the controller 10 controls a “My Instrument” menu screen to be displayed in step 813. In step 814, the controller 10 determines if a “Customizing” sub-item has been selected. When the “Customizing” sub-item has been selected, the controller 10 controls a recording time setting screen to be displayed in step 815. In step 816, the controller 10 determines if the recording time has been set. When it is detected that the recording time has been set, the controller 10 controls prepares recording and displays a bar graph in step 817. When the recording preparation is ended, the recording starts. In step 818, the controller 10 determines if the recording has been completed. When the recording has been completed, the controller 10 inputs the name of a recorded file in step 819. Further, in step 820, the controller 10 stores the recorded file to have the input name and controls the “My Instrument” menu screen to be displayed.
In step 911, the controller 10 controls the motion-beatbox menu screen to be displayed. In step 912, the controller 10 determines if the “My Instrument” item has been selected. When it is determined that the “My Instrument” item has been selected, the controller 10 controls the “My Instrument” menu screen to be displayed in step 913. In step 914, the controller 10 determines if the storage space has been selected. When the storage space has been selected, the controller 10 controls a storage-space-menu screen to be displayed in step 915. In step 916, the controller 10 determines if the “Voice 1” sub-item has been simply selected or the detailed information of the “Voice 1” sub-item has been selected. When the “Voice 1” sub-item has been simply selected, the controller 10 displays a “Replay Of The Voice 1” in step 917. However, when the detailed information has been selected, the controller 10 controls a detailed information screen to be displayed in step 918.
In step 1011, the controller 10 controls the motion-beatbox menu screen to be displayed on the display unit 30. In step 1012, the controller 10 determines if a music box item has been selected. When it is determined that the music box item has been selected, the controller 10 controls a music-box-menu screen to be displayed on the display unit 30 in step 1013. In step 1014, the controller 10 determines if a user has selected one of tunes displayed on the music-box-menu screen. When a predetermined tune has been selected, the controller 10 controls a tune information screen of the selected tune to be displayed on the display unit 30 in step 1015. After displaying the tune information screen, the controller 10 detects an nth shake of the terminal in step 1016 and outputs sounds of an nth musical note of the tune in step 1017. In step 1018, the controller 10 determines if all musical notes of the tune has been played in step 1018. When all musical notes have been played, the procedure ends. However, when all musical notes have not been played, control flow returns to step 1016.
In step 1111, the controller 10 determines if a current mode is a beatbox-learning-play mode. When the current mode is the beatbox-learning-play mode, the controller 10 determines if sounds are output from a first sound source in step 1112. When the sounds are output from the first sound source, the controller 10 outputs a background scene of a third color in step 1116. However, when the sounds are not output from the first sound source, the controller 10 determines if the sounds are output from a second sound source in step 1113. When the sounds are output from the second sound source, the controller 10 outputs a background scene of a second color in step 1117. However, when the sounds are not output from the second sound source, the controller 10 determines if a sound source exists in step 1114. When there is no sound source, the controller 10 outputs a background scene of the first color in step 1115.
Each play depends on whether the motion of the terminal is detected. Further, because a relation between the motion detection and the play has been sufficiently described in
Referring to
When it is determined that the current mode is not the beatbox-learning-play mode in step 1111, the controller 10 determines if the current mode is a music-box-play mode in step 1118. When it is determined that the current mode is the music-box-play mode, the controller 10 displays words of a selected tune on the screen in step 1119. In step 1120, the controller 10 plays sounds corresponding to the musical notes of the correspondent tune and simultaneously changes the color of letters corresponding to the musical notes from the first color (e.g.,
Further, in a state where the musical notes of the corresponding tune are played one by one whenever the shake is detected, it is also possible to output a background scene with colors according to each musical note with reference to the background storage space of the sound file storage unit 23.
Further, in a state where a tune has been selected and then a predetermined quantity of words with one color have been displayed on the screen, it is also possible to change the color of letters corresponding to currently played musical notes to a color corresponding to the musical notes.
Furthermore, it is also possible to play a plurality of musical notes simultaneously or sequentially whenever the shake of the terminal is detected. In this way, it is possible to adaptively achieve chord play or a predetermined interval (e.g., phrase) play based on the detection of the shake. For example, in a state where melodies and chords for a predetermined tune are stored in the sound file storage unit 23 in advance, when a user shakes the terminal once, one sound and a chord corresponding to the sound can be played. In another embodiment, in a state where only chords for a predetermined tune are stored in the sound file storage unit 23 in advance, it is possible to play the tune by one chord whenever a user shakes the terminal once. This case can be used when a user wants to directly sing a song and simultaneously play chords of the song based on the shake of the terminal. In yet another embodiment, in a state where a predetermined tune is separately stored by the phrase in the sound file storage unit 23 in advance, it is possible to sequentially play the tune by one phrase whenever a user shakes the terminal.
According to the present invention as described above, sounds dependent on a motion of a mobile communication terminal are output, so that a user can control the motion of the mobile communication terminal and change and output the sounds vividly as if the user actually plays an instrument. Further, according to the present invention, a user can customize “My Instrument” and store the “My Instrument” in a mobile communication terminal, so that sound sources can be selected and types of a beatbox can also be selected. Therefore, it is possible to be more interesting in a beatbox play. In other words, the present invention allows a user to select various beatboxes and various sound sources in a “Customizing My Instrument”, so that it is possible to provide beatbox environments preferred by each user. Furthermore, according to the present invention, a color of a background scene can change variously during a sound function such as a beatbox play and a music box play, so that it is possible to provide a user with visual entertainment as well as auditory entertainment. Therefore, the entertainment can be doubled.
While the present invention has been shown and described with reference to certain preferred embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the present invention as defined by the appended claims. For example, in the above detailed description, a case where sounds are output in response to a reciprocating motion of a mobile communication terminal is described as an example. However, it will be understood by those skilled in the art that the present invention can be applied when all motions of a mobile communication terminal exist. For example, even when a mobile communication terminal moves in only one direction, the present invention can be realized so that sounds dependent on the motion can be output. Further, a played tune can be stored in a memory and can be reproduced later.
Number | Date | Country | Kind |
---|---|---|---|
67862-2004 | Aug 2004 | KR | national |
16294-2005 | Feb 2005 | KR | national |
30528-2005 | Apr 2005 | KR | national |