Embodiments of the present disclosure relate to the field of communications technologies, and in particular, to a message sending method and a mobile terminal.
With the arrival of the network era, more and more people are used to online chat. For example, a mobile terminal includes chat software, and two users may send a text, a voice, and the like to each other. For another example, in game software, a user may communicate with a teammate by using a text, a voice, and the like.
Using the chat software as an example, when the user uses the chat software to chat, steps that need to be performed include: selecting a chat object, opening a chat window with the chat object, entering a text, a voice, and the like in an input box at the bottom of the chat window, and tapping Send. To switch a chat object, the user needs to close the current chat window, select another chat object, and open a chat window.
It can be learned from the foregoing processes that when the user chat with a plurality of persons at the same time, closing and opening operation actions need to be repeated between a plurality of chat windows. Consequently, user operations are cumbersome, communication efficiency is reduced, and chat experience of the user is affected.
Embodiments of the present disclosure provide a message sending method, to resolve a problem that because a user repeatedly performs closing and opening operation actions between a plurality of chat windows, user operation are cumbersome, communication efficiency is reduced, and chat experience of the user is affected.
To resolve the foregoing technical problems, the present disclosure is implemented as follows:
According to a first aspect, an embodiment of the present disclosure provides a message sending method, including: receiving a first input from a user; obtaining, in response to the first input, a target message written by the user in a message editing window; receiving a second input from the user; identifying, in response to the second input, a target sending window at which the user controls the message editing window to arrive; and sending the target message to the target sending window.
According to a second aspect, an embodiment of the present disclosure further provides a mobile terminal, including: a first input receiving module, configured to receive a first input from a user; a first input response module, configured to obtain, in response to the first input, a target message written by the user in a message editing window; a second input receiving module, configured to receive a second input from the user; a second input response module, configured to identify, in response to the second input, a target sending window at which the user controls the message editing window to arrive; and a sending module, configured to send the target message to the target sending window.
According to a third aspect, an embodiment of the present disclosure further provides a mobile terminal, including a processor, a memory, and a computer program that is stored in the memory and that can run on the processor, and when the processor executes the computer program, the steps of the message sending method are implemented.
According to a fourth aspect, an embodiment of the present disclosure further provides a computer-readable storage medium, where the computer-readable storage medium stores a computer program, and when a processor executes the computer program, the steps of the message sending method are implemented.
In the embodiments of the present disclosure, the user may trigger display of the message editing window on any interface of the mobile terminal through the first input, and write the to-be-sent target message in the displayed message editing window, so that after writing of the target message is completed, the user directly drags the message editing window to the target sending window through the second input, where the target sending window is, for example, a chat window in which a message needs to be sent, and further, after the message editing window arrives at the target sending window, the target message is sent to the target sending window to complete sending. It can be learned from the foregoing processes that, the user may chat with an object corresponding to a chat window without opening any chat window. In particular, when the user chat with a plurality of persons at the same time, the user does not need to sequentially open a plurality of chat windows to send a message but only complete writing of a target message on one interface, and then drag the message editing window to a corresponding chat window, to prevent the user from repeatedly performing closing and opening operation actions between the plurality of chat windows, thereby simplifying a user operation, improving communication efficiency, and optimizing chat experience.
The following clearly describes the technical solutions in the embodiments of the present disclosure with reference to the accompanying drawings in the embodiments of the present disclosure. Apparently, the described embodiments are some but not all of the embodiments of the present disclosure. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments of the present disclosure shall fall within the protection scope of the present disclosure.
Referring to
Step S1: Receive a first input from a user.
A main function implemented by the method in this embodiment is to send a message. The method includes two processes: message pre-editing and message sending. Message pre-editing means that the user pre-edits a message on any interface of the mobile terminal, and message sending means that the user drags the pre-edited message to a window to complete sending.
The first input is used to complete message editing on any interface.
Correspondingly, the first input includes a series of operations such as tapping, sliding, and pressing.
For example, the user triggers display of a message editing window on any interface through the first input, so that the user writes a target message in the message editing window. Herein, the operation of triggering display of the message editing window and the operation of writing the target message are all included in the first input.
Step S2: Obtain, in response to the first input, a target message written by the user in a message editing window.
Content of the target message is obtained in response to an operation input of writing the target message in the message editing window by the user.
Step S3: Receive a second input from the user.
The second input is used to drag a pre-edited message to a window to complete sending.
Correspondingly, the second input includes a series of operations such as tapping, sliding, pressing, and dragging.
For example, any interface in which the user is located includes at least one window. Generally, a window list is displayed on a main interface of chat software, and the window list includes a plurality of chat windows. When the mobile terminal displays the main interface of the chat software, the user may trigger display of the message editing window through the first input, and write the target message in the message editing window through the first input. Further, the user drags the message editing window to one of the chat windows, and confirms that the target message of the message editing window is sent to the chat window. Herein, the operation of dragging the message editing window and the operation of confirming sending are all included in the second input.
Specifically, after the user drags the message editing window to one of the chat windows, a “Send” option is displayed. The user taps the “Send” option to confirm sending. Tap the “Send” option herein is an operation of confirming sending.
Step S4: Identify, in response to the second input, a target sending window at which the user controls the message editing window to arrive.
Optionally, the second input includes the operation in which the user drags the message editing window to the target sending window, so that in response to the second input, the user controls, based on a drag track of the user in the second input, the message editing window to move to the target sending window on an interface based on the drag track.
Further, after the message editing window arrives at the target sending window, the target sending window is identified, and the identified target sending window is highlighted, to facilitate identification by the user.
Step S5: Send the target message to the target sending window.
In this step, the second input further includes an operation of confirming sending by the user, to send the target message to the target sending window in response to the second input.
For example, based on the highlighted target sending window, the user may confirm the target sending window, and then release a finger from a screen, and send the target message to the target sending window to complete sending.
In this embodiment of the present disclosure, the user may trigger display of the message editing window on any interface of the mobile terminal through the first input, and write the to-be-sent target message in the displayed message editing window, so that after writing of the target message is completed, the user directly drags the message editing window to the target sending window through the second input, where the target sending window is, for example, a chat window in which a message needs to be sent, and further, after the message editing window reaches the target sending window, the target message is sent to the target sending window to complete sending.
It can be learned from the foregoing processes that, the user may chat with an object corresponding to a chat window without opening any chat window. In particular, when the user chat with a plurality of persons at the same time, the user does not need to sequentially open a plurality of chat windows to send a message but only complete writing of a target message on one interface, and then drag the message editing window to a corresponding chat window, to prevent the user from repeatedly performing closing and opening operation actions between the plurality of chat windows, thereby simplifying a user operation, improving communication efficiency, and optimizing chat experience.
For example, as shown in
Based on the embodiment shown in
Step S2 includes:
Step S21: Obtain the message editing window in response to the display sub-input.
Optionally, the display sub-input includes a touch gesture action, an air gesture action, and the like on the screen.
Referring to
In this step, before the message editing window is displayed, the message editing window is first obtained.
Step S22: Perform transparency processing on a display area of the message editing window.
Before the message editing window is displayed, transparency processing is performed on the display area of the message editing window. In this way, the message editing window displayed on the interface can be prevented from occluding display content on the interface.
Step S23: Display the message editing window.
Optionally, the message editing window may be displayed on the interface in a pop-up form. Further, the popped-up message editing window is displayed in a hover box, the hovering message editing window is independent of the current interface, and the user may drag the message editing window to move freely in the current interface.
Optionally, the message editing window may be an applet, a widget, or the like. Referring to
In this embodiment, the user may trigger display of the message editing window in any interface through the display sub-input, where the display sub-input is not limited to a simple air gesture action and a touch gesture action. The operation of triggering display of the message editing window by the user is simple and convenient, and the message editing window can be displayed anytime, to complete a message sending function in this embodiment. Further, the message editing window is processed before display, and is displayed on the interface in a transparent hovering state, to prevent the message editing window from occluding display content on the interface.
Based on the foregoing embodiment,
Step S2 further includes:
Step S24: Display the target message in the message editing window in response to the write sub-input.
Referring to
In a process in which the user performs the write sub-input, a size of the display area of the message editing window is adjusted based on content of the target message written by the user until all content of the target message is displayed in the display area of the message editing window.
It should be noted that in the related art, after opening a chat window, the user writes message content in an input box at the bottom of the chat window. Generally, due to a size limitation of the chat window, an area of the input box is relatively small, and message content written by the user is relatively long, the written message content is not displayed for privacy, thereby making it inconvenient for the user to view all message content. In this case, most users write messages in advance in editing software (such as memo and note software), and then write the messages into the input box by copying and pasting. In this way, the users are switched between a plurality of pieces of software, and operations are cumbersome.
In this embodiment, the display area of the message editing window may vary, and changes with message content written by the user in real time. For example, when the user writes less message content, the display area of a message editing window is smaller, to prevent a large display area from affecting a current interface. For another example, as the message content written by the user increases, the display area of the message editing window gradually increases, to ensure that the user can see all written content in the display area of the message editing window at any time. In this way, the user can adjust written message content in real time, and the size of the display area of the message editing window is automatically adjusted without the user's operation, so that user operations are simplified.
Further, to prevent an excessively large display area of the message editing window from affecting the current interface, or to prevent the display area of the message editing window from being greater than a display area of a display screen, a maximum value of an area of the display area of the message editing window may be preset. When the display area of the message editing window is adjusted to the maximum value of the area, an adjustment change is stopped.
Based on the embodiment shown in
Step S25: Obtain, in response to the first input, a target message written by the user in any sub-window of the message editing window.
Optionally, the message editing window includes a plurality of sub-windows, and each sub-window is used by the user to write an independent target message. Referring to
In actual application, after the user calls out the message editing window through the first input, the first sub-window may be first displayed. After writing the target message in the first sub-window, the user taps the indicator arrow 3 of the next sub-window, and the message editing window 1 switches to display the next sub-window. The user continues to write a next target message in the next sub-window. Alternatively, the user taps the indication arrow 2 of the previous sub-window, and the message editing window 1 switches to display the previous sub-window, where the written target message is displayed in the previous sub-window, and the user may edit, view, and the like the target message in the previous sub-window. The user may switch a plurality of sub-windows for display, and the plurality of sub-windows do not affect each other. Blank content may be displayed on a sub-window in which no message is written.
In this step, the user may enable, through the first input, the message editing window to display a sub-window in which the target message is written, so that the mobile terminal obtains the target message currently displayed in the message editing window, and sends the obtained target message to the target sending window after the message editing window is dragged to the target sending window.
In this embodiment, the message editing window includes a plurality of sub-windows, and the user may write a plurality of target messages in advance in the message editing window, to separately send the plurality of target messages to corresponding target sending windows. It can be learned that when the user uses chat software, a plurality of chat windows do not need to be opened repeatedly. The user may view messages of the plurality of chat windows on an interface in which a chat window list is located, and edit a plurality of messages. For example, in
In addition, in one case of the related art, when the user uses chat software, if the user edits message content in an input box of a chat window, the input box is enabled, and a terminal of the other party prompts “Tying . . . ”. Sometimes, the user may only write or delete, or think about organizing a language, and does not want the other party to see the state of “Tying . . . ”. In this embodiment, the user does not need to open a chat window, and therefore, an input box does not need to be enabled, thereby effectively resolving the foregoing problem and improving user experience.
In another case of the related art, if the user is in a chat window, for some special reasons, for example, chat content is relatively important, or a red packet is being snatched, the user cannot close a current chat window in a timely manner. As a result, the user cannot view a message in another chat window in a timely manner. In this embodiment, the user does not need to open any chat window, and may flexibly view messages of a plurality of chat windows on a primary interface, and send messages separately based on the plurality of chat windows, thereby effectively resolving the foregoing problem and improving user experience.
For ease of detailed description, another embodiment of the present disclosure is used as an example for specific explanation, to describe an implementation process of message sending in this embodiment.
In this embodiment, the second input includes a touching and holding operation performed by the user on the message editing window, and in response to the touching and holding operation performed by the user on the message editing window, the message editing window is activated and changes from a still state to a movable state. In this way, a phenomenon that the message editing window is incorrectly moved due to a misoperation of the user can be avoided.
Referring to
Referring to
Referring to
Alternatively, the second input further includes an operation of tapping the “Move” option by the user, the user does not send the target message in response to the operation of tapping the “Move” option by the user, and keeps the message editing window at a current position. Herein, the operation of tapping the “Move” option by the user can be used to move the message editing window without sending a message.
In this embodiment, the second input is further explained in detail. Before the target message is sent, a service for confirming sending is added to the user, to avoid a missending event of the user. Compared with the related art, when the user edits a message in an input box of a chat window, missending is often caused by a mistouching of a “Send” button. This embodiment effectively resolves this problem and further improves user experience. In addition, in a case that the user does not select to send, a service of moving the mobile message editing window is further provided for the user, so that the user can move the message editing window randomly on an interface.
In another embodiment of the present disclosure, if a target message in written in only one of a plurality of sub-windows in the message editing window, the message editing window automatically exits after the target message is successfully sent. If target messages are separately written in at least two sub-windows, after a current target message is successfully sent, the message editing window automatically switches to a next sub-window to display an unsent target message.
In more embodiments, a close button is displayed on the message editing window, and the user taps the close button to exit the message editing window. Alternatively, the user may trigger the message editing window to exit through an air gesture action or a touch gesture action.
In particular, the message sending method in this embodiment of the present disclosure is applied to a single-screen mobile terminal, a dual-screen mobile terminal, and more mobile terminals. In the dual-screen mobile terminal, the user may perform message pre-editing and message sending on one screen, or may perform message pre-editing and message sending on two screens, to make full use of a feature of the dual-screen mobile terminal.
Based on the embodiment shown in
Step S4 includes:
Step S41: Control, in response to the second input, the message editing window to move from the first screen to a target sending window of the second screen.
Step S42: Identify the target sending window of the second screen.
In this embodiment, the user may trigger display of the message editing window on any interface of the first screen through the first input, and the message editing window may be displayed in any screen. After completing editing of the target message in the message editing window, the user may drag the message editing window from one screen to a target sending window of another screen to complete sending.
For example, the dual-screen mobile terminal includes two screens, and the two screens are a primary screen and a secondary screen. Because the primary screen is a screen with higher use frequency, the user may trigger display of the message editing window on the primary screen through the first input, and the message editing window may be optionally displayed on the secondary screen. This prevents the message editing window from interfering with the user's operation on the primary screen, so that the user can edit the target message in a message editing window of the secondary screen. After the user completes editing, the message editing window is dragged to a target sending window of the primary screen to complete sending.
In particular, when a game interface is displayed on the primary screen, a game screen is occluded if the user edits a message on the game interface, and a message input box of the game interface is too small. In addition, the user can only edit a message on a game side, thereby affecting game experience of the user. In this embodiment, the user may write a plurality of commonly used messages in advance in the message editing window displayed on the secondary screen, and when a message needs to be sent, the user directly drags the message editing window to a corresponding position on the primary screen to complete sending, thereby ensuring both game experience and message sending experience.
The dual-screen mobile terminal includes a collapsible screen and a flexible screen.
a first input receiving module 10, configured to receive a first input from a user;
a first input response module 20, configured to obtain, in response to the first input, a target message written by the user in a message editing window;
a second input receiving module 30, configured to receive a second input from the user;
a second input response module 40, configured to identify, in response to the second input, a target sending window at which the user controls the message editing window to arrive;
and a sending module 50, configured to send the target message to the target sending window.
In this embodiment of the present disclosure, the user may trigger display of the message editing window on any interface of the mobile terminal through the first input, and write the to-be-sent target message in the displayed message editing window, so that after writing of the target message is completed, the user directly drags the message editing window to the target sending window through the second input, where the target sending window is, for example, a chat window in which a message needs to be sent, and further, after the message editing window reaches the target sending window, the target message is sent to the target sending window to complete sending. It can be learned from the foregoing processes that, the user may chat with an object corresponding to a chat window without opening any chat window. In particular, when the user chat with a plurality of persons at the same time, the user does not need to sequentially open a plurality of chat windows to send a message but only complete writing of a target message on one interface, and then drag the message editing window to a corresponding chat window, to prevent the user from repeatedly performing closing and opening operation actions between the plurality of chat windows, thereby simplifying a user operation, improving communication efficiency, and optimizing chat experience.
Optionally, the first input includes at least a display sub-input of the message editing window; and
the first input response module 20 includes:
a display sub-input response unit, configured to obtain the message editing window in response to the display sub-input;
a processing unit, configured to perform transparency processing on a display area of the message editing window; and
a display unit, configured to display the message editing window.
Optionally, the first input further includes at least a write sub-input of the target message; and
the first input response module 20 further includes:
a write sub-input response unit, configured to display the target message in the message editing window in response to the write sub-input, where
in a process in which the user performs the write sub-input, a size of the display area of the message editing window is adjusted based on content of the target message written by the user until all content of the target message is displayed in the display area of the message editing window.
Optionally, the first input response module 20 includes:
a sub-window writing unit, configured to obtain, in response to the first input, a target message written by the user in any sub-window of the message editing window.
Optionally, the mobile terminal is a dual-screen mobile terminal, and the mobile terminal includes a first screen and a second screen; and
the second input response module 40 includes:
a cross-screen moving unit, configured to control, in response to the second input, the message editing window to move from the first screen to a target sending window of the second screen; and
a window identifying unit, configured to identify the target sending window of the second screen.
The mobile terminal provided in this embodiment of the present disclosure can implement the processes implemented by the mobile terminal in the method embodiments in
The user input unit 107 is configured to: receive a first input from a user; and receive a second input from the user; and
the processor 110 is configured to: obtain, in response to the first input, a target message written by the user in a message editing window; identify, in response to the second input, a target sending window at which the user controls the message editing window to arrive; and send the target message to the target sending window.
In this embodiment of the present disclosure, the user may trigger display of the message editing window on any interface of the mobile terminal through the first input, and write the to-be-sent target message in the displayed message editing window, so that after writing of the target message is completed, the user directly drags the message editing window to the target sending window through the second input, where the target sending window is, for example, a chat window in which a message needs to be sent, and further, after the message editing window reaches the target sending window, the target message is sent to the target sending window to complete sending. It can be learned from the foregoing processes that, the user may chat with an object corresponding to a chat window without opening any chat window. In particular, when the user chat with a plurality of persons at the same time, the user does not need to sequentially open a plurality of chat windows to send a message but only complete writing of a target message on one interface, and then drag the message editing window to a corresponding chat window, to prevent the user from repeatedly performing closing and opening operation actions between the plurality of chat windows, thereby simplifying a user operation, improving communication efficiency, and optimizing chat experience.
It should be understood that, in this embodiment of the present disclosure, the radio frequency unit 101 may be configured to receive and send information or a signal in a call process. Specifically, after receiving downlink data from a base station, the radio frequency unit 101 sends the downlink data to the processor 110 for processing. In addition, the radio frequency unit 101 sends uplink data to the base station. Usually, the radio frequency unit 101 includes but is not limited to an antenna, at least one amplifier, a transceiver, a coupler, a low noise amplifier, a duplexer, and the like. In addition, the radio frequency unit 101 may communicate with a network and another device through a wireless communication system.
The mobile terminal provides wireless broadband Internet access for the user by using the network module 102, for example, helping the user to send and receive an e-mail, browse a web page, and access streaming media.
The audio output unit 103 may convert audio data received by the radio frequency unit 101 or the network module 102 or stored in the memory 109 into an audio signal and output the audio signal as a sound. In addition, the audio output unit 103 may further provide an audio output (for example, a call signal received voice, or a message received voice) related to a specific function implemented by the mobile terminal 100. The audio output unit 103 includes a speaker, a buzzer, a telephone receiver, and the like.
The input unit 104 is configured to receive an audio signal or a video signal. The input unit 104 may include a graphics processing unit (Graphics Processing Unit, GPU) 1041 and a microphone 1042. The graphics processing unit 1041 processes image data of a static picture or a video obtained by an image capturing apparatus (for example, a camera) in a video capturing mode or an image capturing mode. A processed image frame may be displayed on the display unit 106. The image frame processed by the graphics processing unit 1041 may be stored in the memory 109 (or another storage medium) or sent via the radio frequency unit 101 or the network module 102. The microphone 1042 may receive a sound and can process such sound into audio data. Processed audio data may be converted, in a call mode, into a format that can be sent to a mobile communication base station by using the radio frequency unit 101 for output.
The mobile terminal 100 may further include at least one sensor 105 such as an optical sensor, a motion sensor, or another sensor. Specifically, the optical sensor includes an ambient light sensor and a proximity sensor. The ambient light sensor may adjust luminance of the display panel 1061 based on brightness of ambient light, and the proximity sensor may disable the display panel 1061 and/or backlight when the mobile terminal 100 approaches an ear. As a type of the motion sensor, an accelerometer sensor may detect an acceleration value in each direction (generally, three axes), and detect a value and a direction of gravity when the accelerometer sensor is static, and may be used in an application for recognizing a mobile terminal posture (such as screen switching between landscape and portrait modes, a related game, or magnetometer posture calibration), a function related to vibration recognition (such as a pedometer or a knock), and the like. The sensor 105 may further include a fingerprint sensor, a pressure sensor, an iris sensor, a molecular sensor, a gyroscope, a barometer, a hygrometer, a thermometer, an infrared sensor, and the like. Details are not described herein.
The display unit 106 is configured to display information entered by a user or information provided for a user. The display unit 106 may include a display panel 1061. The display panel 1061 may be configured in a form of a liquid crystal display (Liquid Crystal Display, LCD), an organic light-emitting diode (Organic Light-Emitting Diode, OLED), or the like.
The user input unit 107 may be configured to: receive digit or character information that is input, and generate key signal input related to user setting and function control of the mobile terminal. Specifically, the user input unit 107 includes a touch panel 1071 and another input device 1072. The touch panel 1071 is also referred to as a touchscreen, and may collect a touch operation performed by a user on or near the touch panel 1071 (such as an operation performed by a user on the touch panel 1071 or near the touch panel 1071 by using any proper object or accessory, such as a finger or a stylus). The touch panel 1071 may include two parts: a touch detection apparatus and a touch controller. The touch detection apparatus detects a touch position of the user, detects a signal brought by the touch operation, and sends the signal to the touch controller. The touch controller receives touch information from the touch detection apparatus, converts the touch information into touch point coordinates, and sends the touch point coordinates to the processor 110, and can receive and execute a command sent by the processor 110. In addition, the touch panel 1071 may be of a resistive type, a capacitive type, an infrared type, a surface acoustic wave type, or the like. The user input unit 107 may include another input device 1072 in addition to the touch panel 1071. Specifically, the another input device 1072 may include but is not limited to a physical keyboard, a functional button (such as a volume control button or a power on/off button), a trackball, a mouse, and a joystick. Details are not described herein.
Further, the touch panel 1071 may cover the display panel 1061. When detecting the touch operation on or near the touch panel 1071, the touch panel 1071 transmits the touch operation to the processor 110 to determine a type of a touch event, and then the processor 110 provides corresponding visual output on the display panel 1061 based on the type of the touch event. In
The interface unit 108 is an interface for connecting an external apparatus with the mobile terminal 100. For example, the external apparatus may include a wired or wireless headset port, an external power supply (or a battery charger) port, a wired or wireless data port, a memory card port, a port for connecting an apparatus having an identification module, an audio input/output (I/O) port, a video I/O port, a headset port, and the like. The interface unit 108 may be configured to receive input (for example, data information and power) from an external apparatus and transmit the received input to one or more elements in the mobile terminal 100 or may be configured to transmit data between the mobile terminal 100 and an external apparatus.
The memory 109 may be configured to store a software program and various data. The memory 109 may mainly include a program storage area and a data storage area. The program storage area may store an operating system, an application required by at least one function (such as a sound play function or an image play function), and the like. The data storage area may store data (such as audio data or an address book) created based on use of the mobile phone, and the like. In addition, the memory 109 may include a high-speed random-access memory, or may include a nonvolatile memory, for example, at least one disk storage device, a flash memory, or another volatile solid-state storage device.
The processor 110 is a control center of the mobile terminal, and is connected to all parts of the entire mobile terminal by using various interfaces and lines, and performs various functions of the mobile terminal and processes data by running or executing the software program and/or the module that are stored in the memory 109 and invoking the data stored in the memory 109, to implement overall monitoring on the mobile terminal. The processor 110 may include one or more processing units. Optionally, the processor 110 may integrate an application processor and a modem processor. The application processor mainly processes an operating system, a user interface, an application, and the like. The modem processor mainly processes wireless communication. It may be understood that, alternatively, the modem processor may not be integrated into the processor 110.
The mobile terminal 100 may further include a power supply 111 (such as a battery) that supplies power to each component. Optionally, the power supply 111 may be logically connected to the processor 110 by using a power supply management system, to implement functions such as charging, discharging, and power consumption management by using the power supply management system.
In addition, the mobile terminal 100 includes some function modules not shown, and details are not described herein.
Optionally, an embodiment of the present disclosure further provides a mobile terminal, including a processor 110, a memory 109, and a computer program that is stored in the memory 109 and that can run on the processor 110. When the processor 110 executes the computer program, the foregoing processes of the message sending method embodiment are implemented and a same technical effect can be achieved. To avoid repetition, details are not described herein again.
An embodiment of the present disclosure further provides a computer-readable storage medium. The computer-readable storage medium stores a computer program, and when a processor executes the computer program, the foregoing processes of the message sending method embodiment are implemented and a same technical effect can be achieved. To avoid repetition, details are not described herein again. The computer-readable storage medium may be a read-only memory (Read-Only Memory, ROM), a random-access memory (Random Access Memory, RAM), a magnetic disk, a compact disc, or the like.
It should be noted that, in this specification, the terms “include”, “comprise”, or their any other variant is intended to cover a non-exclusive inclusion, so that a process, a method, an article, or an apparatus that includes a list of elements not only includes those elements but also includes other elements which are not expressly listed, or further includes elements inherent to such process, method, article, or apparatus. In the absence of more restrictions, an element defined by the statement “including a . . . ” does not exclude another same element in a process, method, article, or apparatus that includes the element.
Based on the descriptions of the foregoing implementations, a person skilled in the art may clearly understand that the method in the foregoing embodiment may be implemented by software in addition to a necessary universal hardware platform or by hardware only. In most circumstances, the former is a preferred implementation. Based on such an understanding, the technical solutions of the present disclosure essentially or the part contributing to the prior art may be implemented in a form of a software product. The computer software product is stored in a storage medium (such as a ROM/RAM, a hard disk, or an optical disc), and includes several instructions for instructing a terminal (which may be mobile phone, a computer, a server, an air conditioner, a network device, or the like) to perform the methods described in the embodiments of the present disclosure.
The embodiments of the present disclosure are described above with reference to the accompanying drawings, but the present disclosure is not limited to the foregoing specific implementations. The foregoing specific implementations are merely schematic instead of restrictive. Under enlightenment of the present disclosure, a person of ordinary skills in the art may make many forms without departing from the aims of the present disclosure and the protection scope of claims, all of which fall within the protection of the present disclosure.
Number | Date | Country | Kind |
---|---|---|---|
201910074925.3 | Jan 2019 | CN | national |
This application is a continuation application of International Application No. PCT/CN2020/071693 filed on Jan. 13, 2020, which claims priority to Chinese Patent Application No. 201910074925.3, filed on Jan. 25, 2019 in China, which is incorporated herein by reference in its entirety.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN20/71693 | Jan 2020 | US |
Child | 17383743 | US |