This disclosure relates to the technical field of terminals, and more particularly to a method for controlling a terminal, a terminal, and a non-transitory computer readable storage medium.
With the rapid development of terminal technology, a display screen of a terminal increases in size constantly, applications installed and functions that can be realized are more and more diversified, and a user may spend a lot of time every day to use the terminal. Therefore, power consumption of the terminal is an important indicator for measuring the performance of the terminal.
Control schemes of the terminal such as display and playback have a great influence on the power consumption of the terminal. The terminal has a problem of high power consumption. Therefore, the control schemes of the terminal needs to be improved.
The following is a summary of subjects of the detailed description, and this summary is not intended to limit the scope of the claims.
Implementations of the present disclosure provide a method and device for controlling a terminal, and a terminal, which can optimize control schemes of the terminal.
According to implementations of the present disclosure, a method for controlling a terminal is provided. The method includes the following. An operation area is determined according to a position of an input operation on a screen of the terminal. Power consumption of a target area other than the operation area is reduced with a scheme, where the scheme is selected from a group consisted of reducing an image resolution in the target area and reducing power consumption of layers corresponding to the target area.
According to implementations of the present disclosure, a terminal is provided. The terminal includes at least one processor and a computer readable storage. The computer readable storage is coupled to the at least one processor and stores at least one computer executable instruction thereon which, when executed by the at least one processor, causes the at least one processor to: determine an operation area according to a position of an input operation on a screen of the terminal; and reduce power consumption of a target area other than the operation area with a scheme, where the scheme is selected from a group consisted of reducing an image resolution in the target area and reducing power consumption of layers corresponding to the target area.
According to implementations of the present disclosure, a non-transitory computer readable storage medium is further provided. The non-transitory computer readable storage medium is configured to store a computer program which, when executed by a processor, causes the processor to carry out following actions. A target area is determined according to an input operation on a screen. A scheme for reducing power consumption of the target area is determined according to a layer allocation strategy of a currently running application, where the scheme is selected from a group consisted of reducing an image resolution in the target area and reducing power consumption of layers corresponding to the target area. Power consumption of the target area is reduced with the scheme determined.
According to the schemes for controlling a terminal provided in the implementations of the present disclosure, the power consumption of the target area can be selectively reduced according to the input operation of the user, thereby reducing system power consumption of the terminal.
Other aspects will become apparent upon reading and understanding the drawings and the detailed description.
Before discussing the exemplary implementations in detail, it should be mentioned that some exemplary implementations are described as processes or methods illustrated by flowcharts. In the flowcharts, although steps are depicted as being processed sequentially, some of the steps may be performed in parallel, concurrently, or simultaneously. In addition, the order of the steps can be rearranged without conflict. The process may be terminated when finishing an operation, but the process may include additional steps that are not illustrated in the drawings. The process may correspond to a method, a function, a procedure, a subroutine, a subprogram, and the like.
Hereinafter, implementations of the present disclosure will be described in detail.
According to an implementation of the disclosure, a method for controlling a terminal is provided. The method includes the following. An operation area is determined according to a position of an input operation on a screen of the terminal. Power consumption of a target area other than the operation area is reduced with a scheme, where the scheme is selected from a group consisted of reducing an image resolution in the target area and reducing power consumption of layers corresponding to the target area.
According to an implementation of the disclosure, a terminal is provided. The terminal includes at least one processor and a computer readable storage. The computer readable storage is coupled to the at least one processor and stores at least one computer executable instruction thereon which, when executed by the at least one processor, causes the at least one processor to: determine an operation area according to a position of an input operation on a screen of the terminal; and reduce power consumption of a target area other than the operation area with a scheme, where the scheme is selected from a group consisted of reducing an image resolution in the target area and reducing power consumption of layers corresponding to the target area.
According to an implementation of the disclosure, a non-transitory computer readable storage medium is further provided. The non-transitory computer readable storage medium is configured to store a computer program which, when executed by a processor, causes the processor to carry out following actions. A target area is determined according to an input operation on a screen. A scheme for reducing power consumption of the target area is determined according to a layer allocation strategy of a currently running application, where the scheme is selected from a group consisted of reducing an image resolution in the target area and reducing power consumption of layers corresponding to the target area. Power consumption of the target area is reduced with the scheme determined.
Implementations of the disclosure will be detailed below with reference to the accompanying drawings.
At block 120, an operation area is determined according to a position of an input operation on a screen of the terminal.
For example, the terminal in this implementation may be a device having a display screen, such as a mobile phone, a smart watch, a tablet computer, a game machine, a personal digital assistant, and a digital multimedia player.
In an implementation, the input operation may include a touch input operation or an eye-focus input operation. In the case that the input operation is the touch input operation, the screen of the terminal is a touch screen for example, and the touch input operation of the user is received via the touch screen. In the case that the input operation is the eye-focus input operation, the terminal may include a camera (e.g., a front camera or a rotatable camera disposed at a front surface of the terminal) disposed at the top of the screen of the terminal. The camera is configured to capture a face image. A focus position of user's eyes on the screen is determined by recognizing positions of eyeballs. In this way, the eye-focus input operation of the user is received. The input manner may be applicable to a terminal having a large screen for example.
For example, the position of the input operation of the user on the screen can be understood as a position that the user is interested in. As an example,
In an implementation, the method may include the following. A center position (or referred to as an operation center) of the input operation on the screen is determined. An operation object is determined according to coordinates of the center position on the screen. The operation area is determined according to an area where the operation object locates. One of advantages of this implementation is that the operating area can be accurately determined. Similar to the above example, in
For example, if the input operation is a touch input operation, the method may further include the following. The operation area is determined to be a closed area, where the closed area is formed by an operation trajectory of the touch input operation on the screen.
In an implementation, if the input operation is a touch input operation, the method may further include the following. The operation area is determined to be an area other than the closed area, where the closed area is formed by the operation trajectory of the touch input operation on the screen. Compared with the foregoing manner, this manner can accurately determine the operation area by enhancing the interaction between the terminal and the user, and can also be applicable to a case where an area that the user is not interested in is smaller than an area that the user is interested in.
The above two manners can be set by default or selected by the user. Before the user starts to circle an area, the user may be prompted, such as “please circle your area of interest” or “please circle an area where a power reduction operation needs to be performed”.
At block 130, power consumption of a target area other than the operation area is reduced.
The power consumption of the target area may be reduced by reducing an image resolution in the target area or reducing power consumption of layers corresponding to the target area.
For example, the target area may include all or part of areas on the screen other than the operation area. For example, according to system settings or user settings, a preset area of the screen can be set to be an area that cannot be determined as the target area. In this case, the target area is all the areas on the screen other than the operation area and the preset area. The preset area may vary according to a specific display scenario, where display scenarios can be distinguished according to factors such as currently running applications (or process identifiers of the applications), sensing data acquired by a sensor, touch data, and properties of layers of applications.
In this implementation, an operating system loaded in the terminal may be an Android® system, a windows phone (WP) operating system, a Linux system, an iphone operating system (IOS), or the like. For ease of illustration, in the following, this implementation will be described in conjunction with the Android® system.
The Android® system is taken as an example to briefly describe a process from generation of a to-be-displayed image to display of the to-be-displayed image in the Android® system.
First, in an application layer, each application (hereinafter, “APP” for short) contains one or more layers, and each of multiple applications APP 1, APP 2 . . . APP N performs a layer-rendering operation (i.e., rendering an image on a layer) separately according to application design conditions of each of the multiple applications (generally, determined by a corresponding Android® package (APK)). After the rendering operation, each application sends one or more layers rendered by each application to a layer composite module (e.g., surface flinger) that performs a layer composite operation.
Then, in a framework layer, a layer list containing all layers (including visible and invisible layers) is generated, and the layer list is defined as ListAll. The layer composite module selects the visible layers from the ListAll to form a visible-layer list, and the visible-layer list is defined as DisplayList. Then, the layer composite module finds an unoccupied frame buffer (FB) from three reusable frame buffers in the Android® system. In the unoccupied FB, according to application configuration information, for example, which layer should be bottomed, which layer should be topped, which area is visible and which area is transparent, etc., by means of a composite operation, the layers contained in the DisplayList are superimposed to obtain a final to-be-displayed image.
In a kernel layer, the to-be-displayed image is transmitted to a display hardware, where the display hardware includes a controller and a display screen, so that the to-be-displayed image is finally displayed on the display screen. The type of the display screen is, for example, a liquid crystal display (LCD).
In addition, during a display refresh process, for the Android® system, a vertical synchronization (Vsync) refresh mechanism is introduced.
In a layer-rendering process, a layer composite process, and a layer refresh display process of the terminal, three kinds of frame rates are involved, which are a rendering frame rate, a composite frame rate, and a refresh rate, respectively.
The rendering frame rate is used for triggering the layer composite module to perform the composite operation after each application finishes the layer-rendering operation. The rendering frame rate can also be understood as frames of layers rendered per unit time (e.g., one second). The rendering frame rate includes a rendering frame rate of an application and a rendering frame rate of a layer. There may be more than one application currently running in the Android® system, and each application may include multiple layers. For example, a video player application generally includes three layers: one layer (defined as U1) for displaying video contents, and two SurfaceView-type layers, where one layer (defined as U2) is set to display contents of bullet-screen, and another layer (defined as U3) is set to display user-interface (UI) controls (e.g., a playback progress bar, a volume control bar, and various control buttons, etc.) and display advertisements. The rendering frame rate of the application refers to the number of times that each application performs the rendering operation per unit time, where one or more layers may be rendered when performing a rendering operation. The rendering frame rate of the layer refers to the number of times that a layer of the same number or name (e.g., U1, U2, or U3) is triggered to be rendered per unit time.
The composite frame rate is used for compositing layers rendered by respective applications to a to-be-displayed image. The composite frame rate can also be understood as frames of an image composited per unit time.
The refresh rate refers to a frame rate according to which an image displayed on the display screen of the terminal is refreshed. The display screen may be refreshed at a refresh rate of 60 Hz.
As an implementation, a processing manner of reducing the image resolution in the target area may be applicable to a case where an area is directly divided only according to a display image, regardless of whether the operation area and the target area are in a same layer. Reducing the image resolution in the target area may be implemented in a layer-rendering stage or a layer composite stage. In an implementation, a coordinate range of the target area in a display image is sent to a layer composite module. According to the coordinate range of the target area, respective sub-coordinate ranges of the target area in multiple layers are calculated via the layer composite module. Each of the sub-coordinate ranges is sent to a corresponding application via the layer composite module, whereby respective applications reduce the image resolution in the target area according to respective sub-coordinate ranges when rendering the multiple layers corresponding to the target area. In this way, a rendering process can be simplified during a layer-rendering stage, thereby saving rendering time and reducing power consumption. In another implementation, the coordinate range of the target area in the display image is sent to the layer composite module, and the image resolution in the target area can be reduced in the layer composite process according to the coordinate range of the target area. In this way, the composite process can be simplified during a layer composite stage, and so composite time can be saved and power consumption can be reduced. The image resolution can be reduced by means of related schemes of image processing.
The main difference between reducing the power consumption of the layers corresponding to the target area and reducing the image resolution in the target area described above is the following. A resolution of a part of an image in a same layer can be reduced when reducing the image resolution in the target area. However, when reducing the power consumption of one layer corresponding to the target area, the power consumption of the entire layer is reduced.
In an implementation, reducing the power consumption of the layers corresponding to the target area may include the following processing manners. In a first processing manner, a rendering frame rate of each of the layers corresponding to the target area is reduced. In a second processing manner, a resolution of each of the layers corresponding to the target area is reduced. In a third processing manner, the layers corresponding to the target area are removed from a set of layers to-be-composited (e.g., the DisplayList). In a fourth processing manner, play volume of each of the layers corresponding to the target area is decreased.
For the first processing manner, as an implementation, the rendering frame rate of each of the layers can be reduced by lowering the frequency of a reference signal (e.g., a Vsync signal) for layer rendering. At this time, the frequency of a Vsync signal for rendering the layers corresponding to the target area is reduced during a layer-rendering process, and the frequency of a Vsync signal during a layer composite process and the frequency of a Vsync signal during a refresh display process remain unchanged. For example, the frequency of the Vsync signal during the layer-rendering process, the frequency of the Vsync signal during the layer composite process, and the frequency of the Vsync signal during the refresh display process are all 60 Hz, and when the frequency of the Vsync signal for rendering the layers corresponding to the target area is reduced from 60 Hz to 50 Hz, the frequency of the Vsync signal during the layer composite process and the frequency of the Vsync signal during the refresh display process are still 60 Hz. In addition, the rendering frame rate of each of the layers can also be reduced by changing a response mechanism of a layer-rendering operation in response to reference signals. As an example, make a response to a (2n+1)th signal received, and for a (2n)th signal received, no response is made. As another example, take n signals (e.g., five signals) as a group, make a response to signals having a first type of preset number (e.g., 1, 2, 4, 5) in each group, and for signals having a second type of preset number (e.g., 3) in each group, no response is made.
For the second processing manner, reducing the resolution of each of the layers corresponding to the target area refers to reducing an image resolution of each of the layers corresponding to the target area, which can be implemented in the layer-rendering stage.
For the third processing manner, the operation of removing the layers corresponding to the target area from the set of layers to-be-composited may be implemented via a layer composite module. For instance, a coordinate range of the target area in a display image is sent to the layer composite module. The layer composite module identifies identifiers (such as a name or a number) of the layers corresponding to the target area according to the coordinate range of the target area, and removes corresponding layers from the set of layers to-be-composited according to the identifiers. As such, contents of the layers corresponding to the target area are not contained in a composite image to-be-displayed.
For the fourth processing manner, as an implementation, the layers corresponding to the target area may be layers corresponding to a video advertisement or layers with a sound effect. In this situation, play volume of each of the layers corresponding to the target area can be decreased, and so power consumption of the terminal can be reduced.
The above four types of power reduction processing manners may be combined with each other. For example, the first processing manner and the fourth processing manner are adopted simultaneously, the first processing manner and the second processing manner are adopted simultaneously, the first processing manner, the second processing manner, and the fourth processing manner are adopted simultaneously, or the like.
According to the method for controlling a terminal provided in this implementation, the input operation of the user on the screen of the terminal can be received, the operation area can be determined according to the position of the input operation on the screen, and the power consumption of the target area other than the operation area can be reduced, where reducing the power consumption of the target area includes reducing the image resolution in the target area or reducing the power consumption of the layers corresponding to the target area. By adopting the above technical solution, the power consumption of the target area can be selectively reduced according to the input operation of the user, thereby reducing system power consumption of the terminal.
At block 701, a touch input operation of a user on a screen of a terminal is received.
At block 702, an operation area is determined to be an area other than a closed area, where the closed area is formed by an operation trajectory of the touch input operation on the screen.
At block 703, power consumption of a target area other than the operation area is reduced.
The power consumption of the target area may be reduced by reducing an image resolution in the target area or reducing power consumption of layers corresponding to the target area.
In an implementation, an area other than the operation area may be a closed area formed by an operation trajectory of the touch input operation on the screen.
The manner of reducing the power consumption of the target area may be selected according to a layer allocation strategy of each application. Similarly, taken the game application A as an example, if display contents of the game application A on the game area 801 and display contents of the game application A on the background area 802 are all rendered on a same layer, the image resolution in the target area can be reduced. If the display contents of the game application A on the game area 801 and the display contents of the game application A on the background area 802 are respectively rendered on different layers, the power consumption of the layers corresponding to the target area can be reduced, that is, power consumption of layers corresponding to the closed area 803 is reduced. For the specific processing, reference may be made to the related descriptions above, and it will not be described in further detail herein.
In this implementation, the area that the user is not interested in can be determined according to a circle selection operation on the screen, and power consumption of images or layers corresponding to the area that is not of interest can be reduced, thereby reducing power consumption of the terminal and prolonging battery life.
The input-operation receiving module 901 is configured to receive an input operation of a user on a screen of the terminal.
The operation-area determining module 902 is configured to determine an operation area according to a position of the input operation on the screen.
The processing module 903 is configured to reduce power consumption of a target area other than the operation area, where reducing the power consumption of the target area includes reducing an image resolution in the target area or reducing power consumption of layers corresponding to the target area.
According to the device for controlling a terminal of this implementation, the power consumption of the target area can be selectively reduced according to the input operation of the user, and so system power consumption of the terminal can be reduced.
In an implementation, the power consumption of the layers corresponding to the target area is reduced as follows. A rendering frame rate of each of the layers corresponding to the target area is reduced. A resolution of each of the layers corresponding to the target area is reduced. The layers corresponding to the target area are removed from a set of layers to-be-composited. Play volume of each of the layers corresponding to the target area is decreased.
In an implementation, the input operation includes a touch input operation or an eye-focus input operation.
In an implementation, the operation-area determining module includes an operation-center determining unit, an operation-object determining unit, and an operation-area determining unit. The operation-center determining unit is configured to determine a center position of the input operation on the screen. The operation-object determining unit is configured to determine an operation object according to coordinates of the center position on the screen. The operation-area determining unit is configured to determine the operation area according to an area where the operation object locates.
In an implementation, the input operation is the touch input operation, the operation-area determining module is configured to: determine the operation area to be a closed area, where the closed area is formed by an operation trajectory of the touch input operation on the screen; or determine the operation area to be an area other than the closed area, where the closed area is formed by the operation trajectory of the touch input operation on the screen.
In an implementation, the image resolution in the target area are reduced as follows. A coordinate range of the target area in a display image is sent to a layer composite module. According to the coordinate range of the target area, respective sub-coordinate ranges of the target area in multiple layers are calculated via the layer composite module. Each of the sub-coordinate ranges is sent to a corresponding application via the layer composite module, whereby respective applications reduce the image resolution in target area according to respective sub-coordinate ranges when rendering the multiple layer corresponding to the target area. In another implementation, the image resolution in the target area are reduced as follows. The coordinate range of the target area in the display image is sent to the layer composite module. According to the coordinate range of the target area, the image resolution in the target area are reduced via the layer composite module in a layer composite process.
It should be understood that the implementation process of the foregoing components will not be detailed herein and reference may be made to the foregoing method implementations.
Implementations of the present disclosure provide a terminal, the device for controlling a terminal, as part of the terminal, may be disposed in the terminal.
In an implementation, the at least one computer executable instruction operable with the at least one processor to reduce the power consumption of the layers corresponding to the target area is operable with the at least one processor to: reduce a rendering frame rate of each of the layers corresponding to the target area; reduce a resolution of each of the layers corresponding to the target area; remove the layers corresponding to the target area from a set of layers to-be-composited; or decrease play volume of each of the layers corresponding to the target area.
In an implementation, the input operation is an eye-focus input operation, the at least one computer executable instruction operable with the at least one processor to determine the operation area according to the position of the input operation on the screen is operable with the at least one processor to: obtain a face image; determine a focus position corresponding to the eye-focus input operation on the screen by recognizing positions of eyeballs in the face image; determine an operation object according to coordinates of the focus position on the screen; and determine the operation area according to an area where the operation object locates.
In an implementation, the input operation is a touch input operation, the at least one computer executable instruction operable with the at least one processor to determine the operation area according to the position of the input operation on the screen is operable with the at least one processor to: determine a center position of the touch input operation on the screen; determine an operation object according to coordinates of the center position on the screen; and determine the operation area according to an area where the operation object locates.
In an implementation, the input operation is a touch input operation, the at least one computer executable instruction operable with the at least one processor to determine the operation area according to the position of the input operation on the screen is operable with the at least one processor to: determine a closed area formed by an operation trajectory of the touch input operation on the screen; and determine the operation area according to the closed area.
In an implementation, the at least one computer executable instruction operable with the at least one processor to reduce the image resolution in the target area is operable with the at least one processor to: send a coordinate range of the target area in a display image to a layer composite module; calculate, via the layer composite module, respective sub-coordinate ranges of the target area in a plurality of layers according to the coordinate range of the target area; and send, via the layer composite module, each of the sub-coordinate ranges to a corresponding application, whereby respective applications reduce the image resolution in the target area according to respective sub-coordinate ranges when rendering the plurality of layers corresponding to the target area.
In an implementation, the at least one computer executable instruction operable with the at least one processor to reduce the image resolution in the target area is operable with the at least one processor to: send a coordinate range of the target area in a display image to a layer composite module; and reduce, via the layer composite module, the image resolution in the target area in a layer composite process according to the coordinate range of the target area.
The terminal further includes a peripheral interface 1003, a radio frequency (RF) circuit 1005, an audio circuit 1006, a speaker 1011, a power management chip 1008, an input/output (I/O) subsystem 1009, a touch screen 1012, other input/control devices 1010, and external ports 1004. These components communicate with each other via one or more communication buses or signal lines 1007.
It should be understood that the terminal 1000 illustrated is merely one example of a terminal, and the terminal 1000 may have more or fewer components than those illustrated in the figures. For example, two or more components may be combined, or different component configurations can be adopted in the terminal. The various components illustrated in the figures can be implemented in hardware, software, or a combination of hardware and software including one or more signal processing and/or one or more application specific integrated circuits.
The following describes a terminal, which takes a mobile phone as an example.
The memory 1001 can be accessed by the CPU 1002, the peripheral interface 1003, and so on. The memory 1001 may include a high-speed random access memory and may further include a non-transitory memory such as one or more magnetic disk storage devices, flash memory devices, or other volatile solid-state storage devices.
The peripheral interface 1003 is configured to connect input and output peripherals of the device to CPU 1002 and memory 1001.
The I/O subsystem 1009 can be configured to connect the input and output peripherals on the device, such as the touch screen 1012 and other input/control devices 1010, to the peripheral interface 1003. The I/O subsystem 1009 may include a display controller 10091 and one or more input controllers 10092 configured to control other input/control devices 1010. One or more input controllers 10092 are configured to receive electrical signals from or send electrical signals to other input/control devices 1010, and where other input/control devices 1010 may include a physical button (a press button, a rocker button, etc.), a dial, a slide switch, a joystick, a click wheel. It should be noted that the input controller 10092 can be coupled with any of a keyboard, an infrared port, a USB interface, and a pointing device such as a mouse.
The touch screen 1012 is an input interface and an output interface between the terminal and a user, and is configured to display a visual output to the user. The visual output may include graphics, text, icons, video, or the like.
The display controller 10091 in I/O subsystem 1009 is configured to receive electrical signals from or send electrical signals to touch screen 1012. The touch screen 1012 is configured to detect contact on the touch screen, and the display controller 10091 is configured to convert the contact detected into an interaction with a user interface object displayed on the touch screen 1012, that is, to realize human-computer interaction. The user interface object displayed on the touch screen 1012 may be an icon of a running game, an icon indicating connection to corresponding networks, and the like. It should be noted that the device may also include a light mouse, which is a touch sensitive surface that does not display a visual output, or an extension of a touch sensitive surface formed by the touch screen.
The RF circuit 1005 is configured to establish communication between the mobile phone and the wireless network (i.e., network side) and to transmit and receive data between the mobile phone and the wireless network, for example, transmit and receive short messages, emails, and the like. The RF circuit 1005 is configured to receive and transmit RF signals (which is also referred to as electromagnetic signals), to convert an electrical signal into an electromagnetic signal or convert an electromagnetic signal into an electrical signal, and to communicate with a communication network and other devices through the electromagnetic signals. The RF circuit 1005 may include known circuits for performing these functions, including but not limited to an antenna system, an RF transceiver, one or more amplifiers, a tuner, one or more oscillators, a digital signal processor, a coder-decoder (codec) chipset, a subscriber identity module (SIM), and so on.
The audio circuit 1006 is configured to receive audio data from the peripheral interface 1003, to convert the audio data into an electrical signal, and to transmit the electrical signal to the speaker 1011.
The speaker 1011 is configured to restore the voice signal received by the mobile phone from the wireless network via the RF circuit 1005 to sound and to play the sound to the user.
The power management chip 1008 is configured for power supply and power management of the hardware connected to the CPU 1002, the I/O subsystem 1009, and the peripheral interface 1003.
According to the terminal provided in the implementations, the power consumption of the target area can be selectively reduced according to the input operation of the user, and so it is possible to save system power consumption of the terminal.
The foregoing device for controlling a terminal and the foregoing terminal may execute methods for controlling a terminal provided by any of the implementations of the present disclosure, and have function modules and advantageous effects for executing the methods. For technical details that are not described in detail in the above implementations, reference may be made to the method for controlling a terminal of any of the above implementations.
It should be understood that the implementation process of the foregoing components will not be detailed herein and reference may be made to the foregoing method implementations.
A non-transitory computer readable storage medium is provided. The non-transitory computer readable storage medium is configured to store a computer program which, when executed by a processor, causes the processor to carry out following actions. A target area is determined according to an input operation on a screen. A scheme for reducing power consumption of the target area is determined according to a layer allocation strategy of a currently running application, where the scheme is selected from a group consisted of reducing an image resolution in the target area and reducing power consumption of layers corresponding to the target area. Power consumption of the target area is reduced with the scheme determined.
In an implementation, the input operation includes a touch input operation or an eye-focus input operation, for determining the target area according to the input operation on the screen, the computer program is executed by the processor to carry out following actions. Coordinates of a center position of the input operation on the screen are determined. An operation area having a coordinate range containing the coordinates of the center position is determined. The target area is determined to be an area other than the operation area.
In an implementation, the input operation is a touch input operation, for determining the target area according to the input operation on the screen, the computer program is executed by the processor to carry out following actions. A closed area formed by an operation trajectory of the touch input operation on the screen is determined. determining the target area according to the closed area.
In an implementation, for reducing the power consumption of the layers corresponding to the target area, the computer program is executed by the processor to carry out at least one of: reducing a rendering frame rate of each of the layers corresponding to the target area; reducing a resolution of each of the layers corresponding to the target area; removing the layers corresponding to the target area from a set of layers to-be-composited; decreasing play volume of each of the layers corresponding to the target area. The above four types of power reduction processing manners may be combined with each other, which is not limited herein.
In an implementation, for reducing the image resolution in the target area, the computer program is executed by the processor to carry out following actions. A coordinate range of the target area in a display image is sent to a layer composite module. According to the coordinate range of the target area, respective sub-coordinate ranges of the target area in a plurality of layers are calculated via the layer composite module. Each of the sub-coordinate ranges is sent to a corresponding application via the layer composite module, whereby respective applications reduce the image resolution in the target area according to respective sub-coordinate ranges when rendering the plurality of layers corresponding to the target area.
In an implementation, for reducing the image resolution in the target area, the computer program is executed by the processor to carry out following actions. A coordinate range of the target area in a display image is sent to a layer composite module. According to the coordinate range of the target area, the image resolution in the target area is reduced via the layer composite module in a layer composite process.
It should be understood that the implementation process will not be detailed herein and reference may be made to the foregoing method implementations.
According to the method for controlling a terminal of the present disclosure, the power consumption of the target area can be selectively reduced according to the input operation of the user, thereby reducing system power consumption of the terminal.
Number | Date | Country | Kind |
---|---|---|---|
201710142958.8 | Mar 2017 | CN | national |
This application is a continuation of International Application No. PCT/CN2018/078565, filed on Mar. 9, 2018, which claims priority to Chinese Patent Application No. 201710142958.8, filed on Mar. 10, 2017, the disclosure of both of which are incorporated herein by reference in their entireties.
Number | Date | Country | |
---|---|---|---|
Parent | PCT/CN2018/078565 | Mar 2018 | US |
Child | 16565996 | US |