This application is based on and claims priority under 35 U.S.C. § 119(a) to Indian patent application number 201841009451, filed on Mar. 15, 2018, in the Indian Intellectual Property Office, the disclosure of which is incorporated by reference herein in its entirety.
The disclosure relates to electronic devices. More particularly, the disclosure relates to a method and electronic device for enabling contextual interaction.
In general, electronic devices are ubiquitous in all aspects of modern life. Over a period of time, the manner in which the electronic devices display information on a user interface has become intelligent, efficient, spontaneous, and less obtrusive. The users interact on the user interfaces to navigate and to direct functionality to the electronic device. However, the user interface of the electronic device is mostly static i.e., the user interfaces are not customized based on any parameter such as context, conditions, etc. of the user and display a pre-defined set of applications. Further, the static user interfaces might cause inconvenience to the user in accessing the electronic device due to the increased number of steps involved to access a feature in the electronic device.
In an example, consider that the user is driving. The user interface (UI) of the electronic device is a home screen containing date, time and the applications that the user has selected to be displayed on the home screen, which are all static. When the user wants to play some preferred music, the user will have to navigate through the electronic device to access a music application to play the preferred music. Further, if the user wants to switch to a radio player, then the user will have to repeat the above mentioned steps. Furthermore, if the user wants to make a payment at a toll booth then the user will again have to browse through the applications to find a payment application and make the payment.
Further, the user may have to manually change settings of the electronic device to change the applications appearing on the user interface of the electronic device, which are both inconvenient and time-consuming.
The above information is presented as background information only to help the reader to understand the disclosure. Applicants have made no determination and make no assertion as to whether any of the above might be applicable as prior art with regard to the disclosure.
Aspects of the disclosure are to address at least the above-mentioned problems and/or disadvantages and to provide at least the advantages described below. Accordingly, an aspect of the disclosure is to provide a method and device for enabling contextual interaction.
Another aspect of the disclosure is to automatically determine a context of a user and to display functions based on the context on the screen of the electronic device.
Another aspect of the disclosure is to identify one or more functions from a pre-defined set of functions based on the detected context.
Another aspect of the disclosure is to identify one or more functions from a pre-defined set of functions and present the functions for user interaction.
Another aspect of the disclosure is to display the one or more functions distinctively based on the detected context for user interaction.
Another aspect of the present disclosure is to provide a method to determine the context of the user based on at least one of digital context associated with the user, physical context associated with the user, or user persona including usage pattern and a behavioral pattern of the user.
Additional aspects will be set forth in part in the description which follows and, in part, will be apparent from the description, or may be learned by practice of the presented embodiments.
In accordance with an aspect of the disclosure, a method for enabling contextual interaction on an electronic device is provided. The method includes detecting a context indicative of user activities associated with the electronic device and identifying one or more functions from a pre-defined set of functions based on the detected context. Further, the method also includes causing to display the one or more functions, where the one or more functions are capable of executing at least one of applications or services for accessing content relevant to the context, and dynamically performing an action relevant to the context in response to an interaction with a function.
In accordance with another aspect of the disclosure, an electronic device for enabling contextual interaction is provided. The electronic device includes a memory, a processor, a context detection engine, a function identification module and an output component. The context detection engine is configured to detect a context indicative of user activities associated with the electronic device. The function identification module is configured to identify one or more functions from a pre-defined set of functions based on the detected context. The output component is configured to cause to display the one or more functions, wherein the one or more functions are capable of executing at least one of applications or services for accessing content relevant to the context and dynamically perform an action relevant to the context in response to an interaction with a function.
Accordingly, an aspect of the disclosure is to provide a method for enabling interaction on an electronic device. The method includes identifying one or more functions from a pre-defined set of functions in the electronic device and causing to display the one or more functions, where the one or more functions are capable of executing at least one of applications or services for accessing content. Further, the method also includes dynamically performing an action in response to an interaction with a function.
Accordingly, an aspect of the disclosure is to provide an electronic device for enabling interaction. The electronic device includes a memory, a processor, a function identification module and an output component. The function identification module is configured to identify one or more functions from a pre-defined set of functions in the electronic device. The output component is configured to cause to display the one or more functions, wherein the one or more functions are capable of executing at least one of applications or services for accessing content and dynamically perform an action in response to an interaction with a function.
These and other aspects of the disclosure will be better appreciated and understood when considered in conjunction with the following description and the accompanying drawings. It should be understood, however, that the following descriptions, while indicating various embodiments and numerous specific details thereof, are given by way of illustration and not of limitation. Many changes and modifications may be made within the scope of the disclosure without departing from the spirit thereof, and the embodiments herein include all such modifications.
Other aspects, advantages, and salient features of the disclosure will become apparent to those skilled in the art from the following detailed description, which, taken in conjunction with the annexed drawings, discloses various embodiments of the disclosure.
The above and other aspects, features, and advantages of certain embodiments of the disclosure will be more apparent from the following description taken in conjunction with the accompanying drawings, in which:
Throughout the drawings, it should be noted that like reference numbers are used to depict the same or similar elements, features, and structures.
The following description with reference to the accompanying drawings is provided to assist in a comprehensive understanding of various embodiments of the disclosure as defined by the claims and their equivalents. It includes various specific details to assist in that understanding but these are to be regarded as merely exemplary. Accordingly, those of ordinary skill in the art that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the disclosure. In addition, descriptions of well-known functions and constructions are omitted for clarity and conciseness.
The terms and words used in the following description and claims are not limited to the bibliographical meanings, but, are merely used by the inventor to enable a clear and consistent understanding of the disclosure. Accordingly, it should be apparent to those skilled in the art that the following description of various embodiments of the disclosure is provided for illustration purpose only and not for the purpose of limiting the disclosure as defined by the appended claims and their equivalents.
It is to be understood that the singular forms “a,” “an,” and “the” include plural referents unless the context clearly dictates otherwise. Thus, for example, reference to “a component surface” includes reference to one or more of such surfaces.
Also, the various embodiments described herein are not necessarily mutually exclusive, as some embodiments can be combined with one or more other embodiments to form new embodiments.
Herein, the term “or” as used herein, refers to a non-exclusive or, unless otherwise indicated. The examples used herein are intended merely to facilitate an understanding of ways in which the embodiments herein can be practiced and to further enable those skilled in the art to practice the embodiments herein. Accordingly, the examples should not be construed as limiting the scope of the embodiments herein.
As is traditional in the field, embodiments may be described and illustrated in terms of blocks which carry out a described function or functions. These blocks, which may be referred to herein as units, engines, manager, modules or the like, are physically implemented by analog and/or digital circuits such as logic gates, integrated circuits, microprocessors, microcontrollers, memory circuits, passive electronic components, active electronic components, optical components, hardwired circuits and the like, and may optionally be driven by firmware and/or software. The circuits may, for example, be embodied in one or more semiconductor chips, or on substrate supports such as printed circuit boards and the like. The circuits constituting a block may be implemented by dedicated hardware, or by a processor (e.g., one or more programmed microprocessors and associated circuitry), or by a combination of dedicated hardware to perform some functions of the block and a processor to perform other functions of the block. Each block of the embodiments may be physically separated into two or more interacting and discrete blocks without departing from the scope of the disclosure. Likewise, the blocks of the embodiments may be physically combined into more complex blocks without departing from the scope of the disclosure.
Accordingly, the embodiments herein provide a method for enabling contextual interaction on an electronic device. The method includes detecting a context indicative of user activities associated with the electronic device and identifying one or more functions from a pre-defined set of functions based on the detected context. Further, the method also includes causing to display the one or more functions, where the one or more functions are capable of executing at least one of applications or services for accessing content relevant to the context; and dynamically performing an action relevant to the context in response to an interaction with a function.
In an embodiment, the one or functions are identified based on at least one of digital context associated with the user, physical context associated with the user, or user persona including usage pattern and a behavioral pattern of the user.
In an embodiment, each of the function comprises a plurality of relevant functions associated with the function.
In an embodiment, dynamically performing the action relevant to the context in response to the interaction with the function includes determining a plurality of relevant functions associated with the function. Further, the method also includes identifying a relevant function from the plurality of relevant functions using the detected context and performing the action based on the determined relevant function selected by the user.
In an embodiment, the plurality of relevant functions associated with the function is displayed along with the function for user interaction.
In an embodiment, the one or more functions are displayed distinctively based on the detected context for user interaction.
In an embodiment, the one or more functions are displayed automatically via a screen of the electronic device based on the detected context.
In an embodiment, the one or more functions are displayed via the screen of the electronic device for the detected context based on an input received from the user, wherein the input is one of a gesture input or a voice input.
In an embodiment, the one or more functions and the plurality of relevant functions for the one or more functions are displayed on a pre-defined portion via the screen of the electronic device.
Related-art methods and systems provide user interfaces (UIs) that are static i.e., the UI is not customized based on context of the user. In an example, the related-art UI of the electronic devices do not change based on whether the user is at home or driving.
Unlike related-art methods and systems, the proposed method allows the electronic device to determine the context of the user and provide the list of relevant functions based on the determined context, via a screen of the electronic device.
Unlike related-art methods and systems, the proposed method allows the electronic device to provide the relevant functions distinctively via a screen of the electronic device i.e., by highlighting the functions, enlarging the size of the relevant functions as compared to the other functions and the like.
Unlike related-art methods and systems, the proposed method links the notifications received from various applications to the context of the user and provide the list of relevant functions via a screen of the electronic device 100.
Related-art methods and systems are application based where the user has to follow a pre-defined path to access content on the electronic device which makes the process time consuming. Unlike related-art methods and systems, the proposed method is function based wherein the functions provide an easy access to a group of applications which are utilized for a specific purpose i.e., a function “play” may include all the applications like an audio player, a video player, radio, podcasts etc. combined into one function.
Referring now to the drawings, and more particularly to
In an embodiment, the electronic device 100 can be a mobile phone, a smart phone, personal digital assistants (PDAs), a tablet, a wearable device, a display device, an internet of things (IoT) device, electronic circuit, chipset, and electrical circuit (i.e., system on chip (SoC)), etc.
Referring to the
In an embodiment, the input component 110 can be configured to receive the input from the user on the screen of the electronic device 100. The input from the user can be one of gesture (e.g., touch, tap, drag, swipe, pressing of a dedicated button etc.), voice and the like. The set of functions associated with the context of the user is invoked by the user by providing the input on the UI of the electronic device 100. The UI can also be a voice input interface associated with a voice assistant application. The input component 110 can be hardware capable of receiving the user input. For example, the input component 110 can be a display or a microphone.
In an embodiment, the context detection engine 120 can be configured to determine the user context. The context indicates user activity and/or user intention. The context defines the activity of the user associated with the electronic device 100. The context can be one of a physical context or a digital context. The physical context can be determined based on one of the current location of the user (e.g., shopping mall, theatre, restaurant, etc.), time of the day (e.g., morning, afternoon, evening, night) and the activity performed by the user (for example walking, jogging, driving, sitting, etc.). The digital context can be determined based on one of notifications, ongoing task of the user, upcoming activities, status of connected device(s), browsing history (e.g., the user has searched for finance related sites, etc.), frequently used application, kind of profile used by the user (i.e., work profile, home profile) s and the like. The context detection engine 120 can be determine user context by using a plurality of sensors. For example, the plurality of sensors can be GPS sensor module, proximity sensor, acceleration sensor module, a gyro sensor, a gesture sensor, a grip sensor, color sensor or infrared sensor.
For example, when user is driving, the electronic device 100 automatically detects the context of the user based on the activity performed by the user i.e., driving and identifies the functions related to driving. Further, the electronic device 100 displays the functions and the associated sub-functions on the screen of the electronic device 100.
In an embodiment, the function identification module 130 can be configured to identify the functions based on the context of the user. The function can mean a function executable in at least one application installed in the electronic device 100. If the application installed in the electronic device 100 is modified, the function can also be modified. Initially, the function identification module 130 determines whether the function identification module 130 has access to the digital persona of the user. On determining that the function identification module 130 has access to the digital persona of the user, the function identification module 130 uses the digital persona of the user to determine the function based on the digital context of the user. The digital persona of the user is developed by the electronic device 100 based on a continuous learning of the user's behavior. Further, to access the digital persona of the user, the function repository 140 is assumed to be located outside the electronic device 100 (e.g., a cloud server) and accessed with wireless communication techniques through the communication module 150.
On determining that the function identification module 130 does not have access to the digital persona of the user (i.e., the function repository 140 is offline), the function identification module 130 uses only the physical context and the usage pattern analysis of the user to determine the functions. Further, the function identification module 130 filters the functions and prioritizes the functions based on the usage pattern analysis of the user.
In an embodiment, the function repository 140 can be configured to store the list of functions associated with the physical context and the digital context identified by the function identification module 130. Further, the function repository 140 also stores the digital persona of the user which is created based on learning the usage pattern of the user. Further, the function repository 140 can be embedded within the electronic device 100 and readily accessed. In another embodiment, the function repository 140 can be located outside the electronic device 100 (e.g., a cloud server) and accessed using wireless communication techniques through the communication module 150.
In an embodiment, the communication module 150 can be configured to communicate with the function repository 140. Further, the communication module 150 determines whether the function repository 140 is online and implements one or more suitable protocols for communication. The protocols for communication can be for example, Bluetooth, near field communication (NFC), ZigBee, RuBee, and wireless local area network (WLAN) functions, etc.
In an embodiment, the output component 160 can be configured to provide one or more functions on the screen of the electronic device 100. The functions are fetched and displayed automatically on the screen of the electronic device 100 based on the detected context. The one or more functions are capable of executing at least one of applications or services for accessing the content relevant to the context. Further, the output component 160 is configured to display the one or more functions distinctively on the screen of the electronic device 100 based on the detected context. For example, the functions which have received notifications from the associated applications are highlighted and presented on the screen of the electronic device 100. In various embodiments, the output component 160 can be implemented on the display.
Further, the output component 160 is configured to dynamically perform one or more actions relevant to the context in response to the interaction by the user, with a function i.e., the output component 160 initiates the action to be performed based on the function selected by the user. Furthermore, the one or more functions and the plurality of relevant functions associated with the one or more functions are displayed on a pre-defined portion on the screen of the electronic device 100. For example, the user invokes the list of functions by performing a gesture on the bottom portion of the screen of the electronic device 100 and the list of functions are displayed in the bottom portion of the screen of the electronic device 100 (as illustrated in
In an embodiment, the processor 170 can be configured to interact with the hardware elements such as the input component 110, the context detection engine 120, the function identification module 130, the function repository 140, the communication module 150, the output component 160 and the memory 180 for providing the UI of the electronic device 100.
In an embodiment, the memory 180 may include non-volatile storage elements. Examples of such non-volatile storage elements may include magnetic hard discs, optical discs, floppy discs, flash memories, or forms of electrically programmable memories (EPROM) or electrically erasable and programmable (EEPROM) memories. In addition, the memory 180 may, in some examples, be considered a non-transitory storage medium. The term “non-transitory” may indicate that the storage medium is not embodied in a carrier wave or a propagated signal. However, the term “non-transitory” should not be interpreted that the memory 180 is non-movable. In some examples, the memory 180 can be configured to store larger amounts of information than the memory. In certain examples, a non-transitory storage medium may store data that can, over time, change (e.g., in random access memory (RAM) or cache).
Although the
Referring to
In various embodiments, the context detection engine 120 can identify the user context or context of the electronic device 100. For example, the context detection engine 120 can identify the user context including the location of the user by using the GPS sensor. The context detection engine 120 can identify the user context including the velocity of the user by using the acceleration sensor. The context detection engine 120 can identify the context of the electronic device 100. The context of the electronic device 100 can include a capability, performance, or status of the electronic device 100.
In an example, consider that the user is at home. Based on the location of the user (i.e., the physical context of the user), the context detection engine 120 determines the context as ‘home’. The function identification module 130 identifies the relevant functions associated with ‘home’ based on the digital persona of the user, the context of the user and the usage pattern analysis of the user. Further, the relevant functions are populated on the screen of the electronic device 100. One of the relevant functions can be ‘connect function’. The sub-functions associated with the connect function can be contacts, text messaging applications, instant messaging applications and the like. The ‘connect function’ enable the user to automatically send a text message to a frequently messaged contact etc. Other relevant functions are ‘utility function’ which would enable the user to access reminders, the ‘connectivity IoT function’ which would enable the user to control the connected devices, the ‘share function’ which would enable the user to share multimedia or location data with the contacts, the ‘play function’ which would enable the user to play multimedia content such as videos, music etc.
In various embodiments, the context detection engine 120 can identify the context of user by using the plurality of sensors. In response to identifying the context of user, the context detection engine 120 can transmit the identified context to the function identification module 130.
In various embodiments, the related functions can be grouped according to index of the context. The processor 170 can identify the application installed in the electronic device 100 and the function being provided by the application every predetermined period. The processor 170 can group the identified functions based on the context, and store the grouped functions in the function repository 140. The information of the related function can be stored in the function repository 140.
In various embodiments, the function identification module 130 can identify the related functions corresponding to the identified context. For example, the function identification module 130 can request the function repository 140 to transmit the related functions corresponding to the identified functions. The function identification module 130 can control the output component 160 to provide at least one of the related functions on the display of the electronic device 100.
In various embodiments, the output component 160 can receive information of the related functions from the function identification module. The output component 160 can configure a plurality of GUI (for example, icon) and display the plurality of GUI on the display of the electronic device 100. Each of the plurality of GUI corresponds to each of the related function.
In various embodiments, the processor 170 can receive the user input for requesting execution of the related function and display at least one icon corresponding to application providing the related function. The processor 170 can recommend or suggest the application based on the context information or a priority information of the application when the processor 170 identifies that a plurality of applications supporting the related function are installed in the electronic device. The processor 170 can identify at least one application among application installed in the electronic device 100 based on at least one of user activity, capability or status of the electronic device 100. For example, the processor 170 can identify application capable of executing a function corresponding to the selected object. The processor 170 can recommend or suggest the identified application. The processor 170 can execute the function in the selected application by executing the identified application.
Referring to
The notifications related to the functions are described in Table 2.
Referring to the
At operation 404, the electronic device 100 identifies one or more functions from a pre-defined set of functions based on the detected context. For example, in the electronic device 100 as illustrated in the
At operation 406 the electronic device 100 causes to display the one or more functions. For example, in the electronic device 100 as illustrated in the
At operation 408 the electronic device 100 dynamically performs the action relevant to the context in response to an interaction with a function. For example, in the electronic device 100 as illustrated in the
The various actions, acts, blocks, steps, or the like in the method may be performed in the order presented, in a different order or simultaneously. Further, in some embodiments, some of the actions, acts, blocks, steps, or the like may be omitted, added, modified, skipped, or the like without departing from the scope of the disclosure.
Referring to the
At operation 506, the electronic device 100 captures the context of the user. For example, in the electronic device 100 as illustrated in the
At operation 508, the electronic device 100 determines whether the function repository 140 is online i.e., whether the function repository 140 has access to the digital persona of the user. For example, in the electronic device 100 as illustrated in the
On determining that the function repository 140 does not has access to the digital persona of the user, at operation 512, the electronic device 100 identifies the relevant functions based on the physical context and the usage pattern analysis of the user. For example, in the electronic device 100 as illustrated in the
At operation 514, the electronic device 100 provides the functions after filtering and prioritizing the functions. For example, in the electronic device 100 as illustrated in the
At operation 516, the electronic device 100 dynamically performs the action relevant to context, in response to the function selected by the user. For example, in the electronic device 100 as illustrated in the
On determining that the function repository 140 is online i.e., the function repository 140 has access to the digital persona of the user, at operation 510, the electronic device 100 accesses the digital persona of user and identifies the relevant functions based on the digital context of user. For example, in the electronic device 100 as illustrated in the
The various actions, acts, blocks, steps, or the like in the method may be performed in the order presented, in a different order or simultaneously. Further, in some embodiments, some of the actions, acts, blocks, steps, or the like may be omitted, added, modified, skipped, or the like without departing from the scope of the disclosure.
Referring to
The functions associated with driving can be for example navigation applications, camera or video applications, payment applications, music applications, map applications, applications providing information pertaining to the surroundings, applications providing information pertaining to traffic, etc. Further, the user is allowed to select the required function from the set of functions 604 associated with driving which are presented on the home screen 602 of the electronic device 100.
The user performs a gesture 606 and selects the required function from the set of functions 604 i.e., the user selects the play music function. The electronic device 100 on receiving the gesture 606 electronic device 100 filters the sub-functions to be associated with the play music function based on the context of the user. Since the context of the user is determined to be driving, the electronic device 100 excludes video related sub-functions and associates only the audio related sub-functions to the play music function. Further, the electronic device 100 provides the relevant functions (i.e., sub-functions) associated with the play music function. The relevant functions associated with the play music function can be for example, play lists 608 such as happy songs play list, emotional songs play list, ambitious songs play list, travel songs play list etc., radio player, podcasts and the like. Further, the play list associated with driving is automatically selected from the set of play lists and played without the user having to select the play list, as shown in
Further, the user is also provided with the option of selecting a different playlist or music application from the home screen 602 (e.g., select to play radio) in case the user wants to play music from some other playlist or application. The user can select a different playlist or application with voice command, touch command, etc.
Referring to
The user selects the play music function from the set of functions 704 by performing the gesture 708 on the home screen 702 of the electronic device 100. On determining that the user has selected the play music function, the electronic device 100 automatically launches the music app2 and plays the received music file, as shown in
Further, the user is allowed to select any other sub-function other than the sub-function being played. Furthermore, if the user wants to access any other function apart from the highlighted functions, the electronic device 100 provides the set of functions 710 at the bottom of the screen of the electronic device 100, as shown in
Referring to
Consider a scenario where the user of the electronic device 100 is dining at a restaurant. The user performs the gesture 804 to invoke the functions on the application menu screen 802. The electronic device 100 determines the context of the user based on the user location (i.e., restaurant) as dining. Further, the electronic device 100 provides the set of functions 806 associated with dining on the application menu screen 802 of the electronic device 100, as shown in
As the user selects the payment function from the set of functions 806, the electronic device 100 automatically initiates the payment using a pre-saved diner card 810, as shown in
Referring to 9A, 9B, and 9C, a scenario where the user checks the electronic device 100 in the early hours of the day i.e., morning. The user invokes the set of functions 906 associated with the context of the user on the existing UI of the contact screen 902 by performing a gesture 904, as shown in
Further, if the user regularly accesses the New York Times application in the morning, then the electronic device 100 learns the user behavior pattern and adds the New York Times application to the play function. The user can launch the New York Times application by selecting the play function.
Further, the set of functions 906 associated with the time of the day is provided on the screen of the electronic device 100 and the user can select any of the functions to access a different application.
Referring to
Consider a scenario where the user of the electronic device 100 accesses the virtual assistant application 1002. A panel of enabler functions 1004 is provided at the top portion of the UI of the virtual assistant application 1002 and the set of functions 1006 associated with the context of the user is provided at the bottom portion of the UI of the virtual assistant application 1002, as shown in
The user selects the communication enabler function from the panel of enabler functions 1004 provided at the top portion of the UI of the virtual assistant application 1002 by providing the voice command (indicated by the circle 1008), as shown in
Referring to the
Referring to
The user invokes the set of functions 1106 on the application menu screen of the electronic device 100 by performing a gesture 1104, as shown in
Furthermore, if the user frequently accesses the email-3 application to check for work related emails then the electronic device 100 will learn that the user frequently accesses the email-3 and provides higher priority to the email-3 application associated to the contact function, so that the email-3 application is automatically launched when the user selects the contact function.
Furthermore, the set of functions 1106 associated with the frequently used applications is provided on the screen of the electronic device 100 and the user can select any of the functions to perform a different function.
Various aspects of the disclosure can also be embodied as computer readable code on a non-transitory computer readable recording medium. A non-transitory computer readable recording medium is any data storage device that can store data which can be thereafter read by a computer system. Examples of the non-transitory computer readable recording medium include read-only memory (ROM), RAM, CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices. The non-transitory computer readable recording medium can also be distributed over network coupled computer systems so that the computer readable code is stored and executed in a distributed fashion. Also, functional programs, code, and code segments for accomplishing the disclosure can be easily construed by programmers skilled in the art to which the disclosure pertains.
At this point it should be noted that various embodiments of the disclosure as described above typically involve the processing of input data and the generation of output data to some extent. This input data processing and output data generation may be implemented in hardware or software in combination with hardware. For example, specific electronic components may be employed in a mobile device or similar or related circuitry for implementing the functions associated with the various embodiments of the disclosure as described above. Alternatively, one or more processors operating in accordance with stored instructions may implement the functions associated with the various embodiments of the disclosure as described above. If such is the case, it is within the scope of the disclosure that such instructions may be stored on one or more non-transitory processor readable mediums. Examples of the processor readable mediums include ROM, RAM, CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices. The processor readable mediums can also be distributed over network coupled computer systems so that the instructions are stored and executed in a distributed fashion. Also, functional computer programs, instructions, and instruction segments for accomplishing the disclosure can be easily construed by programmers skilled in the art to which the disclosure pertains. Also, the embodiments disclosed herein may be implemented using at least one software program running on at least one hardware device and performing network management functions to control the elements.
While the has been shown and described with reference to various embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the disclosure as defined by the appended claims and their equivalents.
Number | Date | Country | Kind |
---|---|---|---|
201841009451 | Mar 2018 | IN | national |