1. Field
The disclosed embodiments generally relate to user interfaces and, more particularly, to classifying and presenting multimedia data.
2. Brief Description of Related Developments
Management of different media types in a device can be done in various ways such as by file extensions or types and dates. Adding metadata to the media files improves the ability to search and find files, but generally metadata relies on textual information added to the media file. Adding very descriptive metadata is a time consuming and tedious task that is more often than not postponed by a user of the device leaving a search function of the device to operate with automatically added metadata (e.g. dates, file size and file type). Metadata searches generally work best for searching one media type at a time and do not provide for linking and associating different types of media items.
It would be advantageous to be able to establish links and associations between different types of items and present those different types of items based on the established links and associations.
In one aspect, the disclosed embodiments are directed to a method. In one embodiment the method includes providing different types of data in a device, automatically extracting data features from the data for comparison and automatically presenting the data on a display of the device where a multidimensional spatial relationship between the data on the display depends on a strength of similarities between the data features.
In another aspect, the disclosed embodiments are directed to an apparatus. In one embodiment the apparatus includes a processor and a display connected to the processor wherein the processor is configured to access different types of data associated with the apparatus, extract data features from the data for comparison and present the data on the display where a multidimensional spatial relationship between the data on the display depends on a strength of similarities between the data features.
In another aspect, the disclosed embodiments are directed to a user interface. The user interface includes an input device, a display and a processor connected to the input device and display, the processor being configured to access different types of data associated with the apparatus, extract data features from the data for comparison and present the data on the display where a multidimensional spatial relationship between the data on the display depends on a strength of similarities between the data features.
The foregoing aspects and other features of the embodiments are explained in the following description, taken in connection with the accompanying drawings, wherein:
The disclosed embodiments generally allow a user of a device 101 to re-live and explore connections and links between different items or data accessible by or stored in the device 101 where the connections and links may or may not be known to the user. The data can be any suitable data including, but not limited to, bookmarks, global positioning information, playlists, instant messaging presence, programs, shortcuts, help features, images, videos, audio, text, message files or any other items that originate from the device's operating system and/or applications or from a remote location. Generally the disclosed embodiments classify the data based on a number of different criteria including, but not limited to, metadata and other qualities of the data (e.g. all available information pertaining to each file or item can be extracted and used) as will be described in greater detail below.
The characterized data are grouped together or sorted and presented to a user of the device 101 through a display 114 of the device 101. The sorted data are presented on the display 114 in the form of a map, grid or other visual representation of the files where the data include one or more types of data as described above. The manner in which the data are grouped may be unexpected to the user so that browsing or exploring the items is fun to the user. It is also noted that relationships are built between the data so that, for example, photos taken and songs listened to during an event will be presented to the user as a group.
Referring also to
Each of the features F in the common features Fc is considered as a vector such that, for example, two or more feature vectors that belong to a set of common features Fc (equation [1])
{right arrow over (F)}1,{right arrow over (F)}2εFc [1]
can be weighted with a weight vector (equation [2]) which allows a user to influence how the media files of the media file types T1-T3 are grouped and presented to the user.
{right arrow over (w)} [2]
These feature vectors can be compared for similarity using any suitable distance metrics d such that
d:Fc×Fc→ [3]
where d is the distance and R is the vector space (which can have any suitable number of dimensions to account for the different features of the media items). The disclosed embodiments will be described with reference to Euclidean distance metrics that can be defined as
d
E(x,y)√{square root over ((ε1−η1)2+(ε2−η2)2+ . . . +(εn−ηn)2)}{square root over ((ε1−η1)2+(ε2−η2)2+ . . . +(εn−ηn)2)}{square root over ((ε1−η1)2+(ε2−η2)2+ . . . +(εn−ηn)2)} [4]
where εn and ηn are points in the vector space. It is noted that in other examples other metrics including, but not limited to, direction cosines, Minkowski metric, Tanimoto similarity and Hamming distance can be used.
The methods for measuring the distances between the features F (i.e. the feature vectors) of the media file types T1-T3 are used to classify and visualize the media file types T1-T3 and their features F1-F3. For example, the classification of the media item features F belonging to the common set of features Fc can be mapped to a discrete set of classes using a classifier algorithm Mc as shown in equation [5].
Mc:F→C [5]
Where C is a set of classes (e.g. class space) used in the classification and the features F can be weighted by the weighting vector. The classifier algorithm Mc can be any suitable classifier algorithm including, but not limited to neural networks, learning vector quantization, thresholding and different statistical methods.
The visualization of the media item features F belonging to the common set of features Fc can be mapped to a discrete set of classes using a mapping function, such as visualizer/mapping function Mv as shown in equation [6].
Mv:F→n [6]
Where Rn is a vector space having n dimensions and the features F can be weighted by the weighting vector. The visualizer/mapping function Mv can be any suitable visualizer/mapping function.
The connections and links formed between media files of the media file types T1-T3 through mapping the features to the discrete classes as described above with respect to equations [3]-[6] are used to visually present the media files on the display 114 in dependence on those connections and links as will be described in greater detail below. The media files of the media file types T1-T3 can be presented in any suitable number of dimensions on the display 114 such as, for example, in a two dimensional view or a three dimensional view. The relationships between the media files can be represented on the display 114 as a distance between the media items. For example, items that are connected or related to each other through one or more of the media item features Fi are located close to each other and/or placed in groupings on the display while items that are not connected to each other are spaced apart. In another example, media items that share features may appear larger in size than items that do not share features. In still other examples, the items can be arranged on the display in any suitable manner to indicate to the user that the items are related or not related.
It is noted that the above equations for grouping the items and their respective features are provided for exemplary purposes only and that any suitable equations, methods and functions can be used to group and present the items in the manner described below.
In one embodiment, still referring to
In one embodiment, the user interface 102 of the disclosed embodiments can be implemented on or in a device that includes a touch screen display or a proximity screen device 112. In alternate embodiments, the aspects of the user interface disclosed herein could be embodied on any suitable device that will display information and allow the selection and activation of applications or system content. The terms “select” and “touch” are generally described herein with respect to a touch screen-display. However, in alternate embodiments, the terms are intended to encompass the required user action with respect to other input devices. For example, with respect to a proximity screen device, it is not necessary for the user to make direct contact in order to select an object or other information. Thus, the above noted terms are intended to encompass that a user only needs to be within the proximity of the device to carry out the desired function. For example, the term “touch” in the context of a proximity screen device, does not necessarily require direct contact, but can include near or close contact, that activates the proximity device.
Similarly, the scope of the intended devices is not limited to single touch or contact devices. Multi-touch devices, where contact by one or more fingers or other pointing devices can navigate on and about the screen are also intended to be encompassed by the disclosed embodiments. Non-touch devices are also intended to be encompassed by the disclosed embodiments. Non-touch devices include, but are not limited to, devices without touch or proximity screens, where navigation on the display is performed through, for example, keys 110 of the device 101 or through voice commands via voice recognition features of the device 101.
In one embodiment, the user interface 102 includes a menu system 124. The menu system 124 can provide for the selection of different tools, settings and application options related to the applications or programs running on the device 101. In one embodiment, the menu system 124 may provide for the selection of applications or features associated with the presentation of media items such as, for example, any suitable setting features including, but not limited to, the settable features described herein. In one embodiment, the menu system 124 provides a way for the user of the device 101 to configure how the media file features Fi are grouped and compared against one another. The media file features Fi or grouping parameters can be set in any suitable manner. In one embodiment the menu system 124 can provide a way for the user to adjust any suitable number of parameters for grouping the media items. The menu system 124 can include any suitable text or graphics based menu or features that can be manipulated by the keys 110, touch screen 112 and/or microphone (e.g. through voice commands) of the input device 104. In another embodiment the menu system 124 can be configured to allow a user to configure the device 101 so that the grouping and visualization of the media items can be performed with great specificity. For example, the user can, through the menu system 124, specify any or all of the parameters or media item features that are used when grouping the media items.
The user may also be able to assign a weighting factor to groups of media item features and/or to each individual media item feature through the menu system 124. Assigning a weight to one or more media item features allows, for example, media item features with a heavier weight to influence the grouping of the media item more than media item features with a lesser weight. In one embodiment, it is noted that entering specific parameters for each individual media item feature will aid the user in quickly finding a media item. In other embodiments, one or more of the media item features can be hidden from the user so that the user has a generalized control over which parameters are used in grouping the media items. In another embodiment, the grouping parameters can be separated into different categories where the weighting within the different categories can be manipulated to provide some control over how the media items are grouped together. In still other embodiments the device 101 can be configured so that the grouping parameters are set by the device 101. For example, the grouping parameters can be set during manufacturing of the device, sets of parameters can be downloaded and/or installed into the device or the device can randomly select the grouping parameters. Having limited control over the grouping parameters could provide a source of entertainment to a user and group the media items in unexpected and surprising ways that the user may not think of. It is noted that the embodiments described herein will be described with respect to graphics based control over the grouping parameters for exemplary purposes only.
Referring now to
In this example, the media items or files can be gathered from the peripheral device 190, the network 191 and/or the storage 180 of
In the examples described herein the multidimensional relationships between the media items have a two or three dimensional representation but in other embodiments any suitable number of dimensions can be used. The spatial item dataset 340 is mapped to a spatial coordinate system 390 of the display 114 so that the media items 310 are presented on the display 114 as groups according to the relationships established by the self organizing map engine 350. The mapping of the spatial item dataset 340 can be done in any suitable manner such as, for example, with an artificial neural network algorithm of the self organizing map engine 350 that can learn the interdependencies between the media items 310 in an unsupervised way. The self organizing map engine 350 can group or place each media item 310 into the most suitable cell of the neural network based on the feature vectors. Each cell's location in the neural network can be a crude spatial location that is later refined using a local gradient of the self organizing map engine 350 and some randomness. In one embodiment, it is noted that the presentation of the media items 310 is “fuzzy” or unclear in that the cells do not describe an exact relationship between the grouped media items 310. In other embodiments, the device can be configured to provide exact relationships between the grouped media items 310.
After the spatial location of the media items 310 on the display is determined any suitable indicators of the media content are created and placed in the spatial coordinate system and projected on the display in the explorer view 380 using any suitable rendering features of the device 101. In this example, as can be seen in
It is noted that the difference between media item features determines a media item's position relative to other media items on the display 114. In one embodiment, when creating the feature vectors described above and comparing media items having, for example, the different file types described above any suitable combinations of the media item features can be used. For exemplary purposes only, one combination of features that can be used to form feature vectors for comparing the different media types can include a date, usage count and recording date of music so that files of the different data types having at least these features in common (or at least having similar features) are grouped via the comparison. Another exemplary combination can include tags, keywords, file name, title, metadata and words in a text file. Still another exemplary combination of features can include lightness/darkness of an image/video, slow/fast tempo of music, genre of music and length of words in text. One example of a media file grouping created from comparison of the different combinations of media item features is that music files having similar tempos can be closely grouped together. In another example, bright images/video and text having short words can be associated and grouped with music files having a fast tempo while text having long words and dark images/video can be grouped with music having a slow tempo. The larger the difference between the media item features of one media item with respect to the media item features of another media item, the further apart the media items will be on the display. As can be seen in
In one embodiment, informational data can be presented along with each grouping of items. This informational data can indicate, for example, features that the items within the group share with each other. For example, information 480 shown in
The position of and relationship between each of the media items 310 in the spatial coordinate system can be dynamically changed in any suitable manner. In one embodiment the position and relationship of the media items 310 can be changed by manipulating, for example, the weighting factors applied to the feature data 320. These weighting factors can be manipulated in any suitable manner such as through the menu system 124 described above. In this example, manipulation of the weighting factors will be described with respect to weighting sliders 490-493, which may be part of the menu system 124. Here each of the weighting sliders 490-493 can be associated with any suitable number and/or types of feature data 320. In one embodiment the sliders 490-493 may hide specific weighting parameters from the user and provide a way to generally modify the weighting parameters for grouping the media items 310. For exemplary purposes only slider 490 may be associated with text related feature data, slider 491 can be associated with time and locational feature data and slider 492 can be associated with music tempo, image and/or video brightness and word length feature data and slider 493 can be associated with a size or length of the media items. In other embodiments, there may be a slider for each specific weighting parameter to allow specific searches to be performed.
As can be seen in
In one embodiment, the device 101 can be configured so that the media items 310 can be manually moved from one group to another group in any suitable manner such as, for example, drag and drop, cut an paste, etc. For example, media item 310A can be removed from group 502 and placed in group 501. The device can be configured to track the manual placement of the media items within the different groups 501-503 and apply this information to the neural network so that the device “learns” how to arrange the media items according to, for example, a user preference or relationships known by the user but not previously defined within the device 101. These learned relationships can be applied to other media items to refine the grouping of the media items. The manual placement of the media items can also cause the device 101 to copy a corresponding metadata to the manually placed media items. For example, if one item is moved to a group having metadata related to a certain location, the metadata pertaining to the location will be copied or otherwise added to the moved item.
Still referring to
Navigating through the media items 310 in the explorer view 380 can also be done in any suitable manner. In one embodiment, the device can be configured so that the media items are translated in the X-Y plane of the display 114 by dragging a pointing device across a touch screen. In other embodiments, navigational controls 371 can be provided in the explorer view for translating the media items in the X-Y plane. A zoom feature can also be provided in any suitable manner such as through, for example, navigation controls 371 to allow a user to zoom media items in or out. In another embodiment, there may be a “fit to screen” feature 373 that is configured to fit all the media items on the display 114 without having to adjust the zoom feature. In still other embodiments, the device can be configured for navigating the explorer view and/or switching between two and three dimensional views through any suitable combination of the device's 101 input features 110, 111, 112.
As noted above the device 101 can be coupled to peripheral devices 190 and one or more networks 191. The explorer view 380 can be configured to allow for file transfers between the device 101, peripheral devices 190 and networks 191 for any suitable reasons including, but not limited to, file sharing, backups, synchronization or otherwise managing the files. For example, referring to
Examples of devices on which aspects of the disclosed embodiments can be practiced are illustrated with respect to
In the embodiment where the device 800 comprises a mobile communications device, the device can be adapted for communication in a telecommunication system, such as that shown in
The mobile terminals 900, 906 may be connected to a mobile telecommunications network 910 through radio frequency (RF) links 902, 908 via base stations 904, 909. The mobile telecommunications network 910 may be in compliance with any commercially available mobile telecommunications standard such as for example global system for mobile communications (GSM), universal mobile telecommunication system (UMTS), digital advanced mobile phone service (D-AMPS), code division multiple access 2000 (CDMA2000), wideband code division multiple access (WCDMA), wireless local area network (WLAN), freedom of mobile multimedia access (FOMA) and time division-synchronous code division multiple access (TD-SCDMA).
The mobile telecommunications network 910 may be operatively connected to a wide area network 920, which may be the Internet or a part thereof. An Internet server 922 has data storage 924 and is connected to the wide area network 920, as is an Internet client computer 926. The server 922 may host a worldwide web/wireless application protocol server capable of serving worldwide web/wireless application protocol content to the mobile terminal 900.
A public switched telephone network (PSTN) 930 may be connected to the mobile telecommunications network 910 in a familiar manner. Various telephone terminals, including the stationary telephone 932, may be connected to the public switched telephone network 930.
The mobile terminal 900 is also capable of communicating locally via a local link 901 to one or more local devices 903. The local link 901 may be any suitable type of link with a limited range, such as for example Bluetooth, a Universal Serial Bus (USB) link, a wireless Universal Serial Bus (WUSB) link, an IEEE 802.11 wireless local area network (WLAN) link, an RS-232 serial link, etc. The above examples are not intended to be limiting, and any suitable type of link may be utilized. In one embodiment the local devices 903 may be antennas and supporting equipment forming a wireless local area network implementing Worldwide Interoperability for Microwave Access (WiMAX, IEEE 802.16), WiFi (IEEE 802.11x) or other communication protocols. In other embodiments the local devices 903 can include the device 101 as described above. The wireless local area network may be connected to the Internet. The mobile terminal 900 may thus have multi-radio capability for connecting wirelessly using mobile communications network 910, wireless local area network or both. Communication with the mobile telecommunications network 910 may also be implemented using WiFi, Worldwide Interoperability for Microwave Access, or any other suitable protocols, and such communication may utilize unlicensed portions of the radio spectrum (e.g. unlicensed mobile access (UMA)). In one embodiment, the device 101 of
Although the above embodiments are described as being implemented on and with a mobile communication device, it will be understood that the disclosed embodiments can be practiced on any suitable device incorporating a display, processor, memory and supporting software or hardware. In one embodiment, the device 101 of
It is noted that in other embodiments the features described herein can be modified in any suitable manner to accommodate different display sizes and processing power of the device in which the disclosed embodiments are implemented. For example, in one embodiment, when the disclosed embodiments are implemented on devices with smaller displays, one or more toolbars and/or areas auxiliary to the explorer view may be omitted from the display. In another embodiment where the implementing device has limited processing power, the number of media item features used to sort and group the media items may be limited. In other embodiments, media items can be presented as frames (as opposed to thumbnails) with or without any generic content or text describing the items and/or metadata. In still other embodiments, any suitable indication of the media items can be presented on the display in any suitable manner when the capabilities of the implementing device are limited in some way.
The disclosed embodiments may also include software and computer programs incorporating the process steps and instructions described above that are executed in different computers.
Computer systems 1002 and 1004 may also include a microprocessor for executing stored programs. Computer 1004 may include a data storage device 1008 on its program storage device for the storage of information and data. The computer program or software incorporating the processes and method steps incorporating aspects of the invention may be stored in one or more computers 1002 and 1004 on an otherwise conventional program storage device. In one embodiment, computers 1002 and 1004 may include a user interface 1010, and a display interface 1012 from which aspects of the invention can be accessed. The user interface 1010 and the display interface 1012 can be adapted to allow the input of queries and commands to the system, as well as present the results of the commands and queries.
The embodiments described herein unify the management of different media types and provide new ways to explore media content stored in or accessed by a device. The disclosed embodiments provide for grouping different types of media items together in ways that a user of the device may not envision to provide the user with a fun and entertaining experience. The disclosed embodiments provide an easy way to browse and discovers content among, for example, a large collection of media content by building relationships between similar and/or different types of media items. The disclosed embodiments provide a way, through the relationships between media items, to re-discover media item content that may have been forgotten by a user of the device. In one aspect the disclosed embodiments also provide for a way to search for a specific media item or group of media items. The content discovery of the disclosed embodiments can function with or without metadata associated with the media files as features can be extracted from the media files themselves to build the relationships needed to group and present the media items.
It is noted that the embodiments described herein can be used individually or in any combination thereof. It should be understood that the foregoing description is only illustrative of the embodiments. Various alternatives and modifications can be devised by those skilled in the art without departing from the embodiments. Accordingly, the present embodiments are intended to embrace all such alternatives, modifications and variances.
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/IB08/03242 | 11/26/2008 | WO | 00 | 9/14/2010 |
Number | Date | Country | |
---|---|---|---|
60991366 | Nov 2007 | US |