Digital jukebox device with improved karaoke-related user interfaces, and associated methods

Information

  • Patent Grant
  • 10782853
  • Patent Number
    10,782,853
  • Date Filed
    Wednesday, April 25, 2018
    6 years ago
  • Date Issued
    Tuesday, September 22, 2020
    4 years ago
Abstract
Certain exemplary embodiments relate to entertainment systems and, more particularly, certain exemplary embodiments relate to jukebox systems that incorporate digital downloading jukebox features along with karaoke jukebox and/or photobooth features. A combined karaoke/photobooth/jukebox may enable more integrated performance-like experiences in an in-home or out-of-home location or venue. By leveraging vast audio media libraries, trusted rights-respecting network infrastructure, and on-site image/video capturing from integrated recorders and/or remote portable devices, a more sociable experience may be created for karaoke jukebox patrons, e.g., where custom content can be generated and shared in a safe and legally appropriate manner.
Description
TECHNICAL FIELD

Certain exemplary embodiments relate to entertainment systems and, more particularly, certain exemplary embodiments relate to jukebox systems that incorporate digital downloading jukebox features along with karaoke jukebox and/or photobooth features.


BACKGROUND AND SUMMARY

Jukeboxes have been around for decades and provide users with the ability to select desired music for reproduction in a convenient and advantageous manner. Jukeboxes conventionally have been provided in commercial establishments, such as restaurants and bars, in order to provide desired music on demand for patrons thereof for a fee. Over the last several years, a new generation of jukebox devices have become available that provide significant improvements in the operation thereof for all parties involved. More specifically, the conventional standalone phonorecord and CD jukeboxes are being replaced by digital downloading jukeboxes that are controlled by and communicate with a central server. An example of this new generation jukebox system is shown in U.S. Pat. No. 6,308,204, the entire disclosure of which is incorporated herein by reference. A leading provider of this new generation of jukebox systems is TouchTunes Music Corporation.



FIG. 1 shows an overview of an exemplary embodiment of a digital downloading jukebox system 10. As shown in FIG. 1, the jukebox system 10 includes a central server 12 that contains a master library of audio content (typically music), as well as or alternatively audiovisual content (typically music and associated video or graphics), that can be downloaded therefrom. The jukebox system also includes a series of remote jukebox devices 16, 16a-16f. Each of these jukebox devices are generally located in a bar, restaurant, club, or other desired location, and are operable to play music (e.g., from a suitable storage location such as, for example, from a local server, a central and potentially remote server, from local storage, etc.) in response to receiving a payment from a user, such as coins, bills, credit/debit card, etc., and having one or more songs selected by the user for play. In an alternative embodiment, a music service is paid for on a subscription basis by the location, and the selected music is free for the end-user. The jukebox device 16 typically includes a screen 18 that presents information to the user and allows the user to select songs therefrom, as well as an audio system 20 that plays the selected songs. The screen 18 may also be used for displaying song-related video or graphics. The screen 18 may also be used to display advertisements for the jukebox itself in order to attract customers thereto, to display other types of advertisements, and/or to display any other desired information.


The jukebox devices 16 (sometimes referred to herein as simply “jukeboxes”) are operable to communicate with the central server 12 through a communications network 14, such as, for example, the Internet. The jukeboxes 16 periodically communicate with the server 12 in order to provide information to the server 12 regarding the specific songs that have been played on the jukebox. The central server then uses this information in order to determine the appropriate royalties and/or other payments that are owed for songs played on each jukebox. Thus, one advantage of this new generation of jukeboxes is that the sound reproduction and/or other applicable music rights can be adhered to in a more accurate and reliable manner, thereby assuring the proper royalties are paid to the artists or music owners. The central server 12 can also provide new songs to the jukebox 16 in order to assure that the appropriate or most popular songs are maintained on the jukebox based on the specific customers at that location. Thus, the songs available on each jukebox can be customized through communication with the central server in order to provide the songs and/or types of music that customers generally request at each jukebox location. As described in the above-referenced U.S. Pat. No. 6,308,204, the central server can also advantageously be used to update the operating software on the jukeboxes in order to, for example, change the operation of the jukebox, such as to provide new or improved features. Thus, another advantage of this new generation of jukeboxes is that the songs (or other audio and/or visual content), and the operation of the jukebox itself can be remotely changed as desired without the need to have someone (such as a routeman) personally service the jukebox. Instead, such updates can be done using the central server 12.


As indicated above, the jukebox devices 16 each include a mass storage device, such as a hard drive, which stores the songs and associated video/graphics data (if any), as well as any other desired graphical information for reproduction on the jukebox. The mass storage device of the jukebox typically has limited storage capacity relative to the storage device of the central server 12. As a result, only a fraction of the songs stored on the central server are typically stored on the mass storage device of the jukebox at any one time. There may be other reasons as well, such as for security of the data or limited room in the jukebox itself, for having limited storage capacity on the jukebox and/or limiting the number of songs stored thereon. For example, physical space may be limited on wall-mount jukeboxes or the like, which are designed to be small in size as compared to free-standing models. As explained above, the songs on the jukebox can be changed through communication with the central server, but typically any one jukebox only stores a relatively small subset of the complete library of songs maintained by the central server at any one time.


In order to increase the revenue that a jukebox generates, making the most desired or popular songs available on the jukebox over time may be seen as an advantage. If customers cannot find songs they like on the jukebox, usage of the jukebox (and the revenue generated thereby) can decrease. On the other hand, it is not always possible to predict in advance exactly what a customer at any particular location will desire to play on the jukebox. In fact, there are likely many instances where a customer would have selected a song that exists on the central server but is not currently present on the jukebox. As a result, the jukebox may not be enjoyed and used to its fullest extent. In order to address this problem and increase revenue, jukebox systems have in the past provided a feature that enables the user to search for songs on the central server from the jukebox and request an immediate download of a desired song from the central server to the jukebox for an additional fee. This feature enables the user to play any song in the master library of songs maintained by the central server using the jukebox, regardless of whether or not the specific song is presently stored in the mass storage of the jukebox itself. Thus, the user can first look for desired songs on the local storage of the jukebox and then, if desired, search further on the central server for desired songs (e.g., in connection with search functionality, potentially accessible by selecting a central server search function button on the screen). The jukebox device typically charges an additional fee (such as five credits instead on one credit) for an immediate download and play of a song from the central server as opposed to a standard play directly from the jukebox's local storage.


In addition to all the features of a conventional jukebox, there is a need for an integrated karaoke experience. In that regard, current jukeboxes typically are turned off or otherwise deactivated during karaoke nights, where karaoke providers bring in their own separate equipment and set it up at the venue, or the venue itself simply sets up the separate karaoke system in which it has separately invested. The deactivation of the jukebox device, even if only temporary, represents a loss of potential revenue during such karaoke establishments. The setup and tear-down of the separate karaoke systems also is inconvenient and sometimes burdensome, e.g., for short-staffed venues, for venues where there is not a lot of space, etc. Moreover, conventional jukeboxes have been largely shut out from “karaoke-only” establishments. Because jukeboxes and karaoke device belong in different classes in the real world, a whole market remains unavailable and thus untapped.


As is known, the karaoke environment involves an organized process of selecting music from an exclusive karaoke catalog that contains legally licensed music. Unfortunately, it has been very challenging to develop a sufficiently feature-rich solution to allow a true karaoke experience to take place in a bar or public venue. Karaoke systems in place today for commercial uses typically present music that the “KJ” or “Karaoke Jockey” (as opposed to the “DJ”) has accumulated. It has been observed that few KJs make any effort to obtain legally licensed versions of the songs that make use of. Even genuine efforts to license songs from appropriate karaoke music suppliers are fraught with risk, as the suppliers themselves oftentimes do not have coherent and effective licensing capability. And even genuine efforts to license songs still cannot provide to KJs the vast music catalog available on jukebox devices.


Thus, it will be appreciated that there is a need in the art for a new catalog management and royalty payment strategy that goes beyond what has been put in place for conventionally licensed music for mechanical reproduction and performance, e.g., so as to accommodate karaoke environments. For instance, certain exemplary embodiments provide a way to comply with obligations to respect author rights, in the more complicated karaoke environment where a different set of rights is implicated, e.g., in connection with vocal elimination and replacement, generation of new content, etc.


In this vein, as is known, the karaoke experience creates a new work, namely, a combination of the original music and the performer(s) singing. There oftentimes is a desire for performers to retain audio or audio and video copies of their performances. To meet this need, however, would imply the potentially recurring usage of the original music track and the appropriate tracking and management of these produced works, e.g., to comply with current copyright laws. Thus, another challenge facing the development of a karaoke application involves providing the ability to create these new performance works, while also managing them in a manner that respects the licensing demands of the original rights holder(s). Media management, however, is hampered currently, as digital media and ubiquitous player technology has become widespread.


One aspect of certain exemplary embodiments relates to providing, in the venue or location corresponding to the physical jukebox and karaoke environment, the ability to seamlessly transition between conventional jukebox usage or, jukebox plus broadcast TV (as in using the jukebox audio system to play the sound for a football game being presented on the venues televisions), and/or a new karaoke-appropriate audio and video configuration. In so doing, it may be desirable to make adjustments to a number of settings of the combined karaoke jukebox system.


For instance, the volume settings for all zones may be altered to fairly present the stage environment and provide the performer with suitable audio monitor capability. In addition, video screens may be repurposed to provide information about upcoming performers and lyrics for the current song. As alluded to above, it would be desirable to have this transition take place in a seamless manner, demanding very little skill from venue staff or the karaoke master of ceremonies or the de facto “KJ.”


Thus, it will be appreciated that there is a need in the art for techniques for improving jukeboxes by also incorporating karaoke features that provide karaoke performs a “take-away” while addressing licensing, reporting, and/or other requirements, both at and away from the venue and location, while also enabling for a seamless transition between jukebox and karaoke modes of operation.


The inventors of the instant application have recognized that karaoke can be a positive, social, experience that is fun and sometimes even competitive. Performance is a part of the karaoke experience, for both performers and the audience. The more people involved in the experience, the more fun it tends to be. Smart technology can provide new KJ features. Smart technology also can leverage opportunities to create a connected karaoke experience for a connected world.


Thus, an aspect of certain exemplary embodiments relates to encouraging more people to perform, and/or connecting more people to the performance—before, during, and/or after a performance.


An aspect of certain exemplary embodiments relates to providing a karaoke jukebox connected system with collaborative touch points (including, for example, user devices such as mobile phones, tablets, etc.; jukeboxes themselves; game or other fixed or portable terminals in a location; etc.) that define unique moments.


Another aspect of certain exemplary embodiments relates to treating a performance as a moment of the karaoke experience that other moments both lead to and stem from.


Another aspect of certain exemplary embodiments relates to defining moments in the experience by providing unique or signature interactions that help make the experience immersive and unique for patrons, whether they are performing or watching at the location or remote from the location. For instance, digital songbooks for karaoke-type events may replace paper printouts that feel clunky and in many cases are dated. Karaoke in general may be made less hokey and more unique and tailored to an out-of-home experience in some cases.


Certain exemplary embodiments relate to a jukebox device usable in a jukebox mode, and (a) a karaoke mode, and/or (b) a photobooth mode, as shown and described herein.


Certain exemplary embodiments relate to a method for operating a jukebox device usable in a jukebox mode, and (a) a karaoke mode, and/or (b) a photobooth mode, as shown and described herein.


Certain exemplary embodiments relate to a system including a jukebox device usable in a jukebox mode, and (a) a karaoke mode, and/or (b) a photobooth mode, as shown and described herein.


Certain exemplary embodiments relate to a user interface, or portion thereof, that is suitable for use during karaoke.


In certain exemplary embodiments, a jukebox device is provided and comprises processing resources including at least one processor and a memory, a touch screen display device, a payment acceptor, and a storage device comprising instances of media available for playback via the jukebox device. The processing resources, the touch screen display device, and the payment acceptor cooperate with one another to cause the jukebox device to function in at least first and second different modes. The first mode is a normal jukebox operation mode that is configured to enable a first partial subset of the instances of media to be played back on a pay-per-play basis. The second mode is a karaoke jukebox operation mode that is configured to enable a second partial subset of the instances of media to be played back on a pay-per-play basis, with the first and second partial subsets being different from one another. As a part of the second mode, the touch screen display device presents a first set of screens that each list at least a portion of the instances of media in the second partial subset of the instances of media in a manner that resembles a book having multiple pages. Each said page includes at most a predefined maximum number of entries corresponding to the portion of the second partial subset of instances of media, and each said entry indicates a song name and an artist associated with the corresponding instance of media that is selectable for playback via the jukebox device. The first set of screens each include a common set of control elements for re-organizing and navigating through the book.


According to certain exemplary embodiments, non-transitory computer readable storage mediums tangibly store programs that, when executed, implement these and/or other methods.


The exemplary embodiments, aspects, and advantages disclosed herein may be provided in any suitable combination or sub-combination to achieve yet further exemplary embodiments.





BRIEF DESCRIPTION OF THE DRAWINGS

These and other features, aspects, and advantages of the instant invention will be further understood by review of the following detailed description of the exemplary embodiments when read in conjunction with the appended drawings, in which:



FIG. 1 is a block diagram of a conventional downloading digital jukebox system;



FIG. 2 is a block diagram of an exemplary embodiment of an improved jukebox system 10′;



FIG. 3 is a schematic view showing flight content resolution in accordance with an exemplary embodiment;



FIG. 4 is a schematic view showing display group resolution in accordance with an exemplary embodiment;



FIG. 5 is a flight segment appearing in one anchor point and moving and zooming to appear closer to the user in accordance with an exemplary embodiment;



FIG. 6 illustrates browsing a music collection for a given artist in accordance with an exemplary embodiment;



FIG. 7 illustrates browsing a music collection by artist and album in accordance with an exemplary embodiment;



FIG. 8a illustrates a set of eight (8) example tags associated with “artist 6” in accordance with an exemplary embodiment;



FIG. 8b is a wireframe used to construct an example music map on a jukebox display in accordance with an exemplary embodiment;



FIG. 9 is a block diagram illustrating one arrangement that may enable custom browse state and other features to be implemented in accordance with an exemplary embodiment;



FIG. 10 shows a 3D visualization of a recommended song collection in accordance with an exemplary embodiment;



FIG. 11 is an illustrative screen showing how objects may be navigated in accordance with an exemplary embodiment;



FIG. 12 is an illustrative screen showing how an object may be expanded upon selection in accordance with an exemplary embodiment;



FIG. 13 is another illustrative screen showing how objects may be navigated in accordance with an exemplary embodiment;



FIG. 14 is an illustrative screen showing how a user may enter lyrics to reduce the number of songs being proposed for playback in accordance with an exemplary embodiment;



FIG. 15 is a schematic representation showing how songs may be selected for presentation to a user in accordance with an exemplary embodiment;



FIG. 16 is a flowchart illustrating an example process for determining which content should be available for playback on a jukebox based on location-specific inclusion criteria and actual jukebox usage in accordance with an exemplary embodiment;



FIG. 17 is a flowchart illustrating an example process for initially setting up a jukebox in accordance with an exemplary embodiment;



FIG. 18 is a flowchart illustrating an example process for customizing the display for a recognized user in accordance with an exemplary embodiment;



FIG. 19 is a flowchart illustrating an example process for customizing the display for a special event in accordance with an exemplary embodiment;



FIG. 20a is a first example “artist portal” in accordance with certain exemplary embodiments;



FIG. 20b is a second example “artist portal” in accordance with certain exemplary embodiments;



FIG. 21 is an example of a partial or incremental presentation of the jukebox's play queue in accordance with certain exemplary embodiments;



FIG. 22a uses a colorized “blob” to highlight the center selection, whereas other non-patron selectable elements may be blurred, in accordance with certain exemplary embodiments;



FIG. 22b shows the enlargement of an album jacket to highlight its selection, in accordance with certain exemplary embodiments;



FIG. 22c shows the enlargement of an album jacket to highlight its selection, and further blurs background UI elements to help bring focus to the selected album, in accordance with certain exemplary embodiments;



FIGS. 23a-d show example screenshots of a kinetic motion display in accordance with an exemplary embodiment;



FIG. 24 is an example screenshot of an improved music discovery user interface in accordance with certain exemplary embodiments;



FIGS. 25a-b show components of a jukebox device in accordance with certain exemplary embodiments;



FIG. 26 is an image of an example jukebox device in accordance with certain example embodiments;



FIG. 27 is an illustrative view of visual components on a jukebox device being color matched and synchronized in accordance with certain exemplary embodiments;



FIG. 28 is an example screenshot enabling a user to access a plurality of different features of a combined karaoke/photobooth jukebox in accordance with certain exemplary embodiments;



FIG. 29 is an example screenshot enabling a user to access karaoke jukebox features in accordance with certain exemplary embodiments;



FIG. 30 is an example artist selection screen for a karaoke jukebox in accordance with certain exemplary embodiments;



FIGS. 31-32 show example karaoke songbook page-turning operations in accordance with certain exemplary embodiments;



FIG. 33 is an example screenshot of a detailed karaoke song selection screen in accordance with certain exemplary embodiments;



FIG. 34 is an example screenshot including a display of karaoke song lyrics in accordance with certain exemplary embodiments;



FIG. 35 is an example karaoke user identification screen in accordance with certain exemplary embodiments;



FIG. 36 is an example karaoke song confirmation screen in accordance with certain exemplary embodiments;



FIGS. 37-38 are example karaoke attract mode screens that may be displayed in accordance with certain exemplary embodiments;



FIG. 39 is an example screenshot showing the display of lyrics for a performance in accordance with certain exemplary embodiments;



FIGS. 40-41 demonstrate how performers may be signaled to recognize instrumental breaks in accordance with certain exemplary embodiments;



FIG. 42 is a screenshot of karaoke-related musical rights information that may be displayed in accordance with certain exemplary embodiments;



FIG. 43 is an example screenshot showing a karaoke search mechanism in accordance with certain exemplary embodiments;



FIG. 44 is an example carousel-type display of predefined karaoke songbooks in accordance with certain exemplary embodiments;



FIG. 45 is an example song selection screen from a selected songbook for a karaoke jukebox in accordance with certain exemplary embodiments;



FIG. 46 is an example screenshot showing separation between first and second parts of a karaoke duet in accordance with certain exemplary embodiments;



FIG. 47 is an example screenshot of a karaoke games selection screen suitable for use in connection with certain exemplary embodiments;



FIG. 48 is an example screenshot showing how one example karaoke-related game may be played in accordance with certain exemplary embodiments;



FIG. 49 is an example screenshot that illustrates incremental karaoke song queue reveal features of certain exemplary embodiments;



FIG. 50 is an example screenshot showing how an authorized person may edit the karaoke jukebox queue in accordance with certain exemplary embodiments; and



FIG. 51 is a block diagram illustrating example components of a karaoke jukebox system in accordance with certain exemplary embodiments.





DETAILED DESCRIPTION OF EXEMPLARY EMBODIMENTS

Referring now to the drawings, FIG. 2 is a block diagram of an exemplary embodiment of an improved jukebox system 10′. The jukebox system 10′ includes similar elements as shown in FIG. 1 and described above, including a central server 12, communications network 14, and remote jukebox devices 16, 16a-16f. However, the jukebox system 10′ further includes local servers 22, 22a-22f respectively connected to each of the jukebox devices 16, 16a-16f. The central server 12 includes a master library of songs (and/or other content). Each of the jukebox devices includes a subset of the master library on a local storage device of the jukebox. The central server may be used to individually manage the contents of the jukebox device, by monitoring usage of and updating the subset of songs on each of the jukebox devices with the intent of maximizing the usage thereof. The central server 12 periodically receives data from each of the jukeboxes for the purpose of royalty accounting and payment for songs played. The jukebox devices may connect to the network in any suitable manner, such as dial-up modem or broadband modem (e.g., DSL, cable, wireless broadband, or satellite). The communications network 14 may be any suitable network capable of distributing data (e.g., audiovisual data) from the central server 12 to the jukeboxes 16 and enabling data to be uploaded from the jukeboxes 16 to the central server 12.


The songs (and/or other data) may be digitized, compressed and encrypted by the central server 12 prior to sending songs to the jukeboxes for security and bandwidth purposes using known techniques. The songs are then decompressed and decrypted by the jukeboxes for storage and reproduction thereon. Thus, each of the jukeboxes maintains in a database a library of digitized songs for play on the jukebox, wherein the library can be changed or updated through communication by the central server. The jukeboxes may also receive and store data constituting images (e.g., still and/or moving video and/or graphical images) that can be displayed on the display 18 of the jukebox device 16. In one exemplary embodiment of the invention, the jukebox devices have similar structure and operation described in U.S. Pat. No. 6,308,204 referenced above. Thus, the jukebox devices 16 each may include one or more microprocessors, such as a main CPU and an audio DSP, a memory, such as a hard drive, for storing songs and/or other content, a display of displaying visual items, an audio arrangement 20 for providing audio, a communication system for enabling the jukebox to communicate with the central server 12 through the communications network 14, and operating software, including a multitasking operating system, that controls the operation of the jukebox. The operating software also may be updateable through communication with the central server 12 as described, for example, in U.S. Pat. No. 6,308,204 referenced above. The jukeboxes 16 further include one or more payment devices, such as coin, bill and/or credit card input devices, for enabling a customer to pay for usage of the jukebox device in a convenient manner. The screen 18 may be a touch screen that enables the user to input selections by touching the screen.


Each jukebox device has, in one embodiment, a local server 22 that can be accessed by the jukebox device. The local servers are respectively connected to the jukebox devices using Ethernet or other type of local connection. In another embodiment, the local server may simply be a logical extension (e.g. partition, directory, or area) of the jukebox's hard drive, rather than a separate hardware device. The local servers 22 may each include a mirror copy of the master library of musical recordings maintained by the central server 12. The local server 22 can be loaded with the master library by the entity that owns and/or controls the jukebox network prior to shipping the local server and jukebox device to the jukebox distributor or operator. Of course, over time, the local sever will no longer correspond identically to the central server, due to the fact that the central server may be continually updated with additional or new songs. Thus, the local servers 22 also may be updated periodically to maintain a correspondence with the library on the central server 12. This updating can be done, for example, by the central server 12 through communication with the jukebox devices connected with the local servers 22 using, for example, either dial-up or broadband modems. Alternatively, the updating can be done personally with an update tool that can be connected by a routeman or other person directly to the jukebox or local server for the purpose of updating the contents of the local server. The portable tool could include a removable storage medium, such as a hard drive, that could be returned to and reused by the owner of the jukebox system for future updates. The tool itself could be kept by the operator or other person in charge of maintaining specific jukeboxes for use upon receipt of the updated removable storage medium from the owner of the jukebox system.


For security reasons, the local server 22 may not include all of the digital data that constitutes any one song that is stored on the local server 22. In addition, the part of the song that is on the local server is encrypted. The jukebox device 16 contains the missing part of each of the songs on the local server, thereby enabling the jukebox to assemble the complete song based on the contents of the local server and the memory on the jukebox device. The missing data located on the jukebox is needed in order to decrypt the songs. For example, a single block (or other small fraction) of data for each song may be missing on the local server but present on the jukebox device, and the encryption may be based on the missing block and may proceed on a block by block basis. Thus, none of the blocks can be decrypted without obtaining and/or decrypting a preceding block. This feature provides significant security and prevents or deters theft or other type of unauthorized use or copying of the songs on the local server. Thus, in this embodiment, each local server must be specifically assigned to a specific jukebox device so that the decryption described above can be properly performed.


In accordance with an exemplary embodiment, the local servers may also each be individually registered with and identified to the central server 12, so that the central server can individually manage and monitor each local server. The same is true for the jukebox device itself, i.e., it may also be registered with the central server so that it too can be individually monitored and managed by the central server. As will be understood from the foregoing description, the local servers become an important and advantageous part of the jukebox system by allowing the contents thereof to be accessed by the jukebox device to provide additional services (such as providing additional songs) not available on the jukebox device itself. As will be explained below, the song library of the central server and/or the storage capacity itself can be advantageously used to provide services to other jukeboxes, such as fee-based residential and commercial jukeboxes and/or other fee-based equipment. One use of the local servers may be to provide an immediate song downloading feature. As indicated above, certain exemplary embodiments include a new three-dimensional user interface for exploring and browsing media content, e.g., using a touch-screen. Given an arbitrary collection of content (e.g., music, video, advertisement, and/or other content), the new user interface may create a multi-dimensional (e.g., two- or three-dimensional) visualization virtual space, which allows the user to freely navigate in this collection. This type of navigation is facilitated in certain exemplary embodiments by using a social tagging model and/or collaborative popularity model.


Certain exemplary embodiments are described as implementing touch area sensing techniques. That is, in a sense, certain exemplary embodiments obtain a location a direction and a speed of a sensed touch. Using this data, certain exemplary embodiments may cause the prominent object plane to be moved forward or backwards in the object presentation by moving the touch forward (upwards) were backward (downwards) or selected by a swipe to the left or right. In addition, or in the alternative, by holding a touch for a period of time greater than some threshold (greater than 2 seconds, for example), a choice that is represents it could be presented.


The space is the general presentation area on which all graphic elements are presented. Collections generally refer to logical collections of media and may be subject to preferential filtering, e.g., to present and make available only media that satisfies criteria for different levels of actors. They may sometimes be represented by graphic images. Objects generally refer to songs, artists, playlists, games, or media sources that are represented in a sequence from a central catalog. Objects that are presented may sometimes be thought of as being subject to a match between the user or locations preferences and the attributes of the object was being browsed. As described in certain of assignee's co-pending applications (which are referenced above), an authentication mechanism may be provided to, among other things, identify the user and provide security credential authorization. In the context of certain exemplary embodiments, the authentication action may superimpose user-based preferential filtering over location-based preferential filtering. When user-based preferential filtering is in effect, it may further reduce whatever filtering has been imposed by the location attributes. Once authenticated, new collections such as playlists, favorite games, interrupted games, chat sessions, records of media consumption, high scores, etc., may become available as new possible collections to be presented as objects in the object viewer.


As alluded to above, in certain exemplary embodiments, the space operates in three states. In the flight state, a programmed series of media segments are presented based on the behavior of the jukebox and centralized media preparation activities. In the browsing state, users browse collections through a variety of different modes with the goal of selecting a media item for action. Media items typically include songs, games, and/or other content. In the search state, patrons enter data through a virtual keyboard and option panel or other appropriate input mechanism to directly locate media items that satisfy the search criteria and the preferential filtering in effect at that location. It will be appreciated that these states are exemplary and that these and/or other states may be provided in any suitable combination or sub-combination in different embodiments of this invention. In any event, illustrative details of exemplary flight, browse, and search states will be provided, in turn, below.


Details regarding an example flight state will now be provided. As will be appreciated, for jukeboxes, and music systems in general, only a portion of every day is spent interacting directly or wirelessly with the user who is searching for or playing a song. Most of the time, the lighting and display apparatus are engaged in presenting a sequence of video images. The goal of the video images is to attract users to the system, make them aware that the device is in fact a user controlled music service or jukebox, present advertisements and public service announcements, and/or to promote the system itself showing new features, new media available, and to promote artists and their works.


A problem with jukeboxes today is that they offer a largely homogenous presentation across a particular network. This means that the visuals presented may be wholly inappropriate for many of the locations in which it is installed. Perhaps more importantly, many restaurants or specialty bars or lounges have invested significant resources in creating a specific mood and appearance to attract and satisfy their intended clientele. As commercial music systems increasingly offer larger and larger screen display areas, the opportunity exists to tune the appearance of a jukebox via its visuals to a particular type of location. In addition, the larger screen area presents an opportunity to support increasingly sophisticated digital signage applications interleaved with the visual theme for the location. The downside risk of not responding to implicit or explicit requests for customized designs in order to make the jukebox blend in with the overall environment relates to having the idea of a jukebox placed in the location rejected out of hand.


One solution to this challenge is to offer custom flight packages, which may be thought of as being collections of video and composite imagery that are “authentic” to or otherwise “in tune with” the sensibilities of location. For example, a sophisticated urban Martini lounge could be provided with an urban, chic collection of videos appropriate for the location. A country and western bar could be provided a cowboy theme more in keeping with the surroundings. This has at least two advantageous effects on the industry. First, it opens up an entirely new population of venue types to the music service. Second, combined with sophisticated venue music styles, it creates a music system that has both a look and a sound that is appropriate for the clientele the venue operator seeks to attract and please.


In certain exemplary embodiments, flights may be composed of individual video segments. Video segments may occupy all or a portion of the available screen area and may be overlaid with other video segments of differing duration. Video segments may be resident within the music system in certain exemplary embodiments (e.g., having been distributed from a central server in advance), or they may be dynamically retrieved from music server or other content providers in certain other exemplary embodiments. In still other exemplary embodiments, channels that may be tuned into may be provided. Each flight may have a duration after which the flight would repeat. As such, it may be desirable to include dynamic content so as to keep the presentation fresh and interesting.


Video segments may be generally classified as being editorial or advertising. Editorial segments are generally seen as informative or amusing facts intended to engage viewers and encourage use of the music system, either directly or through a remote access device. Advertising segments are generally seen as being dynamically presented based on advertising materials located on the music system or on a remote server.


Flights can be thought of as a timeline where each segment represents a displayable group. Displayable groups are independent data systems designed to resolve the requirement for editorial or advertising content. A displayable group may be resolved in a single step where only one editorial segment satisfies the criteria of the displayable group, or a displayable group may require a great many rules and references to ultimately resolve the correct video segment to be presented.



FIG. 3 is a schematic view showing flight content resolution in accordance with an exemplary embodiment, and FIG. 4 is a schematic view showing display group resolution in accordance with an exemplary embodiment. As can be seen from FIG. 3, a flight content resolution service 302 may resolve the flight content to be displayed from a variety of sources such as, for example, local content 304, server content 306, and/or web-based content 308 sources. The resolved flight content may be displayed on a screen 310. In the FIG. 3 exemplary embodiments, the “current” segment is provided in the center of the screen 312a and retreats into the background 312b to make room for a “new” segment entering in the foreground 312c. For example, each video segment 312 may have an anchor position, a timeline, and an ending anchor position. This may help provide the ability to add motion and the illusion of three-dimensional space to the presentation. This information may be scripted in accordance with a predefined metadata format.



FIG. 4 may be thought of as a graphical representation of a script, for instance. That is, a flight control table 402 includes plural segments 404a, 404b, . . . 404n. These segments may involve displayable elements that, in turn, may include further displayable elements. The displayable elements in other words may be nested within each other, e.g., in dependence on a script. A “sub-display element” embedded within a display element may be programmed to move within that display element, across display elements, between segments, etc. Transitions thus may be defined as between segments, and as between display elements, and as between display elements across segments, in exemplary embodiments.


As a simple example, FIG. 5 is a flight segment appearing in one anchor point 502a and moving and zooming to a second anchor point 502b appear closer to the user in accordance with an exemplary embodiment. The flight segments may also enter and/or exit with transitions. It will be appreciated that anchor points and zooming may be from top to bottom, bottom to top, left to right, right to left, in or out, etc., and need not necessarily go from bottom to top when giving additional focus to a new content piece (e.g., as in the FIG. 5 example). Movements also may be more complicated in certain exemplary embodiments, e.g., including spiral-type movements, arbitrary movements, random or seemingly random type movements, etc.


As alluded to above (e.g., in connection with FIG. 4), a displayable group may contain one or more displayable groups. In other words, the displayable group resolution process may be recursive or nested in certain exemplary implementations. By maintaining a one-to-many relationship for each flight segment, it may be possible to create dynamic, engaging flights and reduce the effort of producing the content to satisfy all the permutations and combinations of a large display network. Once resolved, video segments of transition attributes for appearance and disappearance may be consulted to help control the appearance of multiple segments.


Until now, it has been extremely difficult or impossible to provide this type customized content across a large-scale network because of the enormous manual labor involved with strict playlist-based systems. The introduction of display groups and the recursive nature provides the ability for the display system to make systematic substitutions for content based on a combination of venue attributes, patron interaction, content availability, flight segment mix, etc.


Periodically, during the flight, a segment may resolve to a displayable group that encourages a patron to use the system. Once a system is touched, the flight mode may be altered. Once the user enters browse or search state, resolution and presentation of flight segments may be restricted to content that is of a space and duration appropriate for coexistence with the browse or search screen displays. For example, based on a setting, editorial content segments could be deferred until the end of the interactive session, but advertising result segments could be presented. In different examples, the flight could be interrupted and resume only after the interactive session was completed, or the flight could stop and a new flight specifically intended for presentation during interactive sessions could begin. In any event, the flights may conform to a style that is consistent with attributes of the venue created when last configured.


As identified above, the need for content reducing, personalizing, and/or features is becoming increasingly important as music collections grow in size and diversity. The browse and search features are a part of the new, natural 3D user interface described herein. This user interface is designed to help patrons find new music, re-find forgotten music, create coherent playlists, and/or the like. The new 3D interface is designed to be intuitive, engaging, and entertaining. The following descriptions provide additional details regarding the interactive 3D visualization techniques of certain exemplary embodiments that allow users to explore the overall music space, receive recommendations for music, login to social networking sites, and generate coherent playlist. These descriptions are provided in the context of exemplary browsing and search states mentioned above.


In the browsing state, users can browse the music collection by artist, by album, by song title, or by genre. In terms of exploring the music space, album artwork or jackets have always been a part of local content browsing. However, browsing content residing on the central server relies mainly on artist name. The interface of certain exemplary embodiments may help unify local content and central server content, as well as content available only from social networking sites, via “play-only” licenses (e.g., where content can only be streamed from a source and cannot be saved locally or on a remote server aside, perhaps, from limited music buffering), etc. Thus, music exploration may be driven by a combination of artist name and artist artwork in certain exemplary embodiments. The combination of artist and/or artist artwork based-discovery is advantageous in that there is a trend away from patrons recognizing albums and instead associating music with a strong brand or image (e.g., often an image of an artist, which helps explain the increasing incidence of artists putting their own likenesses on their album covers). For instance, FIG. 6 illustrates browsing a music collection for a given artist in accordance with an exemplary embodiment, and FIG. 7 illustrates browsing a music collection by artist and album in accordance with an exemplary embodiment.


In FIG. 6, for example, a user can drag albums or a screen location to the left or right to “flip through” different albums. FIG. 6 is more “traditional” in the sense that it sorts music for a particular artist by album. Any album may be directly selectable in certain exemplary embodiments, whereas only the center album (which has the focus) may be selectable in certain other exemplary embodiments. In any event, the ability to drag or swipe left and right makes the albums appear as if they are coming into the foreground as the approach the center and are fading into the background as they move away from the center. In this way, it is possible to simulate a more three-dimensional look-and-feel, and provide more of a sense of immersion when interacting with the user interface. In certain exemplary embodiments, similar albums may be grouped more closely to one another, whereas more disparate albums may be located more remote from one another. In certain exemplary embodiments, newer albums may be grouped more closely to one another and/or may be initially provided at the center. In general, general jukebox location and/or user preferences that are specifically entered and/or inferred from actual or predicted usage may be used to control the arrangement of the albums in the continuum presented in the FIG. 6 exemplary embodiment. New music first; most popular music first; user, location, or “hyped” albums first; rarely played albums first; and/or the like are all display/arrangement paradigms that may be specifically entered or inferred from actual or predicted usage and further may be implemented in connection with certain exemplary embodiments. It will be appreciated that multiple albums for multiple artists may be presented in a single FIG. 6 like arrangement, with similar artists and/or albums being grouped together (e.g., based on metadata or user tags associated with the artists, songs, albums, etc.) and/or displayed in accordance with the jukebox location and/or user specific parameters identified above. Furthermore, it will be appreciated that some forms of collections may place more than one object of the collection on each virtual plane of the series of virtual planes. In certain exemplary embodiments, a sort type may be specified or predefined, and metadata associated with the albums, songs, artists, jukebox plays, etc., may be consulted such that a processor of the jukebox is able to automatically generate the order and arrangement for the items to be displayed.


In FIG. 7, each artist is associated with a 3D object representing the artist name and artist artwork. Each artist object can have any number of albums. Albums related to the same artist form a group that satisfies a relationship constraint. When an artist object is moved in the 3D scene, all descendants in the scene graph move with it. The constraint relationships are stored in a graph called the scene graph, described in greater detail below. Similarly, each album may have any number of songs associated therewith. In the FIG. 7 exemplary embodiment, artists can be moved up and down to control inward and outward movement, respectively, and albums can be moved left or right to control the card shuffle like arrangement shown in FIG. 7. The albums displayed may update with the selection of an artist, and/or vice versa. This sort of collection view mode may alter one of more of the images, video segments, and text that is/are presented as the representative of each song, albums, artists, games, or content sources. In one version of the view mode for artists, the artist name may be presented in text along with an image of the artist that is different from an album cover. In one exemplary implementation, this image may be a promotional image that has recently been released by the label and is updated periodically. In certain exemplary embodiments, the collection viewer may not present an artist that does not match a sufficient number of the user recommendation properties regardless of whether it is available through the device or the network service.


The exemplary displays in FIGS. 6 and 7 may be enabled by mapping each axis to a characteristic of the music. For instance, one, two, and three-dimensional displays may enable a user to move in different directions, with each axis being mapped to a particular value in a data set of characteristics, and with the displayed items (e.g., album jackets or other identifiers) being updated as the user interacts with the display. X, Y, and Z axes may be used for a three-dimensional browsing experience, for example. In addition, or in the alternative (e.g., for 2D browsing), size, color, blurriness, and/or other effects may be applied with similar effect. Each element to be displayed may be assigned a depth value for each axis. As on example, album jackets may be displayed alphabetically by artist, song, album name, etc., on a first axis (with the alphabetical order being a first depth value), whereas song release date may be provided on a second axis, etc. (with the year being a second depth value). In this particular example, a user may move left and right to scroll between a first alphabetical arrangement of artists or albums (along the first alphabetical order depth value), whereas the user may move in or out to move forward or backward in time (along the second year depth value), e.g., to provide a sense of time tied to song or album release date, etc. This may help create the sensation of a curated, temporal browsing experience.


Of course, it will be appreciated that other characteristics also may be used such as, for example, beats per minute, genres, etc. It also will be appreciated that the depth values may be numeric in nature (e.g., as in the case with release date), representable in connection with numeric values (e.g., for names), or discrete in nature (e.g., for genres). For discrete values, discrete selections may be made. However, even discrete values like genres can be represented on a continuum, as rap and hip hop may be thought of as being “closer” to one another than they are to country or zydeco, for example.


The characteristics that are used to assign depth values may be obtained as metadata, e.g., from an outside provider, gathered from social networking sites (e.g., by examining a recognized jukebox user's preferences, “favorites,” “pages,” etc., and expanding that search outwardly to the user's friends' preferences up to a predetermined number of degrees), developed based on the usage of one or more jukeboxes, etc. The originating metadata may be stored in a database or other suitable non-transitory computer readable storage medium. It may be incorporated into the music catalog directly, a link to the originating metadata source or an outside data structure having the depth values may be incorporated into the music catalog or accessed by the jukebox upon a corresponding display request, etc. Further information regarding potential metadata sources is provided below.


In certain exemplary embodiments, a jukebox device comprises a display; a non-transitory computer readable storage medium storing a plurality of instances of media available for playback on or via the jukebox device; and at least one processor configured to cause a user interface to be displayed on the display and further configured to respond to inputs to the user interface. The user interface comprises a display area including virtual axes defining a coordinate space in which individual song, artist, and/or album items are to be displayed. Each individual song, artist, and/or album item has a plurality of characteristics associated therewith, each said axis is associated with one of said characteristics so that each individual song, artist, and/or album item has a defined location in the coordinate space. The display area is updatable in response to user input corresponding to movement within the coordinate space. The coordinate space may be two-dimensional, three-dimensional, substantially linear, etc., in different embodiments.


In certain implementations, items are arranged alphabetically along a first axis and chronologically along a second axis. In certain implementations, a release date associated with the item determines the item's positioning along the second axis. Discrete areas of one said axis may be designated for a predefined arrangement of discrete genres, and the predefined arrangement of discrete genres may be decomposable into a predetermined sub-arrangement of discrete sub-genres. The discrete genres may be grouped such that based on degrees of similarity therebetween such that similar genres are provided in closer relative proximity to one another as compared to disparate genres. Elements may be grouped alone one axis based on popularity in certain exemplary embodiments.


Song items may be selectable so as to trigger their playback, display a popup screen enabling the user to confirm that the selected song is to be played back, etc. Upon detecting an album or artist selection, movement within the coordinate space may be caused such that the selected item is centered therein.


Methods of making and/or using such jukebox devices also may be provided, as may non-transitory computer readable storage mediums tangibly storing instructions for using such jukebox devices/providing such user interfaces. Jukebox systems including plural jukebox devices, for example, also may be provided in different exemplary embodiments.


Similarly, having a database that contains a set of user or automatically generated descriptive words or “tags” associated for each artist, album and song titles, may enable certain exemplary embodiments to visually represent music collections in an interesting way. For instance, FIG. 8a illustrates a set of eight (8) example tags associated with “artist 6” in accordance with an exemplary embodiment. Each time a user presses and holds a music object such as artist artwork (e.g., in FIG. 6 or in FIG. 7), an album artwork or a song title label for a period of time greater than a predefined threshold, a set of floating objects may be displayed, e.g., as shown in FIG. 8a.


Each tag may be represented by one of the floating objects. The size of the floating objects may in certain exemplary embodiments correspond to the weight of the tag. The weight may represent how important or relevant that word is for that music object. The weighting may be determined by a source such as an authorized user, based on an aggregation of user-specified tags, a predefined formula that balances factors such as, for example, genre, release date, user-specified information, etc. The tags may be based on how many people have tagged an item a certain way, how a trusted metadata source has tagged it, etc.


Clicking on one floating object may in certain exemplary embodiments display more music objects sharing the same tag. The resulting collection may represent similar artists, albums, or songs by means of the selected tag. For example, clicking on tag 3 in FIG. 8a may present a collection of artists filtered by tag 3 using a user interface depicted in FIG. 6.



FIG. 8b is a wireframe used to construct an example music map on a jukebox display in accordance with an exemplary embodiment. Among other things, the FIG. 8b screen represents a “visualizer” display in accordance with the display techniques described above. In the FIG. 8b example, a main area provides a three-dimensional user interactable area. Items are arranged and displayed in accordance with the techniques described above. An overview area 802 provides the user with an intuitive feel as to how the display is organized on a larger basis showing, for example, the elements within genres (jazz, rock, and pop), how those elements have relative sizes and locations, how the genres are related or organized in relation to one another, etc. These nodes 800 may be user selectable so as to traverse genres, artists, albums, etc., and they may be sized base don tags, e.g., as described above. A “node history” section 804 provides a search history that is related to the user's movements (and thus implied search methodology) and enables quick jumping back to locations within the overall music map. Breadcrumbs 806 may display a more focused path rather than a more detailed traversal of all nodes visited, e.g., as in the node history section 804.


The illustrative wireframe shown in FIG. 8b includes other elements such as, for example, a leader 808 indicating the name of the establishment, a main advertisement leaderboard 810, an indication of the song currently playing 812, an indication of the user current logged in 814, and a “set list” 806 of the user. Further details regarding set lists are provided below. A text-based search area 818 also is provided, as is a multi-area semi-circular display for enabling a user to select different collections and/or sort different selected collections.


The questions of what media is to be displayed and how such media is to be displayed have been partially answered above. However, a more complete description will be provided below. With respect to the former question, location-specific inclusion criteria may be specified and matched with song metadata so as to define a “rough” master list for a location in certain exemplary embodiments. This location-specific master list may be “refined” based on metadata associated with a recognized user and/or based on jukebox usage over time in certain exemplary embodiments. Finally, the location-specific mater list (with or without having been refined with recognized user information) may be wholly or partially overridden by a custom-defined event (e.g., a disco or other theme night, private party, etc.).


With respect to the latter question as to how media is to be displayed, it will be appreciated that the emphasis of certain exemplary embodiments evolves away from album art and towards “stronger” identities or branding associated with an individual song or groups of songs and typically associated with an artist or an artist's likeness. In any event, a “rough” ordering may be based on location-specific criteria such as, for example, newest music first, Billboard's most popular music first, most popular at location last, etc. As above, this information may be “refined” based on metadata associated with a recognized user and/or based on jukebox usage over time in certain exemplary embodiments. For instance, criteria may be new music first, pop music last, music played by, listened to, or rated highly by friends (and friends' friends, etc.) first, music most played on MySpace first, etc. And, as above, overriding event parameters also may be provided (e.g., to place an emphasis on new music, independent artists, 80s hair bands, etc.).



FIG. 9 is a block diagram illustrating one arrangement that may enable custom browse state and other features to be implemented in accordance with an exemplary embodiment. In FIG. 9, a jukebox 902 is connected to a local database 904 that stores content. This local database 904 may be integrated in to the jukebox 902 in certain exemplary embodiments. The jukebox 902 itself provides a user session and at least initially receives information regarding the location's profile. The location's profile information ultimately may be uploaded to a central server 906 to which the jukebox 902 is connected. The central server 906, in turn, may be operable connected to a centralized database 908 of media, and it may store tag, metadata, user profile, and/or other information. As shown in FIG. 9, the jukebox 902 is directed connected to social networks 910 such as, for example, MySpace, Facebook, Last.fm, and/or the like. Each of these social networks may maintain its own tag, metadata, user profile, event, and/or other information. A middleware component (not shown) may help the jukebox interface with the social networking sites and the information associated therewith (e.g., to search for, retrieve, update, and/or otherwise manipulate the data). In certain exemplary embodiments, the jukebox 902 may not be directly connected to the social networks 910 and, in certain exemplary embodiments, the central server 906 may function as a gateway to the social networks 910. This arrangement may be advantageous in certain exemplary implementations, as the central server 906 may serve as a buffer or de facto firewall between the social networking sites 910 and the jukebox 902. This arrangement also may be advantageous in certain exemplary implementations, as the data retrieved from the social networking sites 910 may be of potential value to multiple jukeboxes (e.g., in a broader jukebox system), and storing and processing this information on a central server 906 may be desirable to reduce the loads on individual jukeboxes, particularly where common information can be shared.


In certain exemplary embodiments, a method of recommending a song to a user is provided. The method may comprise: enabling a user to log in to a jukebox device; determining musical preferences of the user based on a social networking site profile of the user, the social networking site profile identifying at least one artist, song, and/or genre; providing the determined musical preferences of the user to a recommendation engine to develop at least one recommendation; and providing the recommendation to the user on the jukebox device. In certain exemplary embodiments, musical preferences of friends of the user may be determined based on social networking site profiles of the friends, and the determined musical preferences of the friends of the user may be provided to the recommendation engine in developing the at least one recommendation. The determining of the musical preferences may be expanded to friends of the user and friends of friends within a predetermined number of degrees of separation (e.g., 1, 2, etc.), and the expanded determinations may be provided to the recommendation engine in developing the at least one recommendation.


Profile information corresponding to the location in which the jukebox device is located also may be of use in developing the at least one recommendation. Such profile information may be maintained on a central server, for example and, similarly, at least one processor of the central server may help determine musical preference data via a social networking interface provided thereto. In certain example embodiments, a plurality of social networking sites may be consulted to determine musical preferences of the user, with each said social networking site having an associated social networking site profile for the user.


Methods of making and/or using such jukebox devices also may be provided, as may non-transitory computer readable storage mediums tangibly storing instructions for using such jukebox devices/providing such user interfaces. Jukebox systems including plural jukebox devices, for example, also may be provided in different exemplary embodiments.


Having a collaborative filtering engine based on the listening behaviors of other media consumers and a tagging system as described above may allow certain exemplary embodiments to represent the recommended media collections in a 3-dimensional space. One issue for a collaborative filtering engine is the need for a significant amount of data to provide good recommendations. For new media or media with few consumers, conventional engine oftentimes cannot generate good recommendations. Certain exemplary embodiments address these shortcomings by recommending media upon the similarity of tags that have been applied to the media.



FIG. 10 shows a 3D visualization of a recommended song collection in accordance with an exemplary embodiment. Songs objects 2, 3, 4, 5, 6, 7, 10 that are close to song object 1 represent the following paradigm: customers who listen to song (A) also listen (or are likely to listen) to song (B). A size dimension may be introduced in certain exemplary embodiments, wherein larger objects represent more frequently played or otherwise favored songs. When a user clicks on a “Generate Playlist” or similar button, the user may select two “end” song objects in the recommended collection. This may prompt the system to generate a path through the collection that connects the closest song object together. This may be accomplished using, for example, a suitable algorithm such as a breadth first search, depth first search, Bellman-Ford, Dykstra, etc. The resulting playlist may be presented to the user for approval, modification, rejection, etc. in different embodiments of this invention. In certain example embodiments, the songs may correspond to nodes in a graph, whereas the size of the songs may represent weights or distances between adjacent nodes or to a center node. Of course, the same or similar techniques may be used for artists, albums, genres, and/or the like.


Once a user is identified, the system provides ongoing application and/or tracking of the individual's musical preferences and affinities. For instance, certain exemplary embodiments may provide artist and song recommendations produced through the aggregation of the individual's explicit selections, metadata-based associations, social network suggestion indicators, and/or the like. As musical tastes are varied, the system may maintain a list of the individual's “core” affinities. Such core affinities may be an artist, artists similar to a specific artist, a song, songs similar to a specific song, musical selections of another individual or individuals (e.g., as available from a social network music site or sites), recommended artists or recommended songs from user-identified music sites, etc. Any match from a user's musical taste web and the music catalog available from the device or from a server or servers providing music catalog may in certain exemplary instances result in a subset of the music, presenting material that has matched the user explicit, user metadata, user social network friends recommendation, and user music site recommendations. As will be further described below, this music may be a compared with the music library available at a given site, and the overlap of these two groups may be presented for user selection.


Further details of an exemplary search state will now be provided. A search typically would be initiated to enable a user to find a list of songs, or a single song, that the user would like to play or add to a collection. Oftentimes, the user does not know the correct spelling of the artist, album, or song. Or the user may know the name of a particular artist but does not necessarily know the exact group or band with which the artist recorded a particular song. Moreover, a user often remembers particular lyrics to a song but not necessarily the song title itself. As still another example, a song may be memorable because of its association with a movie or television show. This information has been very difficult to obtain before the development of entertainment websites focused on providing additional details surrounding the use of music, its popularity, its lyrics, the artists involved in its creation, important performances of the work, and many other music related data. A search using the techniques described herein could help overcome these and/or other challenges.


To accomplish this type of search, artist and song related metadata may be collected from a variety of sources. Searchable criteria may include song genre, popularity (high ranking or frequent presence) on a particular source within a particular time frame, etc. A user may select the song criteria (for example “country”), the timeframe (e.g., “today”), and the source (e.g., heard on “MySpace”). The qualifiers may appear as optional selection buttons offering a choice for each selection criteria.


As public sources for music information become available, a metadata collection system may help provide a relevant source for data ingestion. For instance, users may be able to identify their music styles, e.g., by drawing on source and social data that is relevant to them. An example music style search tool may provide a broad variety of sources, along with a very granular genre and subgenre selection, and/or other music attributes. Further data regarding a song, for example, may be provided by an outside service (e.g., title, artist, album, label, genre, theme, etc.), or may be determined from the instance of media itself (e.g., tempo or beats per minute, instrumentation, etc.).



FIGS. 11-14 illustrate how media may be searched in accordance with an exemplary embodiment. In particular, FIG. 11 is an illustrative screen showing how objects may be navigated in accordance with an exemplary embodiment. The objects in FIG. 11 are represented by triangles, and the objects may correspond to individual or collections of artists, albums, themes, genres, songs, etc. The order in which objects are presented may be determined as specified above. A user may navigate the collections in a 3D-like manner by swiping up or down to “zoom” in or out. Pressing and holding or dragging a particular object to a blank area may indicate a selection.


In this regard, FIG. 12 is an illustrative screen showing how an object may be expanded upon selection in accordance with an exemplary embodiment. For instance, once an object is selected, it may “explode” to reveal the songs associated with the selected object. This may include, for example, songs recommended for the user, event, and/or venue. The recommendations may be based on a combination of the location- and user-specific metadata, as well as any relevant event information, as compared with metadata and/or tags associated with the instances of media themselves.



FIG. 13 is another illustrative screen showing how objects may be navigated in accordance with an exemplary embodiment. In FIG. 13, a user may rotate wheels to specify, for example, whether the objects represent artists, genres, or songs; whether the instances of media are to be associated with friends of a recognized user and, if so, which friend; etc. In general, this mode selection action alters the collections being presented as objects. For example, a mode choice of artist could in certain exemplary instances replace a series of album object virtual pains with a series of artist object virtual pains (e.g., all subject to preferential filtering). In the case of FIG. 13, the objects represent genres, the genres are limited by the recognized user's friends' genres, and the particular friend is Sally. As such, Sally's genres are retrieved. However, only those genres that satisfy the venue's choice and the event parameters, are present on the music device or music device server, etc., are made available for expansion (e.g., as explained above). The FIG. 13 example also enables a user to enter word fragments (e.g., song names, genre names, lyrics, etc.) to reduce the songs being proposed for playback. FIG. 14 is similar to FIG. 13, except that FIG. 14 is an illustrative screen showing how a user may enter lyrics to reduce the number of songs being proposed for playback in accordance with an exemplary embodiment.


If the user is simply searching for songs, songs that are present on music device or available from the music device server may be reduced from the domain of all music as follows: availability on the server, availability on a music device, present in venue music style group, then optionally, all music available for this user, or music from my Music Style as recorded on my profile. FIG. 15 is a schematic representation showing how songs may be selected for presentation to a user in accordance with an exemplary embodiment. In the FIG. 15 example arrangement, starting with the largest sphere and moving inwardly, the collections include all music, music on a server, music on a music playback device, music in the venue's music style, music in the venue's particular event style, and music in the user's system. This or a similar scheme may also be used for browsing and/or other features of the jukebox. For example, it will be appreciated that music may be limited in different orders, some of the limiting spheres may not be applied, different limiting spheres may be applied, etc., in different embodiments of this invention.



FIG. 16 is a flowchart illustrating an example process for determining which content should be available for playback on a jukebox based on location-specific inclusion criteria and actual jukebox usage in accordance with an exemplary embodiment. All music is initially provided to the jukebox in step S1602. For instance, instances of media from the central server and/or a local server are at least initially available. Inclusion criteria is then specified by an authorized person such as a bar manager, etc., in step S1604. The inclusion criteria may be location specific in the sense that the inclusion criteria may not be relevant to other locations. For instance, an authorized person may specify inclusion criteria including one or more genres, plus the Billboard top 100, plus media associated with five friends. Once the inclusion criteria has been specified, in step S1606, a heuristic is applied to determine the appropriate music for presentation on the jukebox. The heuristic may build a “map” of acceptable media for playback. In certain exemplary embodiments, the heuristic may operate such that each song is assigned a likelihood that it would be desirable for play at a location, and each likelihood may have a confidence level associated therewith. A first rule for certain example implementations is that if there is doubt as to whether the song should be included or removed, then the song should be included. In certain exemplary embodiments, this may mean that if either the likelihood for inclusion meets or exceeds a predetermined threshold or the confidence level is at or below a certain threshold, then the song should be included. A second rule certain example implementations is that songs should be removed if they fall below a certain likelihood for inclusion along with a certain confidence. The scale may be adjustable over time as more and more data is acquired. In other words, the requirement for confidence may be lessened as more and more data is acquired. Certain exemplary embodiments aim to avoid offering a song that a patron has no link to, although this may not always be possible since a patron's decisions are not completely predictable. In any event, the appropriate music is displayed at the appropriate time(s) in step S1608, e.g., as a result of a search, browse, etc. The system may monitor for changes to the heuristic map based on jukebox usage over time in step S1610 and update the mapping accordingly, e.g., by returning to step S1606. It will be appreciated that the system may become “smarter” or “learn” over time, e.g., as more and more data is input.


In certain exemplary embodiments, a logit and/or probit calculation may be made for each song to determine the likelihood of leaving it in or removing it as an option for playback. The location information may be used to define the overall domain of the location or, in essence, setup the overall “vibe” of place.


In one or more steps not shown in FIG. 16, event information, if any, may be used to further restrict the overall environment, at least temporarily. In such cases, event information may provide an optional override, even to the overall “vibe” criteria. Similarly, in one or more steps not shown in FIG. 16, the system may be provided with the programmed logic circuitry for individually filtering/sorting media based on user-specific metadata. This user-based information may not necessarily as important as the overall vibe criteria in certain example implementations, but certain exemplary embodiments want to provide the opportunity to drill down and provide more detailed recommendations based on layers and layers of tags and/or metadata associated with the users and/or songs.



FIG. 17 is a flowchart illustrating an example process for initially setting up a jukebox in accordance with an exemplary embodiment. In step S1702, an authorized user (e.g., location staff member) inputs inclusion criteria during initial setup of jukebox. The inclusion criteria may be indicative of the type of location (e.g., Irish bar, hip hop dance club, country line dancing, biker bar, ultralounge, etc.), characteristics or demographics of typical patrons (e.g., age range, race/ethnicity, etc.), content ordering preferences (e.g., most popular first/last, newest first, etc.), etc. The jukebox then converts the inclusion criteria into metadata to be associated with the location in step S1704. This information is then communicated to the central server to determine the appropriate master playlist and overall look-and-feel for the jukebox based on inclusion criteria in step S1706. This process may include matching metadata associated with the location to metadata and/or tags associated with songs, the metadata/tags possibly being stored in a jukebox database of the central server, on social networking sites, and/or the like. Furthermore, additional information may be retrieved from external sources, as appropriate, and matching may then be performed on this further information. For instance, certain exemplary implementations may retrieve information from social networking sites, Billboard, etc. In this way, it may be possible to glean information regarding pure buzz or hype, popularity, etc. The need for determining whether to reach out to external sources may be based in part on a recognized user's information and/or recognized users' information may be a source of data in itself. It will be appreciated that the central server may have a plurality of predefined categories and/or channels in certain exemplary embodiments, and the jukebox may select one or more appropriate category(ies)/channel(s) in such embodiments based on the matching. In any event, information regarding the master playlist and look-and-feel for jukebox is communicated from the central server to the jukebox in step S1708, and the jukebox is provisioned accordingly in step S1710. The provisioning may including setting the flight state content and/or flight state content source, setting the master music library for the jukebox, setting the “skin” for the jukebox, etc. It will be appreciated that steps S1706 to S1710 may be performed periodically or dynamically to help ensure that the jukebox is up to date and providing appropriate music.



FIG. 18 is a flowchart illustrating an example process for customizing the display for a recognized user in accordance with an exemplary embodiment. The master music library is provided in step S1802, e.g., in accordance with the FIG. 17 example process. In step S1804, the user logs in. In step S1806, information associated with recognized user is retrieved. Such information may be stored locally on the jukebox, in the central server, on a social networking site, etc. Such information may include user-specified information concerning song plays, preselected favorites (e.g., artists, songs, albums, genres, etc.), demographics, etc. Such information also may include information gleaned from jukebox usage over time (e.g., the user prefers newest music, prefers particular order, only likes certain music at certain times/days/locations, etc.). The information associated with he recognized user is used as further inclusion criteria to customize selections for the user in step S1808. Such information may be used to further limit playlists, make “appropriate” recommendations, order or reorder the media presented to the user, etc.


In steps not shown, but as a process associated with step S1808, the jukebox may communicate with the central server to determine appropriate user-customized playlist/user-customized presentation based on recognized user inclusion criteria. Similar to the above, this sub-process may involve matching metadata associated with the recognized user to metadata and/or tags associated with songs. Further information may be retrieved from external sources, as appropriate, and then matching may also be performed on this further information. For instance, it is possible to gather information from social networking sites, one or more specified social network sites or sites of which user is a member, etc. User's data, user's friends' data, users' friends' friends' data, etc., may be of interest. This information may be communicated regarding the user-customized playlist/user-customized presentation may be communicated from the central server to the jukebox, and the jukebox may be provisioned accordingly, as above. Similar to the above, this may be performed periodically or dynamically (e.g., on recognized user login). Ultimately, in step S1810, user browsing and/or search is enabled in accordance with the user-customized playlist/user-customized presentation.



FIG. 19 is a flowchart illustrating an example process for customizing the display for a special event in accordance with an exemplary embodiment. The master music library is provided in step S1902, e.g., in accordance with the FIG. 17 example process. Information about the special event is received in step S1904. Such information may include, for example, information about the type of event (e.g., a private party with a DJ, disco night, rave night, etc.), the duration of event, whether the event recurs, etc. In step S1906, the information associated with the special event is used as further/alternate inclusion criteria when the jukebox communicates with the central server to determine appropriate playlist and look-and-feel for the jukebox. For instance, the further inclusion criteria may further limit playlists or sometimes generate a new master music library, make recommendations appropriate for the event, order or reorder media appropriately, etc. Information concerning the event playlist and look-and-feel for the jukebox is communicated from central server to the jukebox in step S1908, and the jukebox is provisioned accordingly in step S1910.


In one or more steps not shown, user login may be enabled. However, some or all of the recognized user functionality may be overridden by virtue of the event taking place. That is, the jukebox system may or may not perform further limiting and/or make suggestions based on the recognized user information. Similarly, the jukebox system may or may not use other user preferences gleaned from general JB usage to provide recommendations/reordering. In certain scenarios, a user may search for a particular artist, e.g., to search for all albums or songs that artist has provided. In such a case, the jukebox may update its display and to become more suitable for the particular artist, album, genre, etc. For example, FIG. 20a is a first example “artist portal” in accordance with certain exemplary embodiments. In the FIG. 20a example, a user has searched for all albums by Alicia Keys. This may cause a change to the display in terms of, for example, the underlying or background image, the color scheme of the user interface, the wireframe layout of the elements on the display, the external lightshow, etc. In the FIG. 20a example in particular, an image of Alicia Keys is displayed as a high resolution underlay beneath the user interface. The albums from the artist's oeuvre are the only one displayed because they are the only ones that match the search. The color scheme has become a more black and white color scheme, e.g., in accordance with Alicia Keys' Internet presence. In certain exemplary embodiments, a link to the artist's homepage, Facebook or other social networking or other site also may be provided. In keeping with the simplified look-and-feel of the Alicia Keys' Internet presence, many optional control elements (e.g., as shown and described in connection with the FIG. 8b arrangement) may be omitted.


It will be appreciated that different artists may have different artist portals, and the updates (e.g., to the wireframe, elements shown, etc.) may be the same or different than those described above. In that regard, FIG. 20b is a second example “artist portal” in accordance with certain exemplary embodiments, e.g., for Santana. As can be seen, the Santana artist portal has a different background art and a different color scheme, as compared to the Alicia Keys artist portal. In addition, the albums are not limited to Santana albums in the FIG. 20b example, as the user arrived at the Santana portal via a different means, navigated away from it (e.g., by implementing a subsequent search), etc.


Various “skins” may be attached to albums, songs, or the like, similar to how depth values may be assigned above. The concept of a skin includes, in certain exemplary embodiments, one or more of images, themes, wireframe displays, web pages, etc., that is/are context specific. Thus, it will be appreciated that the selection of an album, song, artist, etc., may cause a lookup of an appropriate skin and cause a corresponding change in the overall look and feel of the jukebox user interface. The skin may include item selection and/or arrangement information indicating, for example, which control elements are to be selected, where they are to be located, how they are to be sized/shaped/colored, etc. Corresponding information may be specified for non-functional elements such as, for example, background art, etc.


Although certain exemplary embodiments have been described in relation to an artist portal, the techniques may be applied to other concepts. For instance, a skin may be created by a recognized user at the jukebox or remotely, e.g., via the MyTouchTunes user interface, and the skin may be applied upon user login. A user may customize its jukebox skin in accordance with the above defining, for example, which control elements are to be displayed and how they are to be displayed, background art, etc. Similarly, a record label, group of artists, etc., also may have predefined skins that are applied at the above-described and/or other points. As is known, some record labels have distinctive presences, iconic images, etc. It will be appreciated that Def Jam's skin may be markedly different from the MuzikMafia's skin, as well as El Cartel Records' skin, etc.


As is known, jukeboxes typically only display a currently playing song. However, a queue is kept and could be displayed on a display in full or in part. For instance, FIG. 21 is an example of a partial or incremental presentation of the jukebox's play queue in accordance with certain exemplary embodiments. A user may be charged a premium fee for partially and/or incrementally revealing the jukebox play queue. For instance, a pricing scheme may be implemented such that the user pays a first amount for seeing an immediately subsequent selection queued up for playback, and increased amounts for seeing more and more selections in the queue.


The FIG. 21 playlist reveal example shows upcoming song titles and artist names. However, the display techniques described herein may be caused to display on selective bits and pieces of the upcoming queue, e.g., so as to not give away too much information. The selective reveal of information may be tied to the user's proclivity or inclination to use a “jump the queue” or “play it now” feature of the jukebox for a particular selection. A pricing structure may be implemented such that skipping 1 or 2 songs costs X credits, skipping 3 or 4 songs costs X plus some further amount of credits, etc. In other words, the price for skipping songs may vary based on the number of songs to be skipped. It also may be tied to the user's proclivity or inclination to use a “lock in” feature, e.g., that charges a premium price for ensuring that the song will be played, that it cannot be skipped without a user paying at least a user-paid or predefined reservation price, etc. The reveal of information may selectively include one or more attributes of one or more upcoming songs including, for example, title, artist, “mood” or genre, beats per minute, etc. The selective reveal may provide, for example, full title and artist information regarding the next two songs (which may be pop songs) and indicate that 10 more songs (which also may be pop songs) are in the queue. In so doing, jukebox patrons, proprietors, and other visitors to the location may be able to get a sense as to what the mood in the establishment will be like in the future. The selective reveal also may be tied to a timeframe, e.g., so that people get a sense of what the establishment will feel like over a predetermined time period of, for example, the next 20-30 minutes, the next hour, for the whole night, etc.


In certain exemplary embodiments, a jukebox device comprises a display; a non-transitory computer readable storage medium storing a plurality of instances of media available for playback on or via the jukebox device; at least one processor configured to maintain a queue of instances of media to be played back on the jukebox, the at least one processor being further configured to cause a user interface to be displayed on the display and further configured to respond to inputs to the user interface. The user interface may include an indication of a currently playing instance of media including at least a song name and an artist associated with the instance of media; and an incremental queue reveal module configured to cause the user interface to display information about one or more upcoming instances of media, the information at least initially excluding one or both of song names and artists.


The incremental queue reveal module may be configured to cause the user interface to display information about a predetermined number of instances of media, with the predetermined number being greater than 1. In the alternative, or in addition, it may be configured to cause the user interface to display information about a plurality of instances of media, the number of instances of media depending on the number of instances of media in the queue that are playable within a predetermined amount of time. The predetermined amount of time may be, for example, 30 minutes, 1 hour, 2 hours, the amount of time until closing, etc.


In certain exemplary embodiments, the incremental queue reveal module is configured to cause the user interface to at least initially display only information about a genre associated with each of the one or more upcoming instances of media, only information associated with a number of beats per minute associated with each of the one or more upcoming instances of media, etc. In certain exemplary embodiments, the initial display may include information only about an artist of each of the one or more upcoming instances of media.


In certain exemplary embodiments, the at least one processor is configured to activate the incremental queue reveal module upon payment of a fee. The amount of information revealed may be made to vary in dependence on the fee collected such that a larger fee corresponds to more information being revealed. Furthermore, the at least one processor may be configured to administer a first pricing scheme that relates the amount of fees due with the amount of information revealed with respect to each instance or media and with respect to a number of instances of media for reveal.


In certain exemplary embodiments, a queue jumping module may be configured to enable the user to move a selected instance of media to a user-specified location in the queue. The at least one processor may activate the queue jumping module upon payment of a fee. Furthermore, the at least one processor is configured may administer a second pricing scheme that relates the amount of fees due to a number of items in the queue to be jumped. The fee may be made to vary proportionally with the number of items in the queue to be jumped. The user interface may be programmed to enable a user to lock in an instance of media, upon payment of a fee, guaranteeing that it cannot be skipped. The incremental queue reveal module may be further configured to indicate whether instances of media are locked in such that they cannot be skipped.


Methods of making and/or using such jukebox devices also may be provided, as may non-transitory computer readable storage mediums tangibly storing instructions for using such jukebox devices/providing such user interfaces. Jukebox systems including plural jukebox devices, for example, also may be provided in different exemplary embodiments. Because the user interfaces of certain exemplary embodiments may be perceived by some users to be complicated, various techniques may be implemented to help bring focus to patron-selectable elements, e.g., song selection navigation. Blurring, fading, enlarging, and/or other techniques may be used to manage the user's focus in different embodiments, and thus may help simplify the experience, even though there are many more features available and those features are more powerful. For instance, FIG. 22a uses a colorized “blob” to highlight the center selection, whereas other non-patron selectable elements may be blurred, in accordance with certain exemplary embodiments. FIG. 22b shows the enlargement of an album jacket to highlight its selection, in accordance with certain exemplary embodiments. In the FIG. 22b example, the background UI elements have not been modified to bring additional focus to the selected item. By contrast, FIG. 22c shows the enlargement of an album jacket to highlight its selection, and further blurs background UI elements to help bring focus to the selected album, in accordance with certain exemplary embodiments. It will be appreciated that the background UI elements (e.g., in FIGS. 22b and 22c, for example) may be reduced to black-and-white images or otherwise color adjusted to bring focus on one or more items of interest. Although FIGS. 22a-c have been described in connection with the selection of an album, the highlighting techniques described herein may be applied to other aspects of the UI experience.


Certain exemplary embodiments may also build mosaics, e.g., from album covers or other artwork. An algorithm may be run on source artwork to make mosaics for use in the jukebox's user interface. Such an algorithm may run in real-time on the jukebox, e.g., to accommodate various mosaic designs based on the source artwork available (e.g., the number of album jackets available) and dynamically employ such mosaic designs in the user interface. A mosaic construction algorithm may use facial and color density recognition to arrange (or prevent the arrangement of) the source artwork in a desired patter, in certain exemplary embodiments.


Such mosaic techniques involve several concepts, including source image selection and manipulation (e.g., for the images used to build the mosaic), the destination image creation (e.g., for the target mosaic being created), and the destination image context (e.g., for how the target mosaic is used). The source images, destination image, and context in which the destination image is to be used may contain a layer of metadata. This metadata facilitates (e.g., through ID matching) to find a wealth of metadata that can be reused to feed a mosaic-creation algorithm. For instance, an album jacket is associated with a genre, style, artist, year, and other information. Play behavior and social network information about those musical elements may also be included to feed the algorithm information about the relative popularity of music elements, the placement of the music elements in relation to a social network environment, etc. For instance, more popular music may be larger, music played by a patron and a patron's friends (to a predetermined degree of separation in certain examples) may be enlarged, etc. In one context, an avatar of a user may be built from albums jackets matching the user's musical preferences, a map of the USA drawn from actual plays as they occur on the network, etc. Further details about these techniques are provided below.


With respect to source image selection and manipulation, mosaics may be built from source images (tiles) that are aggregated to simulate a larger target image. The tiles may be made from a whole image or from a sub-selection of the whole image, a modified source image (e.g., the original image's colors or hues have been altered), etc. Similarly, tiles may be used in any angle, flipped, mirrored, etc., to assemble the target. Image types may be arbitrary and may include, for example, album jackets, artist artwork, jukebox avatars, musical instruments, etc. Certain tiles may be flagged for special treatment by the algorithm. For example, popular albums, songs, or artists may be enlarged, etc.


With respect to destination image creation, it will be appreciated that the mosaic destination is the image being produced using the source tiles. A target may be a static image (e.g., artist artwork, a cowboy hat, the TouchTunes logo, etc), or a moving concept (e.g., an animation or video), or some combination thereof. Other possible destination concepts also are possible. With respect to destination image context, a variety of potential uses for constructed mosaics are possible. For instance, an artist image (e.g., for use in an artist portal as discussed above) may be built, e.g., using the artists' album jackets. As an example, an image of George Straight may be built using album art from his numerous albums. An image describing genre may be built, e.g., using source tiles relevant to that genre. As an example, a cowboy hat or popular artist like Taylor Swift may be built using only country music album jackets). A target may be built using geographical information (e.g., build a map of the USA using album jackets that are popular for a specific region). A background mosaic also may be built, e.g., using current jukebox profile information (the location's most popular genre, the currently playing song, etc.). Mosaics may represent playlists, albums from an artist, etc. There are, of course, other possibilities for the context in which the mosaics may be used.


A user may interact with a mosaic in certain exemplary embodiments. For instance, a mosaic may be enlarged by a user, and the user may select a song, artist, album, playlist, etc., from the mosaic.


Techniques also may be provided for synchronized partial lyrics display. That is, in certain exemplary embodiments, lyrics may be displayed at an appropriate or predetermined time during a song. For example, if the jukebox were playing the Black Eyed Peas song, “I Gotta Feeling,” the words “I Gotta Feeling!!!” could be displayed at the same time they are sung by the artist.


Lyrics also may be arranged in a pattern that reflects the context of the onscreen lyric snippet, also known as kinetic text. FIGS. 23a-d show example screenshots of a kinetic motion display in accordance with an exemplary embodiment. The lyrics have been synched with the Blink 182 song, “Josie.” FIG. 23a flies in at the start of the song. FIG. 23b includes words that are added as they are sung, with stretching of the word “of” and with the placement of the words within other words as they are moved out of focus. FIG. 23c shows how blurring can be selectively used to match the mood of the song, as it may be made to represent “frustration.” FIG. 23d illustrates how color can be used to emphasize a word as it is emphasized by the artist singing the song.


Kinetic text is enabled in certain exemplary embodiments by creating a script that matches or synchs with the song. Thus, a track of metadata including timing, mood, motion, color, image, etc., may be created for a song, and the track may be stored, e.g., as described above. In certain exemplary embodiments, an XML or XML-like tag structure may be created to link together timing, lyrics, and effects including, for instance, sizing/resizing, entry/exit points, zoom from/to locations, pan directions/amounts, etc.


Similar to the above, synchronized lighting projection also may be provided in certain exemplary embodiments. The user interface coloration may be modified to synchronize with the context of the currently playing song. For example, if the jukebox is playing DJ Funk's “The Roof is on Fire,” the user interface may display bright red and orange colorations, e.g., when the words “THE ROOF IS ON FIRE!” are sung. Themes may be developed and associated with one or more songs. For example, a “fire” theme may be created, wherein the jukebox user interface is modified to red and orange colorations. The theme may be associated with songs like DJ Funk's “The Roof is on Fire,” the Bloodhound Gang's “Fire Water Burn,” Johnny Cash's “Ring of Fire,” the Ohio Players' “Fire,” etc. The themes may be linked to the songs, and they may be applied for the entirety of the song in certain example implementation. Alternatively, or in addition, XML or other scripts may be created such that there is a link between a time (or times) when the theme should be activated and a particular song.


Certain exemplary embodiments may incorporate a camera into the jukebox. The camera may be used to create an artist or patron likeness, and the captured image may be treated and included into the user interface. For example, the jukebox's on-board camera may be used to convert the patron's likeness into an on-screen facsimile or caricature of the currently playing artist. The likeness may be caricatured using known automatic or manual techniques, and the modified or unmodified image may be used in a variety of ways. For example, the image may be used as background art, associated with a playlist, associated with a user avatar, incorporated into scripts used during music playback (e.g., to show who selected a song), provided for a mosaic (e.g., for a playlist, of “regulars” at a venue, etc.), and/or the like. The presence of a network connection along with a camera also may be advantageous. For example, it may be possible to have a “live look-in” at a location. A person contemplating whether to visit a bar or club could determine in advance whether it appears to be fun, exciting, “chill,” etc., in advance. A user could also see what is going on at a location while away from home. Live look-ins may be provided via an Internet connection and thus may be accessed through a webpage, a smart phone, or the like. The images also may be analyzed real-time or after the fact to gather demographic data about the location and/or jukebox patrons.


In a similar manner, the jukebox camera could be used a “mirror” to “reflect” what the jukebox sees is going on in the room, on the dance floor, etc. A feed may be provided to the jukebox display itself, remote terminals operably connected to the jukebox but still within the same venue, to suitably equipped TVs or other displays, etc. The reality may be “augmented,” e.g., by adding virtual elements to the mirror image that do not exist in reality. For instance, the image on-screen may be identical to a mirror except that album jackets may float onscreen in front of the user's “reflection.”


In certain exemplary embodiments, the camera may be used as an attract device. More particularly, in certain exemplary embodiments, the jukebox's camera may serve as the basis for a motion detection system that attracts patrons to the jukebox. The jukebox may attract the patron by, for example, increasing or decreasing the intensity of the jukebox's user interface lighting a the patron approaches; welcoming the patron when they are within a predetermined number of feet of the jukebox; playing a sound, such as a heartbeat, that lets patrons know they are getting “colder” or “warmer” as they approach; using edge detection to estimate the size of a group and addressing the audience appropriately via sound or textual display; and/or the like. The message also may be determined based on the type of music playing. Thus, if a group is detected and country music is playing, a message like “Hey, y'all! How about a song?” may be displayed rather than the more generic message, “You look you could use a song.” The main display or any secondary displayed (e.g., an LED array) may be used to display messages like these, e.g., for patron welcoming or other features, potentially before the patron has even physically touched the jukebox. It will be appreciated that the camera may be made to function as a proximity sensor in certain exemplary embodiments, and that certain exemplary embodiments also may additionally or alternatively include separate proximity sensors for like purposes.


It is believed that, on average, 2.2 people are in front of the jukebox at a time when a selection is being made. The camera may help determine that multiple people are in front of the jukebox and make a recommendation for the group as opposed to the individual. Doing so also may help resolve seeming ambiguities in selections, as different group members may have drastically different song requests that are not easily clusterable or classifiable so as to come up with a good recommendation. Thus, recognizing that a group of people are present may relax certain constraints, cause the jukebox to disregard seeming “outliers,” or simply not provide recommendations to a group.


It will be appreciated that the camera may detect lighting levels and adjust jukebox elements so as to take a good quality picture. This may in certain example instances involve adjusting display, LED, and/or rim lighting intensities in connection with a camera flash in a dark or partially lit environment, lowering lighting levels to reduce washout in a bright environment, etc. In certain exemplary embodiments, the users may select an effect such as, for example, washout, saturation, overexposure, etc., and the camera and/or lighting elements may be made to respond according to known photographic techniques for accomplishing the desired effect(s). Flashes may be timed so as to correspond with movement on the screens, e.g., to draw a user's gaze upwards, downwards, or in a particular direction, so that attention focuses on a camera taking a picture (e.g., when the jukebox is operating in photobooth mode). In certain exemplary embodiments, frames or backgrounds may be applied to pictures and in some cases, they may be sponsored (e.g., a Bacardi-sponsored frame may include a tropical feel, etc.) in order to generate more revenue for the site.


In certain exemplary embodiments, a large glass or other panel may be placed over a substantial portion of the jukebox. The panel in certain exemplary embodiments may be made to function as a touch screen panel in certain instances, thereby giving or at least simulating interactive functionality to jukebox elements that otherwise would not be user manipulable. Thus, the panel may serve as a proxy for communication with elements with which a user ordinarily could not interact. The jukebox software may communicate the position of the touched area and take a corresponding action. Fore example, it may be possible to emulate a touch screen LED by placing it beneath a transparent, veritable tactile screen, where the underlying LED array otherwise would not be user manipulable. Examples of functionality that may be imparted to otherwise static elements include, for instance, tracing a user's finger as the user spells out the user's name or other text (e.g., on an LED array), playing a classic “whack-a-mole” type game, detecting when a user is leaning on or otherwise touching the jukebox (which may cause the jukebox to display a message asking the user to avoid leaning on it, inviting the user to make a selection, etc.), and the like. The camera and/or proximity sensors on the jukebox may be used as a part of a security system for the venue in certain exemplary embodiments. That is, in certain exemplary embodiments, the jukebox's camera may be used to detect and/or report an unwanted present. The camera may be used to create an image or video of the unwanted present, if detected. For instance, after closing, the jukebox may be set to a security mode where it appears to be powered down but its camera continues to function. If an intrusion is detected, e.g., via motion detectors, proximity sensors, and/or the jukebox's camera, the camera may create an image or video of the unwanted presence. The jukebox's audio capabilities to communicate with the unwanted presence, e.g., by sounding an alarm, playing music as loud as possible, etc. The jukeboxes LEDs and/or main screen may communicate that the presence has been detected, and/or the main screen may display a captured or live image or video of the intruders. Furthermore, the jukebox may use its network connection, e.g., to email, text message, or otherwise send an alert to the jukebox owner and/or appropriate authorities to apprise the owner and/or the authorities of the unwanted present, initiate a voice or VoIP call to the jukebox owner and/or appropriate authorities, etc. Similar to the techniques described above, in a more “passive” mode, the camera may provide a live look-in mode for the owner and/or authorities to monitor the situation from a local or remote location. In certain exemplary embodiments, the camera on the jukebox may serve as and sometimes even take the place of a closed circuit monitoring system.


In certain example embodiments, a security mode may be entered into at a predefined time (e.g., after closing), upon an input from an authorized person (e.g., shutting down the jukebox for the night), etc. The security mode may in certain exemplary embodiments power down or otherwise deactivate the display screen so that the jukebox as a whole does not appear to be operation. However, the camera disposed on or otherwise connected to the jukebox may capture a stream of images and/or video. That stream may be related to a security subsystem operating in the jukebox network or a separate network. Authorized users may remotely log into the security subsystem (e.g., using a username/password combination, by inputting a unique address or other identifying information) to obtain live look-ins. The jukebox and/or security subsystem may monitor for abnormalities such as, for example, unexpected motion at the venue, etc. Upon the detection of a disturbance, the security subsystem may generate or relay a message to an authorized person. For example, a VoIP call may be initiated to the venue's proprietor, to law enforcement personnel, to a private security company, etc. The jukebox also may be made to sound an alarm, display flashing lights, etc., as a security measure against intrusion into the location.


It will be appreciated that many of the same streaming techniques may be used in connection with other, recreational live look-in techniques described herein. For instance, the jukebox network or other appropriate server system may obtain streaming information from the jukebox and enable users (e.g., registered users) to remotely view the venue, e.g., from an Internet browser. In certain example embodiments, the jukeboxes themselves may act as the servers.


In certain example embodiments, a jukebox device comprises a display; a camera; a non-transitory computer readable storage medium storing a plurality of instances of media available for playback on or via the jukebox device; and at least one processor configured to: obtain images and/or video from the camera, display said obtained images and/or video from the camera on the display of the jukebox device substantially in real time so that the display substantially mirrors what is detected by the camera, and enable a remote user to view said obtained images and/or video from the camera. The at least one processor may be further configured to cause said obtained images and/or video to be displayed on one or more external display devices provided in a common location with the jukebox device. Similarly, a web server may be configured to enable the remote user to view said obtained images and/or video from the camera.


The at least one processor may be further configured to enter into a security mode at a user-specified time or upon a signal from an authorized user. The at least one processor causes the camera to act as a motion detector when said security mode is entered. The at least one processor is configured to raise an alarm when motion is unexpectedly detected. The alarm may include initiation of a VoIP call to one or more specified parties; sending SMS, emails or other messages; causing the display and/or lighting on the jukebox to become illuminated and/or flash; an audible alarm; and/or the like.


Methods of making and/or using such jukebox devices also may be provided, as may non-transitory computer readable storage mediums tangibly storing instructions for using such jukebox devices/providing such user interfaces. Jukebox systems including plural jukebox devices, for example, also may be provided in different exemplary embodiments.


Certain exemplary embodiments also provide adaptive jukebox personalization features including, for example, personality generation and personality expression. Personality generation relates to the process of collecting various pieces of information and using them to characterize the personality of the jukebox. This may be an ongoing process. For example, each time a user interacts with the jukebox, new information is generated and may be collected.


Different types of information may be considered to define the personality of a jukebox. These factors may include some or all of the following and/or other factors: the jukebox's geographic location, the most popular songs played on a jukebox, the type of music played on a jukebox (i.e., genre, style, mood, etc.), plays made on the whole jukebox network, information as to whether additional weight should be given to a current patron's selections (e.g., if they are a member of MyTouchTunes or bar staff), etc.


The jukebox then may be assigned a personality type, e.g., based on the collected information. Personality types may be discrete in the sense that any given jukebox is assigned only one type at any given time. The personality types, and the criteria matching those types, may be defined by the jukebox provider. For example, as a basic implementation, the most popular music genre in a location may be used to define personality types. In such a case, jukeboxes may be referred to as “Rock type,” “R&B type,” etc. Personality generation may also involve adjustments and refinements to the jukebox personality type over time, e.g., as new information is collected. For instance, a jukebox initially be categorized as a “Rock type” jukebox, but may be refined over time to indicate a preference for “energetic” rock songs rather than “slow” rock songs or “rock ballads.”


Thus, it will be appreciated that information may be gathered, an initial assignment as to one of plural predefined categories may be made, and that adjustments or refinements to the initial assignment may be made over time, e.g., as more data is collected. It will be appreciated that the adjustments or refinements may be sub-classifications within a particular category, re-classification into a new category or sub-category, etc.


The personality expression of a jukebox may be understood as how the user experience is affected as a result of the jukebox's personality. For instance, the general appearance of the jukebox may be modified to match the jukebox personality. Such modifications may include some or all of the following and/or other characteristics: displaying information about the jukebox's location, a custom UI theme representing the personality of the jukebox, a lightshow pattern and coloration that reflects the personality of the jukebox, background image of an artist as a UI backdrop according to the location's personality (see artist portal above), etc.


The example techniques described above also may apply to jukebox adaptive music browsing and/or searching. For instance, when browsing music on the jukebox, emphasis may be placed on preferred music genres, styles, moods, etc., based on the jukebox personality. The same or a similar approach may be applied to music searching. That is, in certain example instances, more emphasis may be placed on preferred music types in search results, e.g., such that songs that match the personality are listed first or near the top of the results, or at least higher than they might otherwise be, etc.


Along similar lines, in certain exemplary embodiments, a jukebox search algorithm may provide incremental auto-complete suggestions to patrons with respect to the songs, artists, and albums most played on that jukebox, and/or that match the personality of the jukebox. The algorithm may draw from a variety sources that identify the preferences of the user such as, for example, the DMA, city, MyTouchTunes user profile, and possibly the Facebook user profile. Similar techniques may be applied as to the jukebox personality, e.g., for unregistered or not recognized patrons.


For example, as a patron enters the letters “Rol” into the search field, the auto-complete may recognize the user, and determine that the Latin artist Rolando should be listed higher on the auto-complete suggestions than the Rolling Stones, who are more popular across the jukebox network, but not necessarily more popular to the jukebox user or on the jukebox in use.


The notion of jukebox personality also may be used to send music content packages to jukeboxes according to their personalities in certain example implementations.


Personalization of the jukebox and/or the user also may be used for targeted advertisements. For instance, the personality of a jukebox may be used to help determine which ads it will receive or display based on some or all of the following and/or other factors: location specific events (e.g., Happy Hour, Ladies' Night, New Year's Party, etc.), local concerts (e.g., Arcade Fire, your favorite indie band is playing in New York this weekend), local events (e.g., local fund-raising, blood collections at the local community center, etc.), and/or the like.


Furthermore, contextual advertisements may be based on the personality of the jukebox and/or the jukebox user. For recognized users who have logged in, the personality may be known. However, for anonymous users, advertisements may be displayed based on a more current browsing behavior and song selection (e.g., during a current selection). For instance, as the user browses the alternative rock genre on the jukebox, the jukebox may display advertisements for an upcoming Lollapalooza event.


Further personalization may be possible in connection with contextual advertisements by taking a picture at the time of the user's interaction (or from an archive if the user is a recognized user) and then incorporating the user's likeness into the advertisement. For example, if the user is browsing Jimmy Buffett music, a picture of the user may be taken and incorporated into a picture with the user on the beach and an advertisement for cheap travel to the islands.


Certain exemplary embodiments may enable a user to create set lists for play. For example, a user may browse the catalog of songs, etc., and select songs for potential playback on the jukebox. The user may reorder songs, delete songs, add songs, etc., to the working set list during a session, much like a person may place items in a shopping cart and remove some other items. Once the set list is defined by the user, it may be submitted for play in whole or in part as one or more packages. Similarly, the user may make some or all of the songs in the set list play immediately for a premium. In that regard, the user may simply drag and drop the set list, or a portion thereof, into a predefined area of the user interface designated for causing immediate play of the song(s). Thus, it will be appreciated that the jukebox user interface may be configured to allow a user to create of a group of songs that are then queued for play by the user in a single action, and an advanced set list may enable the user to modify the set list's contents before it is executed (e.g., such that songs may be added, removed, reordered, given a higher priority via the play next or jump the queue function, etc.). This set list may be executed using a user interface element or a physical activator on the jukebox The set list may be advantageous in certain exemplary embodiments because it does not require a user to be logged in to the jukebox. In other words, it will be appreciated that the set list techniques described herein may be particularly advantageous for ad hoc creation of a list of songs for playback by an anonymous user. It will be appreciated that the user may be able to save the set list as a playlist for subsequent use, e.g., by registering with the jukebox and/or logging in to an already existing account.



FIG. 24 is an example screenshot of an improved music discovery user interface in accordance with certain exemplary embodiments. As shown in FIG. 24, the elements in a collection of media are displayed in the fan-like arrangement 2402 in the approximate center of the screen. The collection may be, for example, selections from a genre, a playlist, etc.


The collection may include a predetermined number of elements. The elements in the selection may be user determined (e.g., in the case of a playlist), determined by the provider of the jukebox (e.g., in the case of a genre), a proprietor of the location (e.g., for a theme night or for the location in general), etc. In the FIG. 24 example, genres 2404 are listed in the arc at the bottom of the screen, and the “pop” genre is selected.


Once a collection is determined, a collection type may be further specified. There are three buttons 2406 on the FIG. 24 example screen for specifying the collection type. These options include artists, albums, and songs. Thus, the elements in the collection may be grouped or organized according to the collection type. In the FIG. 24 example, the collection type is “songs.” Accordingly, pop music selections are provided in the fan-like arrangement at the approximate center of the screen. All pop music selections in the collection are listed separately as songs because that is the specified collection type. If the albums collection type were specified, the albums containing the songs in the collection would be displayed. Similarly, if the artists collection type were specified, the artists who perform songs in the collection would be displayed. In the latter cases, a user could browse by album or artist, make a selection, and then further select a song from a selected album or artist. Regardless of how the elements are grouped by collection type, the same songs are represented, just in different ways.


Once a collection type is specified, the items in the collection may be ordered, e.g., using an ordering indicator 2408. For instance, three illustrative ordering buttons are provided below the fan-like arrangement of songs 2402. The “A-Z” button organizes the elements in alphabetical order, the star button organizes the elements in terms of popularity, and the calendar button provides a time-based arrangement of the elements (e.g., by release date). The selection of a particular order will, in turn, cause a “scrubber bar” 2440 to appear. In the FIG. 24 example, the order selected is alphabetical order and, thus, the scrubber bar shows the letters of the alphabet. A user may quickly navigate through the collection using the scrubber bar (and, in particular, in the FIG. 24 case, by selecting a letter). The selection of a letter may cause it to glow or become larger for emphasis. It will be appreciated that the elements shown in the scrubber may change based on the order selected. For popularity, a series of numbers ranging from 1 to the number of elements in the selection may be provided, and a series of dates, years, or the like may be provided if the calendar order is selected. It will be appreciated that the selection of an artist may not allow the calendar order to selected in certain example instances, as an artist typically is not associated with a year the same way an album or song has a release date. However, in certain cases, the date of an artist's first single or first album may be used as a relevant date.


Thus, in certain exemplary embodiments, a jukebox device comprises a display; a non-transitory computer readable storage medium storing a plurality of instances of media available for playback on or via the jukebox device; and at least one processor configured to cause a user interface to be displayed on the display and further configured to respond to inputs to the user interface. The user interface includes: at least one first user interface element configured to enable a user to select a collection of instances of media available for playback from a plurality of possible collections of instances of media each said collection representing a different partial subset of media available for playback on or via the jukebox device, at least one second user interface element configured to enable the user to select one of plural possible grouping modes for grouping the instances of media in the collection, the grouping modes including artist, album, and song groupings, wherein the selected grouping mode determines items to be presented to the user for possible selection, at least one third user interface element configured to enable the user to select one of plural possible ordering modes for ordering the items to be presented to the user for possible selection, and a display area for displaying at least some of the items to be presented to the user for possible selection, the items being ordered within the display area according to the selected ordering mode. The at least one first user interface element, the at least one second user interface element, and the at least one third user interface element are all displayable on the display at the same time. The collections may include predefined genres and/or musical themes, playlists, and/or the like in different implementations.


The ordering modes may include, for example, an alphabetical order mode, a popularity-based order mode, and a chronological or date order mode. Each item to be presented to the user for possible selection may be assigned a popularity value, with the popularity-based order mode ordering the items to be presented to the user for possible selection based on the assigned popularity values. The popularity values may be assigned based on one or more of: item popularity across a jukebox network, item popularity at a venue at which the jukebox device is located, item popularity on an industry chart, and/or item popularity at one or more social networking sites. The chronological or date order mode may in some cases be disabled when the artist grouping mode is selected.


The user interface may in certain exemplary embodiments further display at least one fourth user interface element, with the at least one fourth user interface element being adaptable based on the selected ordering mode. The at least one fourth user interface element may include letters when the alphabetical order mode is selected, numbers or lines corresponding to popularity values when the popularity-based order mode is selected, a date or calendar display when the chronological or date order mode is selected, etc. In certain implementations, the display area may include a partial subset of the items to be presented to the user for possible selection, with the partial subset being determined based on a selection made with respect to the at least one fourth user interface element.


A partial subset of items may be displayable in the display area in a fan-like arrangement across a general center area of the display, where the exact items displayed are changeable based on user manipulation of the at least one fourth user interface element and/or user input to the display area. In certain exemplary embodiments, a user selection of a displayed album item causes the display of songs within the selected collection and associated with the selected album item to be displayed in the display area, and/or a user selection of a displayed artist item causes the display of songs within the selected collection and associated with the selected artist item to be displayed in the display area. In certain exemplary embodiments, a user selection of a displayed song item causes a popup screen to be displayed, with the popup screen enabling the user to confirm that the selected song is to be played back. The display of the popup screen may be accompanied by the blurring, fading, and/or reduction to grayscale of user interface elements behind the popup screen.


Methods of making and/or using such jukebox devices also may be provided, as may non-transitory computer readable storage mediums tangibly storing instructions for using such jukebox devices/providing such user interfaces. Jukebox systems including plural jukebox devices, for example, also may be provided in different exemplary embodiments.


Many jukeboxes, kiosks, gaming devices, and the like have used exterior lights to attract and/or emphasize various events. These devices typically issue direct commands to a color-processing device. Unfortunately, however, this control technique oftentimes makes it difficult to tailor the overall experience (e.g., acoustic, user interface, electromechanical, environmental, and peripheral) to changing or event-related situations, e.g., to adopt a different appearance in a coordinated fashion. More particularly, what has been absent is the ability to contextually alter these behaviors in a way that allows the device and its user interface to collaborate as an integrated system. A solution to this problem is to create a control system that couples the behavior of visual peripherals and displays to a theme. The theme thus in certain exemplary embodiments would help describe the standard behavior for the user interface and the peripheral lights.


The exemplary embodiments described herein may be used in connection with a jukebox of the type shown and described in, for example, U.S. application Ser. No. 29/371,255, the entire contents of which are hereby incorporated herein by reference. FIGS. 25a-b show components of a jukebox device in accordance with certain exemplary embodiments, and FIG. 26 is an image of an example jukebox device in accordance with certain example embodiments. As can be seen from FIG. 25a, a backwash light segment (1) is provided, along with a rim light segment (2). An LED display (3) is located above a payment mechanism attract light (4), the latter of which helps serve as a main display screen and main portion of the user interface. A near field communication attract light (5) may be provided. A user may provide payment via a payment acceptor that is proximate to a payment mechanism attract light (6). A main activation switch light (7) also may be provided around a main activation switch. FIG. 25b is a close-up of the LED display (3) shown in FIG. 25a, in accordance with certain exemplary embodiments. As can be seen from FIG. 25b, the LED display (3) may comprise a portion of an array of individual LED elements (9). At least the LED display (3) and the main display screen may be covered with a common touch-sensitive panel in certain exemplary embodiments.


As shown in FIG. 26, the jukebox device 2600 includes a main display screen 2602 that is configured to display a main user interface, e.g., under the control of at least one processor built into the jukebox device 2600. A secondary display 2604 (which may, in certain example embodiments comprise an LED array) also is provided above the main display screen 2602. In certain example embodiments, the secondary display 2604 may be configured to display FFT-like data, welcome messages, kinetic or static text, etc. In certain example embodiments, the main display 2602 and/or the secondary display 2604 may be fully or partially covered with a glass or other touch-sensitive substrate 2606. This touch-sensitive substrate 2606 may provide interoperability to otherwise static elements such as, for example, the secondary display 2604. In certain example embodiments, the main display screen 2602 may be a first touch screen display itself, and a touch-sensitive may be provided over only the secondary display 2604. Other arrangements are, of course, possible in different embodiments.


A physical interaction button 2608 may be provided for taking an action designated by the user interface. In certain exemplary embodiments, lighting elements may be disposed around the button 2608. A decorate “play” or other symbol may be placed on the button 2608 to improve its aesthetic appearance.


One or more lighting elements 2610 may be provided around the periphery of the jukebox device 2600. The one or more lighting elements 2610 may be operable as a single element, as segments, etc., in different embodiments. In certain exemplary embodiments, the one or more lighting elements 2610 may be provided, constructed, and/or controlled so as to simulate more classic or vintage-looking neon lights. Although not shown, one or more light elements may be disposed at or pointing towards the rear of the jukebox device, e.g., so as to provide for a “wall wash” or back light. A frame 2612 also may be provided around the outer periphery of the jukebox device 2600, e.g., so as to give it a completed or “seamless” look and feel.


One or more payment acceptors may be provided. As shown in the FIG. 26 example, a credit card acceptor 2614 is provided, as are first and second bill acceptors 2616 and 2618. Coin acceptors may be provided in certain example embodiments, as may RF-ID readers. A common color palette and themed behavior type may be applied to some or all of these visual assets so that they are coordinated and attractive rather than harsh and random. For example, rear facing lights (1), a front facing light rim (2) the text or graphics on LED display (3) and all usage of accent color (12) occurring within the User Interface could be set to the same color attribute. Then, this attribute could change according to a set periodicity and by using a color range or a sequence of specific colors (defined by RGB, pantone or other color description method). The effect of this would be a highly coordinated visual appearance.


A common script or data set may be provided for synchronizing the display elements, e.g., in terms of color or colors, periodicity for change, synchronization rules, etc. It will be appreciated that by substituting a new data set, the complete overall appearance of the device and the rendered user interface may be changed without altering any of the software or hardware. This would, in turn, allow the same system to be used in diverse commercial and/or residential establishments, and further provide the ability to adapt the established design and often trademarked color usage that identifies the brand.


Such techniques also may allow independent locations that have invested in specific lighting and design to tailor the color, intensity, and pace of all lighting so that it would fit with their established decor, all the while benefiting from the manual or automatic coordination of the on-screen user interface that would appear to have been customer colored for this location. In certain exemplary embodiments, the system may be set to use a preplanned color palette but to apply this palette to different rhythms, e.g., either the actual beat and/or frequency of the song being played at this instant, or a set pace that was established for the location. The combination of this color and pattern may help constitute the location's lighting theme. The customization may be enhanced with a set of images that is gathered into an image collection such as “Country” or Urban,” and these images may have associated metadata that describes a lighting theme.


As the images are presented on the system's embedded screen, or on remote screens that are provided with content from the system, the lighting may follow the entrance and exit animations of the images. This lighting rhythm may be “with slideshow” and the light color, luminosity, and rhythm that is to be followed as the image may be presented based on metadata associated with each image, e.g., as in a matte for a print, the lighting color may enhance or complement certain colors within the image. This advantageously may help provide the ability to extend the visual impact of an image beyond its actual size, e.g., by projecting a complementary color or colors at the same time.


In certain exemplary embodiments, the coordination of the presentation of an image and the lighting attributes may involve the presenting of a still or full-motion advertisement on the local or remote screen. The brand colors or signature contrasts could be presented temporarily, e.g., in accordance with the display of the advertisement. The pattern may be overridden based on a defined event. For example, if a music playing event occurs, then the light pattern on a device may adopt the rhythm and color choices to provide accompanying song visualization. In other words, in certain exemplary embodiments, priorities may be defined, and the advertisement coloration effects may be made secondary to music related events, etc.


Each colorization pattern may be described in terms of “key frames” in certain exemplary embodiments. For example, for each channel, there may be multiple key frames, with each one being defined by the end value for the color of the channel to be computed starting from the current color, to a specific point in time, using a given easing mode. The point in time may be defined using a color value representing the progress in the interval of [0, 1]. The easing mode may be a mathematical function that translates the time into a dimensional progress factor used when going from key frame to key frame.


Below is an illustrative script that defines a lightshow pattern in accordance with certain exemplary embodiments. More particularly, the example script below defines a lightshow pattern with a duration of 10 seconds and operating on the color of the 6 channels that constitute the wash wall. For the first channel, the first key frame will move the color from black to blue in the first 10 percent of the duration of the pattern. The second will move the color from blue to green from 10 percent to 20 percent of the duration. The third will move the color from green to red from the following 20 percent of the duration to the 30 percent of the duration, and the last key frame will move the color from red to black from 30 to 40 percent of the duration.


For the second channel, the first key frame will move the color from black to blue from 10 to 20 percent of the duration of the pattern. The second will move the color from blue to green from 20 to 30 percent of the duration. The third will move the color from green to red from the following 30 percent of the duration to the 40 percent of the duration, and last key frame will move the color from red to black from 40 to 50 percent of the duration, and so forth.














[


 /* channels declaration */


  [


   {


    “type” : “TTChannel”,


    “id” : “channel-0”,


    “index” :0


   },


   {


    “type” : “TTChannel”,


    “id” : “channel-1”,


    “index” : 1


   },


   {


    “type” : “TTChannel”,


    “id” : “channel-2”,


     “index” : 2


   },


   {


    “type” : “TTChannel”,


    “id” : “channel-3”,


    “index” : 3


   },


   {


    “type” : “TTChannel”,


    “id” : “channel-4”,


    “index” : 4


   },


   {


    “type” : “TTChannel”,


    “id” : “channel-5”,


    “index” : 5


   },


   {


    “type” : “TTChannel”,


    “id” : “channel-6”,


    “index” : 6


   }


  ],


/* pattern declaration */


  [


   {


    “type” : “TTAnimator”,


    “id” : “animator-default”,


    “duration” : 10000, /* duration in milliseconds


*/


    “is-default” : true, /* if true, the pattern


will loop indefinitely */


    “properties” :


    [


      {


        “object” : “channel-0”,


         “name” : “rgb”,


         “ease-in” : false,


         “interpolation” : “linear”,


         “keys” :


        [


          [ 0.10, “linear”, “blue” ],


          [ 0.20, “linear”, “green” ],


          [ 0.30, “linear”, “red” ],


          [ 0.40, “linear”, “black”]


       ]


      },


      {


        “object” : “channel-1”,


        “name” : “rgb”,


        “ease-in” : false,


        “interpolation” : “linear”,


        “keys” :


        [


          [ 0.20, “linear”, “blue” ],


          [ 0.30, “linear”, “green” ],


          [ 0.40, “linear”, “red” ],


          [ 0.50, “linear”, “black”]


        ]


     },


     {


        “object” : “channel-2”,


        “name” : “rgb”,


        “ease-in” : false,


        “interpolation” : “linear”,


        “keys” :


        [


          [ 0.30, “linear”, “blue” ],


          [ 0.40, “linear”, “green” ],


          [ 0.50, “linear”, “red” ],


          [ 0.60, “linear”, “black”]


       ]


     },


     {


        “object” : “channel-3”,


        “name” : “rgb”,


        “ease-in” : false,


        “interpolation” : “linear”,


        “keys” :


        [


         [ 0.40, “linear”, “blue” ],


         [ 0.50, “linear”, “green” ],


         [ 0.60, “linear”, “red” ],


         [ 0.70, “linear”, “black”]


        ]


     },


     {


        “object” : “channel-4”,


        “name” : “rgb”,


        “ease-in” : false,


        “interpolation” : “linear”,


        “keys” :


       [


         [ 0.50, “linear”, “blue” ],


         [ 0.60, “linear”, “green” ],


         [ 0.70, “linear”, “red” ],


         [ 0.80, “linear”, “black”]


       ]


     },


    {


      “object” : “channel-5”,


      “name” : “rgb”,


      “ease-in” : false,


      “interpolation” : “linear”,


      “keys” :


      [


        [ 0.60, “linear”, “blue” ],


        [ 0.70, “linear”, “green” ],


        [ 0.80, “linear”, “red” ],


        [ 0.80, “linear”, “black”]


      ]


    },


   ]


  }


 ]


]










FIG. 27 is an illustrative view of visual components on a jukebox device being color matched and synchronized in accordance with certain exemplary embodiments. In the FIG. 27 example, an advertisement for Verizon is being shown. In that regard, the LED display says Verizon, and the Verizon logo appears in the main display area. In addition, the LED display text is red, as is a portion of the main display. The wall wash also is red, and the rim light segment is white for contrast. The red Verizon theme may be applied during song play, as shown in FIG. 27 in which the Black Eyed Peas song “Boom Boom Pow” is being played. In certain exemplary embodiments, a custom message may be designated for the song, e.g., to display “BOOM BOOM BOOM!!!” in the LED display when the artists sing the same. In such a case, precedence may be given to the song-specific text. However, the red Verizon theme may be maintained, such that the text is displayed in that color pattern.


In certain example implementations, the coloration might not be exact as between different display elements, even when the same colors are specified. For instance, the LEDs in the array may have a certain CRI or other value influencing the color produced, whereas the main display may have a second CRI or other value influencing the color produced. The LEDs in the LED array may be different still from the rim light segment, etc. The wall wash may be influenced by the paint on the wall, ambient light, etc.


To provide a consistent coloration, the system may specify different colors for output such that they seem to be the same to a viewer. The differences may be predetermined in some respects, e.g., for the LED array, the main display area, and the rim lights. However, it may not be possible to always predict and pre-compensate for location-specific factors that may influence the ultimate coloration. In such cases, the camera may be used to help determine the characteristics of the environment. Data from the camera may be analyzed and/or used to adjust the coloration of one or more elements so that a consistent coloration is possible. A software-based color filtering module may be provided in certain exemplary embodiments for helping to collect the data, calculate offset values (e.g., in accordance with a mathematical formula), and providing instructions or color offset codes for the operating system to consider when indicating what colors are to be output for each of the display elements.


In certain exemplary embodiments, a jukebox device is provided. It may include, for example, at least one display device; a generally elongate array of LED elements provided above the at least one display device; a backwash light arranged at a rear portion of the jukebox device and configured to generate light behind the jukebox device; a rim light disposed around a periphery of the jukebox device; a substantially circular activation switch or button; a non-transitory computer readable storage medium storing a plurality of instances of media available for playback on or via the jukebox device; and at least one processor configured to provide a user interface and further configured to coordinate lighting events based at least in part on events associated with the user interface. At least one payment collector having an associated payment mechanism attract light provided in close relative proximity thereto also may be provided.


The at least one processor may coordinate lighting events in dependence on a script (e.g., an XML-based script), and the script may include indications as to what lights are to be activated, when the lights are to be activated, and what color is to be displayed. At least some instances of media may have themes associated therewith, with the themes being recognizable by the at least one processor such that, when played, the at least one processor creates lighting events in connection with the corresponding theme. Similarly, the jukebox device may be configured to display advertisements, with at least some of the advertisements having themes associated therewith, and with the themes being recognizable by the at least one processor such that, when displayed, the at least one processor creates lighting events in connection with the corresponding theme.


A single touch sensitive substrate may be disposed over both the at least one display device and the array of LED elements. The at least one processor may be configured to monitor for inputs to the touch sensitive substrate and selectively activate LED elements in the array of LED elements in response to a detected touch to an overlying area.


The at least one processor may be configured to coordinate lighting events having consistent colors produced by different light sources by offsetting the color values to be generated to account for differences in the light generated by the light sources and/or in response to the surrounding environment. In certain exemplary embodiments, the jukebox device may further comprise a camera. The at least one processor may be configured to use data from the camera to determine ambient conditions.


Methods of making and/or using such jukebox devices also may be provided, as may non-transitory computer readable storage mediums tangibly storing instructions for using such jukebox devices/providing such user interfaces. Jukebox systems including plural jukebox devices, for example, also may be provided in different exemplary embodiments.


In certain exemplary embodiments, karaoke features may be provided. Karaoke features may be enabled, for example, by linking songs to lyrical scripts. Such scripts may have lyrics associated with times in a song and optionally may further include animation information (e.g., kinetic text) for highlighting text as it is to be sung, removing text once it has been sung, introducing new text that it about to be sung, etc.


In certain exemplary embodiments, the jukebox may receive live streams. For instance, in certain exemplary embodiments, a DJ may simultaneously stream music to multiple jukeboxes at multiple locations. Similarly, a DJ in a location may tie into a venue's jukebox system and turn some or all jukeboxes in the venue into a speaker system.



FIG. 28 is an example screenshot enabling a user to access a plurality of different features of a combined karaoke/photobooth jukebox in accordance with certain exemplary embodiments. As above, the jukebox 2800 in the FIG. 28 example includes an LED array 2802 (which, in the illustration, shows output from an FFT or the like) and a main display 2804 (which may be a touch screen display). The main display 2804 includes a substantial part of the user interface elements with which the user interacts and has a “skin” applied thereto for the artist “Pink”, e.g., as indicated by the optional skin label 2806. The skin in this example includes a background image and a predefined coloration of certain control elements of the jukebox. An indication of the venue's name is displayed in area 2808 and, in this example, is “Karaoke Club 850.” The song that is now playing is displayed in area 2810, along with optional graphical elements related to the same, and a number of available credits is shown in credit meter area 2812. In this example, album art is displayed along with text in area, and -0- credits are available.


If a user wishes to play a song, a user may select any of the control elements displayed in the fan-like array 2814 that is the general center of display 2804. Control elements are provided for browsing songs in connection with lists of predefined genres, local favorites (which may be developed based on actual jukebox usage at one or more locations, specified by a authorized person such as a routeman or location owner), new and popular songs (e.g., as determined based on the new media distributed to the venue or the master catalog of songs; based on plays at one or more locations such as all locations, all locations owned by a particular person; all locations served by a particular operator; in accordance with an industry chart; etc.), in accordance with playlists defined at the site and/or made publically available by registered jukebox patrons, etc. A user also may search for songs locally and/or from the master catalog, e.g., using the appropriate control element. A user may switch between karaoke jukebox mode, normal jukebox mode, and photobooth mode, e.g., by selecting a corresponding control element from the mode array 2816. Ad banners 2818 may also be displayed in certain exemplary embodiments. Example photobooth user interface features are described in U.S. application Ser. No. 13/621,922, filed Sep. 18, 2012, the entire contents of which are hereby incorporated herein by reference. In certain exemplary embodiments, operation in photobooth mode may be limited to a predefined duration (e.g., 2 minutes, 5 minutes, etc.) and/or number of pictures or picture cycles (e.g., 3 pictures or picture cycles, etc.), e.g., to help avoid a situation where the jukebox is dominated and cannot be directly used for normal jukebox and/or karaoke jukebox related operations.



FIG. 29 is an example screenshot enabling a user to access karaoke jukebox features in accordance with certain exemplary embodiments. The FIG. 29 example screenshot has a visual appearance that is similar to the FIG. 28 example screenshot and incorporates some of the same indicators. For instance, the credit meter 2812 and venue label 2808 are provided in FIG. 29. A home link 2902 back to the main selection screen (e.g., shown in FIG. 28) is provided, as is a social network link 2904. See, for example, U.S. Publication No. 2012/0158531, the entire contents of which are hereby incorporated by reference. A fan-like array of karaoke-related options 2900 similar to the fan0like array of 2814 is provided. The user, by selecting an appropriate icon, can search for karaoke songs by artist, by song, through a master search (e.g., that looks at the combined collection of information concerning artist name, album name, song name, lyrics, and/or the like), or from songbooks. A user may also play karaoke-related games. These various sub-modes are described in greater detail below.



FIG. 30 is an example artist selection screen for a karaoke jukebox in accordance with certain exemplary embodiments. The FIG. 30 example screenshot includes a browsable list of songs arranged by the artist. A user may select a lyrics button next to an artist/song listing, e.g., to verify that the song is what is expected. This may be desirable in some instances, because there are many songs that have the same or similar names that are in fact quite different from one another, and a would-be karaoke performer may not always able to sort out in advance whether the correct song is being selected. However, by browsing the lyrics, a would-be karaoke performer may be able to verify that the correct song is being selected. This is not really possible using conventional, printed-out books that do not include lyrics.


A user may move throughout the list by selecting a number of different control elements. For example, left and right arrows 3002a and 3002b may be used to advance the list one “page” at a time. For instance, when the master karaoke song list from which a selection may be made includes a number of songs greater than the number of songs that can be shown on a display, subsets can be organized into pages. Pressing the left and right arrows 3002a and 3002b may turn one “page” at a time to display more and more songs, similar to a conventional printed-out songbook.


An example page-turning action is shown in connection with FIG. 31. It will be appreciated that this animation may look and feel mechanical or automatic to convey to the user that the page is being turned by the jukebox device itself. As the page turns, listings on the front and back 3102a and 3102b of the page being turned may be displayable, and the page being revealed 3104 may be shown and the old page 3106 is being covered. In certain examples, the pages may be transparent, semi-transparent, or opaque when they are being turned.


A user may also press down on a page to achieve a similar action, e.g., as shown in connection with FIG. 32. In contrast with the more automatic feel provided in connection with the FIG. 31 example, this sort of finger-dragging movement may be more organic and may respond more closely to the user's actual movements, just as a person turning a page in a book may grab a corner and move it across the book a number of different ways. Other aspects of the display in FIG. 32 associated with the finger-dragging movement may be similar to those discussed above in connection with FIG. 31, e.g., in terms of text shown on the front and back of a page being turned, etc.


Referring once again to FIG. 30, similar to as described above, a scrubber bar may be used to advance through a list quickly, e.g., by updating the display to directly reflect the start of listings of a particular entry, by scanning across pages as users drag their fingers, etc. Page turning animations may be applied such that, for example, directly selecting a letter shows one turn with a number of pages being turned, scanning across pages reflects leafing through the corners of pages and/or rapid turning of pages, etc. In certain example embodiments, if a user selects a letter, the display may be advanced such that entries beginning with that letter are first. This may, for example, cause entries that ordinarily would not appear at the start of a page to be adjusted accordingly, which is not possible with static books. In other instances, when a letter is selected, the page may be advanced to the page including the first entry with that letter, without the formatting of the pages being adjusted, similar to a normal book.


The ordering indicators 3006 are similar to those described above in connection with ordering indicator 2408. Thus, similar to the above, once a selection is made using the main karaoke screen shown in FIG. 29, for example, the items in the collection may be ordered. As above, three illustrative ordering buttons are provided below the list of songs, with the “A-Z” button organizing the elements in alphabetical order, the star button organizing the elements in terms of popularity, and the calendar button providing a time-based arrangement of the elements (e.g., by release date).


A user may change the collection, and cause the display to be updated, by selecting the artist button 3008a, the songs button 3008b, or the songbooks button 3008c. A user also may turn to karaoke-related games by pressing the games button 3010, and also may institute a search using the search button 3012. Thus, it will be appreciated that the options shown in the fan-like array 2900 in FIG. 29 may be accessible and selected from at least some of the specific sub-mode displays.


When a user selects a song from the list shown in FIG. 30, for example, a detailed song selection screen may be displayed. FIG. 33 is an example screenshot of a detailed karaoke song selection screen in accordance with certain exemplary embodiments. As shown in FIG. 33, the song selection screen 3300 may specify the name of the song and the style in which the song is sung (e.g., the name of the artist that performed the original version of the song). The key in which the song is performed, as well as the duration of the song, may also be displayed. A user may press a button to initiate a query searching for other songs by the artist, and the resulting list may be shown in a new window, an updated version of the main list shown in FIG. 30, or in some other way. The user may also obtain song lyrics for the song, e.g., for reasons similar to those provided above. If a user requests to see the lyrics, they may be shown in a dedicated area 3302 with a scroll bar or other tool for the user's perusal, e.g., as shown in FIG. 34. If a user wishes to sing a song, the select song button 3304 may be depressed. If there are insufficient credits on the jukebox, the user may be prompted to insert additional payment, and payment may be accepted through any suitable payment means (e.g., coins, credit card, automatic debit from a registered account, etc.).



FIG. 35 is an example karaoke user identification screen in accordance with certain exemplary embodiments. This screen 3500 may be displayed after the select song button 3304 from FIG. 34 is pressed. The title bar 3502 of the screen 3500 confirms the name of the selected song and the artist in whose style the song was recorded. The title bar 3502 also enables the user to see the original key the song was recorded in, and use arrows to adjust the key the song is performed in, e.g., in musical half-step increments. Thus, karaoke performers who are familiar with their “natural keys” can have the jukebox device essentially transpose the background music, supporting vocals, and/or the like, to make the song easier to perform. In certain exemplary embodiments, the background instrumentation may not be transposed whereas the supporting vocals may be transposed, or vice versa. Example techniques for finding one's natural key are set forth in U.S. application Ser. No. 13/621,922, filed Sep. 18, 2012, the entire contents of which are hereby incorporated herein by reference. U.S. Publication No. 2012/0158531, the entire contents of which are hereby incorporated by reference, also discloses social network integration techniques that may be used in connection with certain exemplary embodiments.


Below the title bar 3502 of the screen 3500, the karaoke performer is able to identify him/herself. The FIG. 35 example enables this to be done in one of several ways. For non-registered jukebox users, performers can simply enter “stage names” in the area 3504, e.g., using the virtual keyboard 3506 or the like. As another option, registered jukebox users can log into the jukebox or jukebox system by pressing the login button 3508. Non-registered jukebox users who wish to become registered jukebox users optionally can sign up, e.g., by completing a full or abbreviated form directly on the jukebox or a suitable mobile device. If a registered jukebox user is already logged into the jukebox, an avatar 3510 of the user may be displayed.


In certain exemplary embodiments, a jukebox user may be registered through the myTouchTunes social network. In certain exemplary embodiments, users may log in to the jukeboxes by having the jukeboxes communicate with other social network sites that are maintained by companies other than the jukebox provider (such as, for example, Facebook, Twitter, Google+, etc.). The jukebox device may communicate with an external social network over a network and through a suitably configured interface, e.g., after selecting which social network should be used as the credentialing service. In certain exemplary embodiments, users may be recognized by providing credentials to non-social network services, e.g., to email or web providers (such as Yahoo, MSN, or the like). In certain exemplary embodiments, a user may use the camera provided on the jukebox to take a picture to be used as an avatar image if an avatar image is not already set, if a guest wishes to create a temporary avatar that may or may not be persisted for and presented in connection with future performances (e.g., the next performance, performances within a night, performances within a predefined time period such as a week or so, etc.). Example social network related interface techniques are described in for example, U.S. application Ser. No. 13/621,922, filed Sep. 18, 2012, the entire contents of which are hereby incorporated herein by reference.


A user may add a karaoke request to a karaoke queue by pressing the perform button 3512. It will be appreciated that the perform button 3512 may in some ways resemble a physical hardware button of the jukebox, thereby providing a point of similarity that ties together a user's hardware and software interactions, e.g., to create a more holistic and unique user experience. It will be appreciated that in general, the large physical button may be used to complete a predefined action (such as, for example, the triggering of a play, the selection of a song to perform, the initiation of a search based on entered criteria, etc.). In any event, at this point, a suitable number of credits may be deducted from the credit counter. Once the song is added to the karaoke queue, the example song confirmation screen in FIG. 36 may be displayed. This example screen may in certain exemplary embodiments indicate the user's position in queue, e.g., as an absolute number (such as first, second, etc.), an expected or approximate time (with text such as “be ready to sing in about three minutes”, etc.), and/or the like. In other cases, the user's position may not be displayed at all, e.g., if the user's position is so far in the future that the user might be discouraged (e.g., if the user's position in the queue meets or exceeds a predetermined number, if the user's performance is expected to occur more than a predetermined number of minutes in the future), be inclined to cancel the performance, at least temporarily lose interest in the karaoke event, etc.


The user may also have the option to add more songs at this time, or indicate that selections are done. In the former case, the FIG. 36 example screen may be closed and the underlying songbook may be displayed. In the latter case, the user may be logged out and the jukebox may be placed back in a karaoke flight or attract mode.



FIG. 37 is an example karaoke attract mode screen that may be displayed in accordance with certain exemplary embodiments, and FIG. 38 is another example karaoke attract mode screen that may be displayed in accordance with certain exemplary embodiments. These example attract mode screens may be displayed on the jukebox device and/or an external device that is connected to, or a part of a jukebox system in a location. For instance, as will be appreciated from the FIG. 37-38 examples, the displays may be external monitors or the like that may be remote from the jukebox itself. This may be advantageous in certain exemplary embodiments, e.g., because the jukebox may be disposed in a corner or away from most of the “action” in a bar or other location, whereas a centrally located display or multiple easily visible displays may present messages to a potentially broader audience.


In certain exemplary embodiments, one or more external displays may be selectively controllable by the karaoke jukebox. For instance, regular television or other programming may be provided by the displays, e.g., until the jukebox is put into karaoke mode. At that time, the jukebox may take control of some or all of the displays in order to show karaoke-related material. Example karaoke-related material may include, for example, the above-described and/or other attract loops, lyrics, feedback, the results of karaoke-related game (e.g., in connection with the example karaoke-related games discussed herein and/or other games), etc. The external display(s) may in addition or in the alternative display who is “up next” in terms of the performer or a stage name for the performer, an actual or approximate wait time until the next song or next several songs are to be performed, what will be sung next and/or for the next several songs, the key the songs will be sung in, how long the songs are scheduled to last, etc. It will be appreciated that, in general, anything that can be displayed on the karaoke jukebox display(s) can also be displayed on one or more external displays controlled by the karaoke jukebox.


Although the FIG. 37 example is somewhat generic, the FIG. 38 example includes at least some information about the karaoke song queue. For instance, it will be appreciated that the FIG. 38 example shows the song and performer who performed last together with an optional rating of the performance. With respect to the optional rating, certain exemplary embodiments may encourage audience members, friends, a KJ, and/or others to provide encouragement, discouragement, or the like. For instance, in certain exemplary embodiments, a registered and/or non-registered users may use suitably configured mobile devices to check in to a location and give thumbs up and/or thumbs down ratings, “like” and/or “dislike” indications, specific messages or “props” selected from a preset list and/or customized by the sender, etc. The “results” may be displayed during and/or after a performance and tabulated, e.g., in connection with a possible game in which different performers compete against one another for the highest or best rating, etc. The FIG. 38 example also indicates what song is next, who is to perform, how long before the next performance begins, the key the next song is in, and the length of the song. Some or all elements in the karaoke song queue may be displayed, e.g., if there are any. Regardless of whether the queue is empty, in certain exemplary embodiments, a message encouraging users to sign up for a performance may be shown. In certain exemplary embodiments, the message may be tailored to situations where the next song is the last song remaining in the queue for performance.



FIG. 39 is an example screenshot showing the display of lyrics for a performance in accordance with certain exemplary embodiments. A partial subset of lyrics may be displayed, and the lyrics may be highlighted as they are to be sung by a performer. In the FIG. 39 example, four lines are shown, and sung lines are first deemphasized (e.g., by being grayed out) as other lines are highlighted in another color when they are to be sung. The lines of lyrics may be replaced one at a time in certain exemplary embodiments, e.g., to make room for further lines of lyrics. In certain exemplary embodiments, lyrics may be color coded or otherwise selectively emphasized and/or deemphasized in connection with, for example, three different colors. For instance, lyrics in a line that have already been sung may be shown in a first color, lyrics that are being sung may be highlighted in a second color (e.g., on a line-by-line basis so that an entire line is highlighted before it is considered already-sung), and lyrical lines that have yet to be sung may be provided in a third color. In addition, or in the alternative, bolding, enlarging, and/or the like may be used to emphasize lines that are being and/or are to be sung, whereas blurring, narrowing, and/or the like may be used to deemphasize lines that have already been sung in certain exemplary embodiments.


As suggested by the FIG. 39 image, a remote control may be provided to control at least some karaoke jukebox functionality. A KJ or other authorized person may use the remote to skip songs (e.g., if a person does not show up, if a particular selection is unpopular, etc.), restart songs (e.g., if a person or party to sing a song is not situated in time, gets off to a bad start, etc.), provide free plays or credits, end songs (e.g., if a performer is doing particularly poorly), etc. A remote also may be used to adjust microphone volumes for one or more different microphones. For instance, the volume level for “Microphone B” is shown in FIG. 39. In some cases, a KJ or other authorized person may use the remote to increase the volume for particular good or bad singers for desired audience effects, etc. In scenarios where multiple microphones are provided, the microphones' volumes may be adjusted as one, in groups, and/or individually. In certain exemplary embodiments, the remote may be used to control the “house volume,” as well.


There occasionally are pauses or instrumental breaks in a song where performers do not sing. FIGS. 40 and 41 demonstrate how performers may be signaled to recognize such breaks in accordance with certain exemplary embodiments. In FIG. 40, for example, an indicator 4000 shows a short pause. Short pauses may be defined as periods where a performer does not sign for a first predetermined amount of time, e.g., in terms of number of seconds, number of beats or bars, etc. The quarter note shown in the FIG. 40 example may, for example, move from left to right (or in some other direction) to show the status of the pause. For instance, when the quarter note reaches the bar symbol, the performer may know it is time to sing again. Longer pauses, instrumental breaks, and/or the like may be shown, e.g., in accordance with the FIG. 41 display. A textual label that is selectively highlighted with one or more moving graphical elements may be used to signal the length of the delay. For instance, the instrumental break text itself may be highlighted in accordance with the duration of the pause such that, for example, when the text is fully highlighted, the performer will know it is time to sing again. In addition, or in the alternative, a progress bar or the like may be provided. Although lyrics and/or other information may be displayable on the jukebox itself, the jukebox may still be able to receive input (e.g., enabling a user to add money to a queue, add songs to a playlist, etc.). This may be facilitated by providing one or more ancillary displays with lyrics, displaying the lyrics in the LED array while the main screen is used for other purposes, dividing or partitioning the screen in some way to reserve one division/partition for lyrics and another for jukebox operates, etc.



FIG. 42 is a screenshot of karaoke-related musical rights information that may be displayed in accordance with certain exemplary embodiments. The example techniques disclosed herein are advantageous in that they allow rights to be tracked for all parties involved in a karaoke performance. This sometimes is quite difficult, since the rights associated with a karaoke performance may sometimes require payment to an artist or record label, a writer, and performers who cover the original song to lay down the karaoke track (and re-record songs in the style of the original artist, e.g., in the same or similar key, with the same or similar tempo, etc.). Many karaoke operators do not comply with reporting and/or royalty requirements, and many songs played back are unlicensed. Certain exemplary embodiments on the other hand enable plays to be tracked and proper accountings to be made in connection with a system that has trusted and verified by the industry for years as far as “normal” jukebox operations are concerned. Thus, the reporting, auditing, accounting, and/or other features that may be used in connection with “normal” jukebox operations may also be leveraged in connection with karaoke performances.


It will be appreciated that the main karaoke display and/or external displays may provide visual cues to a KJ or other person moderating a karaoke event. These cues may expressly or impliedly suggest that a microphone should be handed over to the person singing next, indicating when the next performer or group of performers should get ready and/or “take the stage,” etc. In some cases, the visual cues may even replace the need for a KJ or other moderator.


Referring once again to FIG. 29, if the search button is selected, then the example screen shown in FIG. 43 may be displayed. The FIG. 43 example karaoke search screen may in certain exemplary embodiments enable a user to perform an “all-in-one” search for words, phrases, and/or the like, that takes into account artists, songs, lyrics, and/or the like. A user may narrow down the search criteria by specifying search criteria such as some or all of these and/or other groups, specifying data ranges, recent chart toppers, or the like. It will be appreciated that the search that may be initiated in accordance with FIG. 43 may be made using a user interface that is basically the same as the search that may be initiated using the search option shown in FIG. 28 for normal jukebox operations. This may provide a continuity of experience across the different modes in which the jukebox can be operated and also make it easier for users to learn to use only one system. It will be appreciated that the results that are returned may be more limited when a karaoke search is initiated, as karaoke libraries, at least at present, are known to be smaller than “normal” songs that can be played back in a “normal” jukebox mode (e.g., because there is a separate set of license rights and related accounting that must be done in some cases).


Referring once again to FIG. 29, if the songbook button is selected, then the example screen shown in FIG. 44 may be displayed. The FIG. 44 example screenshot may include a carousel-type display of predefined songbooks such as, for example, songbooks collecting karaoke songs by “crooners” (like Frank Sinatra, Dean Martin, etc.), disco songs, popular karaoke duets, easy to sing songs, music associated with popular shows or themes, alternative music, “boy band” recorded songs, bubble gum pop songs, country classics, etc. Each songbook may include a preselected or curated selection of songs that matches that particular theme. A location may also create its own songbooks in certain exemplary embodiments, and these songbooks may be shared across locations, for example, if they are designated as being public songbooks (e.g., via myTouchTunes in manners similar to the way that playlists are sharable). In certain exemplary embodiments, the number of songs in each songbook may be displayed for user reference prior to making a selection from among the various provided songbooks. Similar to the description above, the FIG. 44 display enables a user to select between the different karaoke sub-modes and/or initiate a search. It is noted that the carousel-type display in FIG. 44 is somewhat similar to that shown in FIG. 20b and again may provide a level of consistency throughout the jukebox user experience.


Once a particular songbook is selected, the FIG. 45 example screenshot may be displayed. It will be appreciated that this list is similar to that shown in FIG. 30, and similar functionality may indeed be provided. One difference, however, is that the list is limited in accordance with the selected songbook. So that users have an idea which songbook has been selected, indicator 4500 may be provided, e.g., to include a text- and/or graphics-based indication of the same. Once again, similar to the description above, the FIG. 45 display enables a user to select between the different karaoke sub-modes and/or initiate a search.


A duet or other multi-part karaoke song may have its lyrics formatted to reflect the differences in parts. FIG. 46, for instance, is an example screenshot showing separation between first and second parts of a karaoke duet in accordance with certain exemplary embodiments. As can be appreciated from the FIG. 46 example, the screen is divided into male and female parts, as is appropriate for the song “Summer Nights.” In certain exemplary embodiments, different parts may be interleaved with one another but highlighted in different colors to visually cue the correct singer. Lyrics may be highlighted and/or otherwise emphasized and updated in accordance with the techniques above. This may be true even when the parts are to be sung together. In certain exemplary embodiments, because more lines of lyrics may need to be displayed on a screen, fewer lines may be provided for each part. For example, although four lines of lyrics are shown in the FIG. 40 example, only three lines are shown for each performer in the FIG. 46 example. The different parts are stacked one on top of the other in the FIG. 46 example, although it will be appreciated that other arrangements may be provided in different exemplary embodiments.


It will be appreciated that the song sub-mode may operate in a manner that is quite similar to the artist sub-mode. However, instead of organizing songs in the song list by artist, the song list instead will be organized by song title. Otherwise, the paging, scrubber bar, and/or other navigational aspects may function in much the same way.



FIG. 47 is an example screenshot of a karaoke games selection screen suitable for use in connection with certain exemplary embodiments. These games may provide a new way to experience karaoke and/or create opportunities to participate in events that are always different and quite unique. As above, the FIG. 47 display enables a user to select between the different karaoke sub-modes and/or initiate a search. There are three games shown in the FIG. 47 example, although these and/or other games can be provided in connection with different exemplary embodiments. A first game, dubbed “Dareoke”, enables users to pick a song for a friend to perform. That is, a user may select and pay for a song, and then challenge a friend or other person to actually perform that song. This may involve a user selecting a song from a list, selecting a registered or other user to “dare,” paying for the song, and then sending a message to the dared user. Songs may be selected in accordance with any of the example techniques disclosed herein. In certain exemplary embodiments, preselected difficult to sing, easy to sing, embarrassing, humorous, and/or other songs may be organized into lists to make song selection easier. A user may select the person to “dare” by searching for registered jukebox users, selecting a friend the person is connected to via a social media site (such as Facebook or the like), enabling the user to type in a telephone number or email address of a person (and having an SMS, email message or the like sent to the person), etc. “Dares” may be displayed on the jukebox and/or display, sent to a mobile device of the user (e.g., via SMS, MMS, email, social network message, tweets, blog post, wall update, etc.), made available to a KJ or other authorized person to announce, and/or the like. A dared user may “double-dare” the sender, e.g., provided that the person originating the first dare is known to the jukebox (e.g., by a suitable login operation or through the provision of other identifying information), and the “double-dared” sender may then have to perform the originally selected song. A message may be delivered to “double-dared” sender in the same and/or different ways as those set forth above. The performance may take place in connection with the techniques set forth above, e.g., in connection with FIG. 40.


Another example game is dubbed “Duet Roulette” and involves the random selection of a partner and/or a song to be sung. In certain exemplary embodiments, a list of duets may be maintained and selected from at random. Known jukebox users, members of a social networking site who have checked into a location, and/or others may be selected at random in certain exemplary embodiments. In certain exemplary embodiments, users may have to identify themselves as being willing to participate in a duet to be eligible to participate in this game. Messages may be sent to the parties to perform the duet in the above-described and/or other ways. The performance may take place in connection with the techniques set forth above, e.g., I connection with FIG. 46.


Still another example game is dubbed “Kamikaze Karaoke” and involves random song selection. FIG. 48 is an example screenshot showing how this game may be played in accordance with certain exemplary embodiments. In certain exemplary embodiments, a first song may be selected at random from the full song list or from a list of songs curated in the manner discussed above, e.g., in connection with the songbooks and/or example Dareoke game. If the user wishes to perform that song, then the button 4802 may be pressed. Alternatively, if the user wishes for another song to be selected at random, the “re-spin” button 4804 may be pressed. A user may re-spin up to a predetermined number of times (e.g., three) and thus have further songs randomly selected before a selection “must” be made. If the predetermined number is met, then the user may be “forced” to perform the final randomly selected song. In certain exemplary embodiments, the “remaining spin count” may be shown. In certain exemplary embodiments, the act of re-spinning may remove the selection from which the re-spin was taken from the list of possible songs that may be randomly selected in the next spin operation, and this removal may or may not be visually represented in certain exemplary embodiments. In other cases, the selection from which the re-spin was taken will not be removed from the list. Once a song is selected or once the remaining spin count drops to zero, the performance may take place in connection with the techniques set forth above, e.g., in connection with FIG. 40.



FIG. 49 is an example screenshot that illustrates incremental karaoke song queue reveal features of certain exemplary embodiments. The FIG. 49 example screenshot includes an indication of the currently performing song 4902, including the name of the performer and the song being performed. The indication of the currently performing song 4902 may also have a progress bar associated with it in order to demonstrate how far along a particular performance is. In the FIG. 49 example, the progress bar is included as gradually increasing shading or selective re-coloration in the generally rectangular indicator 4902 itself.


A back button 4904 enables a user to see a listing of the songs and/or performers who recently have performed. If a score or rating is associated with a performance, that information may be visible as well (e.g., from the main display itself, or once an indicator corresponding to the performance for which additional information is desired is selected). One or more next song indicators 4906a-b also may be shown. Each subsequent song indicator 4906 may indicate the singer, the song, and/or an approximate wait time for the song. In certain exemplary embodiments, precise times may be provided, whereas other exemplary embodiments may provide estimates (e.g., rounded up or down to the closest minute, always rounded up to the closest minute, etc.). Estimates may be provided at the minute level or at some other incremental level such as, for example, at the 15 second, 30 second, or other interval. In certain exemplary embodiments, estimates may be provided at the 30 second level for songs expected to be performed within 5 minutes, at the minute level for songs expected to be performed in 5-15 minutes, at the 3 minute level for songs to be performed in 15-30 minutes, and/or the like. These estimates may take into account lead-in and/or lead-out audio and/or video segments. In certain exemplary embodiments, a short amount of intro and outro audiovisual information may be provided, e.g., in connection with a custom music programming (CMP) channel. The CMP channel may feature a high-energy playlist to play songs or clips before and/or after a performance. This may help to maintain an energetic or other desired atmosphere in the karaoke jukebox location, provide more natural transitions between performers (even in situations where the is no KJ or other moderator), etc. In certain exemplary embodiments, multiple CMP channels may be provided, e.g., for different themes. The themes may be specified by a location and, for example, country bars may have a first channel, biker bars may have a second channel, etc. In certain exemplary embodiments, the CMP channel may be generic. In certain exemplary embodiments, CMP clips may be selected based on the song just performed and/or about to be performed, e.g., so that a clip with an appropriate “feel” will prepare the audience for a transition. In certain exemplary embodiments, intro and/or outro clips may last about 45 seconds or some other longer or shorter duration that may be consistent or variable across different clips in different exemplary embodiments.


The number of next song indicators 4906 to be displayed on the screen may be fixed or set by the operator, KJ, or other authorized person. In certain exemplary embodiments, the ability to see the queue, in whole or in part, may be an added feature reserved for certain users (e.g., registered jukebox users, jukebox users who sign in through a social network type account, etc.). In addition, or in the alternative, in certain exemplary embodiments, the ability to see the queue, in whole or in part, may be provided upon the payment of additional money. In some cases, more money may be charged to see more of the queue.



FIG. 50 is an example screenshot showing how an authorized person may edit the karaoke jukebox queue in accordance with certain exemplary embodiments. In certain exemplary embodiments, the queue may be displayed, along with summary information. This summary information may include, for example, the number of performers following the current performer, how much time is accounted for, when the queue is expected to run out of entries, etc. In certain exemplary embodiments, the queue listing may include the name of the performer, the song to be performed, a numerical entry corresponding to the placement in the queue, an avatar associated with the performer (if any), and/or the like. In certain exemplary embodiments, an authorized user may drag-and-drop items and/or use another mechanism to reorder the entries in the queue, indicate whether a song should be completely removed from the queue, etc. The FIG. 50 example screenshot may be accessed via the jukebox itself, remotely from a suitably configured device that has logged into the jukebox system with appropriate credentials, etc. In certain exemplary embodiments, the queue may be locked, e.g., such that it cannot be edited and/or added to. This may in some cases helps to ensure that all songs are played through in a particular time period and so that would-be performers do not get frustrated if they pay for songs but cannot perform them (e.g., if they have to leave the location because it is closing, because a dedicated karaoke time has ended, etc.).



FIG. 51 is a block diagram illustrating example components of a karaoke jukebox system in accordance with certain exemplary embodiments. FIG. 51 shows a suitably configured jukebox 5100 (which may in certain example embodiments be a computer or computer like system, e.g., with at least one processor, a memory, and non-transitory computer readable storage media) with a plurality of input and output ports. For instance, the main I/O board of the jukebox 5100 may include inputs and outputs for video, light shows, audio, lighting for floor and/or external sources, etc. In the FIG. 51 example, the jukebox 5100 is connected to a wireless microphone receiver 5102 and a wireless router 5104. The wireless router 5104 may provide a connection to the Internet, as well as to a Video Interface Module (VIM) 5106.


The VIM 5106, in turn, includes video (e.g., HDMI, component, etc.) and/or CATS connections to one or more displays 5108 in a location and thus provides a connection between the jukebox 5100 and the one or more displays 5108. The VIM 5106 may be a hardware component separate from the jukebox 5100 in certain exemplary embodiments, whereas it may be integrated into the jukebox housing and/or part of the main I/O board of the jukebox 5100 in other exemplary embodiments. The VIM 5106 may serve as a sort of relay in certain exemplary embodiments but, as indicated above, it may “cut into” normal programming on the one or more displays 5108 in response to suitable control signals from the jukebox 5100. It is noted that outputs from the VIM 5106 may be split (e.g., using splitters or the like) in certain exemplary embodiments to provide the same display and/or control signals to multiple different displays. As indicated above, a camera (e.g., a 720p camera or the like), touch screen (e.g., based on a 26″ or other sized display), LED array, LED wash lights, etc., may be provided. It will be appreciated that the displays and/or lighting elements, for example, may function independent of one another. For instance, the wash light elements may be set to randomly display lights, programmed to match what is happening on the display, etc.


Because of the few components, karaoke setup may be quite easy.


It will be appreciated that the jukebox, when operating in the karaoke mode, may be at least partially controlled in via a mobile device. A mobile application of the sort described in 2012/0158531 may be extended in this vein, e.g., to enable a user to search for and select karaoke songs to be performed, pay for karaoke performances, etc. The search and/or browse features that are provided may be simplified versions of the search and/or browse features disclosed herein. For instance, users may organize karaoke songbooks by artist, song, predefined collections, etc., and/or may initiate searches across all artists, songs, songbooks, lyrics, and/or the like. If a user checks in to a particular location, or selects a particular location, the user may use the mobile app to search for music available through the relevant jukebox, e.g., by conducting a search based on a karaoke song list accessed from the karaoke jukebox device itself. In a similar vein, users also may see how many credits they have, view lyrics, determine what key a song is in, etc., and generally obtain the same information as that set forth above. The mobile devices also may be synched up with a particular performance so that lyrics appear on the mobile device itself, thereby creating a sense that the jukebox is at least partially controlling the mobile device. This may be beneficial because the mobile device may in essence serve as yet another ancillary display device, e.g., for lyrics presentations, attract modes, etc.


In addition to searching for songs, users may also have the ability to search for karaoke-enabled jukeboxes. This may include, for example, the ability to search for scheduled karaoke nights, locations that have “on-demand” karaoke features enabled, etc. To facilitate the search, operators, location owners, and/or other persons may access a website or use a function provided to the jukeboxes to specify their karaoke schedules and/or indicate whether karaoke is available on-demand. The website may know which locations are involved by virtue of the credentials through which the person logged in, in view of a unique identifier associated with a jukebox used to access the website, through location information specifically provided by the authorized person, etc. In certain exemplary embodiments, karaoke scheduling information may be maintained locally (e.g., on the jukebox, local computer, or the like) and distributed to a central server upon a “commit” or other operation. Once the scheduling information is provided to the central server or other accessible data store, the application may search through it, e.g., in connection with geographic restrictions, time/date restrictions, theme restrictions, and/or the like, that may be specified by the user and/or gleaned from the user's device (through GPS, radio triangulation, and/or other techniques).


Mobile app users may check in to locations, which may synch their devices to the jukeboxes. This may make it possible to use mobile apps to record part of all of karaoke performances, e.g., as disclosed in U.S. application Ser. No. 13/621,922, filed Sep. 18, 2012.


It will be appreciated that the flowcharts and use cases presented above are provided by way of example and without limitation. Other flowcharts and use cases are possible in connection with different exemplary embodiments of this invention.


The metadata and/or tags associated with the songs may be provided by a commercial service, by the jukebox system, users, and/or some combination of these and/or other providers. Some such data may be determined automatically (e.g., through Fourier analysis and/or other transforms). For music, metadata and/or tags may be representative of information including, for example, artist name, album name, song name, genre/style (which may be hierarchical in some instances), Billboard position (when applicable), release year, popularity based on network plays information, lyrics, tempo, beats per minute (BPM), mood, instrumental characteristics (e.g., strident electric guitars, booming drums, twangy guitar, etc.), vocal characteristics (e.g., aggressive, prominent vocal harmonies, screaming, etc.), intensity characteristics (e.g., high energy, moderate, laid back, etc.) timbre (e.g., bright, dark, rich, etc.), and/or the like. Potential suppliers of this and/or other information include, for example, AMG, BMAT, DoubleV3, EchoNest MixZing, Gracenote, Mufin MusicIP, Music Box, Music Genome Project, and Syntonetic.


Certain exemplary embodiments relate to an entertainment center comprising a computer capable of communicating with networks, wherein said computer is further connected to at least one display through standard analog, digital, or network-addressable displays, said computer being operable to interact with a remote device connected to one of said networks in communication with said computer, said remote device being operable to accept a code and transmit said code to said computer, and wherein said computer can validate against a database or against an algorithm the validity of said code and, upon positive validation, said computer is configured to allocate a monetary value or a credit value to said remote device. The remote device may be operable to browse content contained on said computer and said remote device may be further operable to select and pay for said content using said monetary or said credit value, said computer may be operable to reduce said monetary or said credit value upon a selection by said remote device. The code may instead or in addition be sent to the remote device and entered on the computer.


Certain exemplary embodiments relate to an out-of-home entertainment center coupled with at least one Internet-based messaging system and/or a social networking site and coupled with at least one remote device, said remote device being connected to the out of home entertainment center by a wired or wireless local area network or through the Internet, wherein the use of some of the entertainment center services by said remote device causes said entertainment center to send messages to said at least one Internet-based messaging system. Connecting the system through the Internet may require a user to input a code to the remote device that uniquely identifies the entertainment center.


Some of the handheld devices referred to herein may be those described, for example, in application Ser. No. 11/902,790, the entire contents of which is incorporated herein by reference.


The present disclosure has used certain terms that should not be interpreted as limiting the invention to a particular embodiment, hardware components and configurations, software configurations, etc. For example, many features and examples have been described in relation to their existence within a bar, pub, or other environment. However, it will be appreciated that the features present in the exemplary embodiments of the present invention are adaptable for use in any location where a jukebox (or multiple jukeboxes) may be located. Similarly, while certain features and functions are described with reference to usage by “users,” “owners,” “operators,” “patrons,” etc., it will be appreciated that these terms are generic and may, in most cases, be used interchangeably depending on the embodiment chosen and the feature employed. For example, while it may be advantageous to limit the initial song selection to owners and/or operators, in certain exemplary embodiments, patrons may play a role in the initial song selection. It will be appreciated that the term “display” includes, for example, monitors connected to computers directly or remotely, or embedded ICs such as IP TV technology. Displays may be network addressable. Also, standard digital signs (LED based) also may be considered displays and/or may be provided as network addressable displays.


Although certain exemplary embodiments have been described in connection with out-of-home locations, it will be appreciated that the techniques described herein may be adapted for use in an in-home or personal jukebox.


Still further, particular hardware combinations and configurations are disclosed which represent only one way which the embodiments may be constructed. Central servers may, in some exemplary embodiments, be comprised of one or more servers acting together or separately to coherently provide the full range of services necessary to enable a functioning jukebox. For example, a cluster of servers may comprise a virtual central server, with one server providing media, another tracking membership, still another processing licensing, etc. Similarly, the local servers described herein may be incorporated into the jukeboxes. For example, the local servers may appear to function independently, even though they exist as part (e.g. partition) of an integrated mass storage device within the jukebox. Indeed, as hard disks become larger and less expensive, they may preferably serve the functions of local servers.


Also, although the term “song” has been used sometimes in the above-description, this term is not intended to be limiting to the scope of the invention, and any instance or instances of media (e.g., song, video, song/video combination, data, information etc.) can be used in any embodiment herein and still fall within the intended scope of the invention.


Lastly, it will be appreciated that the screen shots and software arrangements presented herein are only one exemplary method for organizing and displaying the features disclosed herein. Other configurations are possible and are therefore contemplated herein.


While the preferred aspects of the invention have been illustrated and described herein, it will be apparent to one of ordinary skill in the art that various changes and/or modifications can be made. Thus, the specific description herein is meant to be exemplary only and is not intended to limit the invention beyond the terms of appended claims.

Claims
  • 1. A jukebox device, comprising: processing resources, including at least one processor and a memory;a touch screen display device;a payment acceptor; anda storage device comprising instances of media available for playback via the jukebox device;wherein the processing resources, the touch screen display device, and the payment acceptor cooperate with one another to cause the jukebox device to function in at least first and second different modes,wherein the first mode is a normal jukebox operation mode that is configured to enable a first partial subset of the instances of media to be played back on a pay-per-play basis, andwherein the second mode is a karaoke jukebox operation mode that is configured to enable a second partial subset of the instances of media to be played back on a pay-per-play basis, the first and second partial subsets being different from one another, andwherein, as a part of the second mode, the touch screen display device presents a first set of screens that each list at least a portion of the instances of media in the second partial subset of the instances of media in a manner that resembles a book having multiple pages,each said page including at most a predefined maximum number of entries corresponding to the portion of the second partial subset of instances of media, and each said entry indicating a song name and an artist associated with the corresponding instance of media that is selectable for playback via the jukebox device, andwherein the first set of screens each include a common set of control elements for re-organizing and navigating through the book, and wherein said re-organizing includes dynamically rearranging, when an input is received indicating a letter for a song name or an artist, the listed portions of the instances of media in the second partial subset such that ones of said entries starting with the indicated letter are listed starting at the top of a newly turned page in said book by dynamically reformatting the newly turned page to cause said entries that ordinarily would not appear at the start of a page to be adjusted accordingly to start at the top of the newly turned page.
  • 2. The jukebox device of claim 1, wherein: the common set of control elements includes previous and next page buttons that are selectively enabled in allowing a user to navigate through the book, andactuation of the buttons causes an animation to be displayed, in relation to the book, to simulate a consistent and mechanical page-turning operation and correspondingly display entries for a new page.
  • 3. The jukebox device of claim 2, wherein: the touch screen display device is configured to receive user input at or near a corner of the book; andthe processing resources are configured to, in accordance with received user input and in relation to the book, create a custom animation that simulates a page turning in a manner that generally tracks received user input across the book and correspondingly display entries for a new page.
  • 4. The jukebox device of claim 3, wherein animations that simulate consistent and mechanical page-turning operations and custom animations that simulate pages turning in manners that generally track received user input both temporarily show portions of entries on each side of the turning pages.
  • 5. The jukebox device of claim 1, wherein: the first set of screens includes a first screen for an artist-organized book, and a second screen for a song-organized book, andthe common set of control elements includes elements for switching between the first and second screens.
  • 6. The jukebox device of claim 5, wherein the books associated with the first and second screens each include entries for all instances of media in the second partial subset of instances of media.
  • 7. The jukebox device of claim 6, wherein the first set of screens includes a third screen for displaying a songbook, the songbook including entries for only some of the instances of media in the second partial subset of instances of media.
  • 8. The jukebox device of claim 6, further comprising a songbook selection screen, the songbook selection screen including representations of a plurality of songbooks, each said songbook including entries for only some of the instances of media in the second partial subset of instances of media.
  • 9. The jukebox device of claim 8, wherein each said songbook includes entries for a preselected list of instances of media in the second partial subset of instances of media, each preselected list being associated with a theme appropriate for the respective songbook.
  • 10. The jukebox device of claim 8, wherein the songbook selection screen includes at least some of the control elements from the set of common control elements.
  • 11. The jukebox device of claim 10, wherein the control elements from the set of common control elements that are included in the songbook selection screen include the elements for switching between the first and second screens, and the common set of control elements includes an element for switching to the songbook selection screen.
  • 12. The jukebox device of claim 11, wherein the first set of screens includes a third screen for displaying a songbook selected from the songbook selection screen.
  • 13. The jukebox device of claim 1, wherein: the common set of control elements includes a generally linear arrangement of selectable alphanumeric characters, andselection of one of the alphanumeric characters via the touch screen display device advances pages in the songbook so that a first entry beginning with the selected alphanumeric character appears with respect to the book.
  • 14. The jukebox device of claim 1, wherein, as a part of the second mode, the touch screen display device is configured to display a performer identifier screen, the performer identifier screen accepting an identifying name of a user who has selected an instance of media for a karaoke-type performance.
  • 15. The jukebox device of claim 14, wherein the identifying name of the user who has selected the instance of media for the karaoke-type performance is specifiable in connection with a virtual keyboard on the touch screen display device.
  • 16. The jukebox device of claim 14, wherein the identifying name of the user who has selected the instance of media for the karaoke-type performance is determined automatically if a user logs into the jukebox device.
  • 17. The jukebox device of claim 14, wherein the identifying name of the user who has selected the instance of media for the karaoke-type performance is determined automatically if a user provides valid credentials for a recognized social networking site.
  • 18. The jukebox device of claim 14, wherein the identifying name of the user who has selected the instance of media for the karaoke-type performance is determined automatically if a user logs into the jukebox device and/or provides valid credentials for a recognized social networking site.
  • 19. The jukebox device of claim 1, wherein avatars are displayable on the jukebox device and/or an external display connected to the jukebox device, in connection karaoke performances by users who have provided corresponding images and who have successfully identified themselves to the jukebox device.
  • 20. The jukebox device of claim 19, wherein avatar images are retrievable from a social networking site and/or capturable using a camera connected to the jukebox device.
CROSS-REFERENCES TO RELATED APPLICATIONS

This application is a continuation of U.S. application Ser. No. 15/074,863 filed Mar. 18, 2016 (U.S. Pat. No. 9,959,012 issued May 1, 2018) which is a continuation of U.S. application Ser. No. 13/833,173 filed Mar. 15, 2013, now U.S. Pat. No. 9,292,166 issued Mar. 22, 2016, which is a continuation-in-part (CIP) of U.S. application Ser. No. 13/621,922 filed Sep. 18, 2012, now U.S. Pat. No. 9,324,064 issued Apr. 26, 2016, which claims the benefit of Provisional Application Nos. 61/584,750 filed Jan. 9, 2012 and 61/536,015 filed Sep. 18, 2011; this application is also a continuation-in-part (CIP) of U.S. application Ser. No. 13/138,660 filed Mar. 5, 2012, now U.S. Pat. No. 9,076,155, which is a National Stage Application of International Application No. PCT/US2010/000799 filed Mar. 17, 2010, which claims the benefit of Provisional Application No. 61/202,617 filed Mar. 18, 2009; and this application is also a continuation-in-part (CIP) of U.S. application Ser. No. 12/929,466 filed Jan. 26, 2011, now U.S. Pat. No. 9,521,375 issued Dec. 13, 2016, which claims the benefit of Provisional Application Nos. 61/431,036 filed Jan. 9, 2011 and 61/298,509 filed Jan. 26, 2010, the entire contents of each of which are hereby incorporated by reference in this application.

US Referenced Citations (728)
Number Name Date Kind
3710137 Stephens, Jr. Jan 1973 A
3807541 Kortenhaus Apr 1974 A
3982620 Kotenhaus Sep 1976 A
4008369 Theurer et al. Feb 1977 A
4064362 Williams Dec 1977 A
4186438 Benson Jan 1980 A
4232295 McConnell Nov 1980 A
4335809 Wain Jun 1982 A
4335908 Burge Jun 1982 A
4336935 Goldfarb Jun 1982 A
4356509 Skerlos et al. Oct 1982 A
4369442 Werth et al. Jan 1983 A
4375287 Smith Mar 1983 A
4412292 Sedam Oct 1983 A
4413260 Siegel et al. Nov 1983 A
4521014 Sitrick Jun 1985 A
4528643 Freeny Jul 1985 A
4558413 Schmidt et al. Dec 1985 A
4572509 Sitrick Feb 1986 A
4577333 Lewis et al. Mar 1986 A
4582324 Koza Apr 1986 A
4588187 Dell May 1986 A
4593904 Graves Jun 1986 A
4597058 Izumi Jun 1986 A
4636951 Harlick Jan 1987 A
4652998 Koza Mar 1987 A
4654799 Ogaki Mar 1987 A
4658093 Hellman Apr 1987 A
4667802 Verduin et al. May 1987 A
4674055 Ogaki et al. Jun 1987 A
4675538 Epstein Jun 1987 A
4677311 Morita Jun 1987 A
4677565 Ogaki Jun 1987 A
4696527 Ding et al. Sep 1987 A
4703465 Parker Oct 1987 A
4704725 Harvey et al. Nov 1987 A
4707804 Leal Nov 1987 A
4722053 Dubno Jan 1988 A
4761684 Clark Aug 1988 A
4766581 Korn et al. Aug 1988 A
4787050 Suzuki Nov 1988 A
4792849 McCalley Dec 1988 A
4807052 Amano Feb 1989 A
4811325 Sharples Mar 1989 A
4814972 Winter et al. Mar 1989 A
4815868 Speicher Mar 1989 A
4825054 Rust Apr 1989 A
4829570 Schotz May 1989 A
4852154 Lewis et al. Jul 1989 A
4857714 Sunyich Aug 1989 A
4868832 Marrington Sep 1989 A
4885694 Pray et al. Dec 1989 A
4905279 Nishio Feb 1990 A
4920432 Eggers Apr 1990 A
4922420 Nakagawa May 1990 A
4924378 Hershey May 1990 A
4926485 Yamashita May 1990 A
4937807 Weitz Jun 1990 A
4949187 Cohen Aug 1990 A
4953159 Hayden et al. Aug 1990 A
4956768 Sidi Sep 1990 A
4958835 Tashiro Sep 1990 A
4965675 Masashi et al. Oct 1990 A
4977593 Ballance Dec 1990 A
4999806 Chernow Mar 1991 A
5008814 Mathur Apr 1991 A
5012121 Hammond Apr 1991 A
5027426 Chiocca Jun 1991 A
5041921 Scheffler Aug 1991 A
5046093 Wachob Sep 1991 A
5053758 Cornett et al. Oct 1991 A
5058089 Yoshimara Oct 1991 A
5077607 Johnson et al. Dec 1991 A
5081534 Geiger et al. Jan 1992 A
5101451 Ash et al. Mar 1992 A
5101499 Streck et al. Mar 1992 A
5106097 Levine Apr 1992 A
5117407 Vogel May 1992 A
D327687 Arbitor Jul 1992 S
5128862 Mueller Jul 1992 A
5138712 Corbin Aug 1992 A
5148159 Clark et al. Sep 1992 A
5155847 Kirouac Oct 1992 A
5159678 Wengelski et al. Oct 1992 A
5163131 Row Nov 1992 A
5166886 Molnar Nov 1992 A
D332096 Wolff et al. Dec 1992 S
5172413 Bradley et al. Dec 1992 A
5180309 Egnor Jan 1993 A
5189630 Barstow et al. Feb 1993 A
5191573 Hair Mar 1993 A
5191611 Lang Mar 1993 A
5192999 Graczyk Mar 1993 A
5197094 Tillery Mar 1993 A
5203028 Shiraishi Apr 1993 A
5210854 Beaverton et al. May 1993 A
5214761 Barrett et al. May 1993 A
5222134 Waite et al. Jun 1993 A
5228015 Arbiter et al. Jul 1993 A
5231157 Herzig et al. Jul 1993 A
5237157 Kaplan Aug 1993 A
5237322 Heberle Aug 1993 A
5239480 Huegel Aug 1993 A
5250747 Tsumura Oct 1993 A
5252775 Urano Oct 1993 A
5260999 Wyman Nov 1993 A
5261104 Bertram et al. Nov 1993 A
5262875 Mincer et al. Nov 1993 A
5276866 Paolini Jan 1994 A
5278904 Servi Jan 1994 A
5282028 Johnson et al. Jan 1994 A
5289476 Johnson et al. Feb 1994 A
5289546 Hetherington Feb 1994 A
5315161 Robinson May 1994 A
5315711 Barone et al. May 1994 A
5319455 Hoarty et al. Jun 1994 A
5321846 Yokota et al. Jun 1994 A
5327230 Dockery Jul 1994 A
5335313 Douglas Aug 1994 A
5339095 Redford Aug 1994 A
5339413 Koval Aug 1994 A
5341350 Frank Aug 1994 A
5355302 Martin et al. Oct 1994 A
5357276 Banker Oct 1994 A
5369778 SanSoucie Nov 1994 A
5375206 Hunter Dec 1994 A
5386251 Movshovich Jan 1995 A
5389950 Martin et al. Feb 1995 A
5404505 Levinson Apr 1995 A
5406634 Anderson et al. Apr 1995 A
5408417 Wilder Apr 1995 A
5410326 Goldstein Apr 1995 A
5410703 Nilsson et al. Apr 1995 A
5418713 Allen May 1995 A
5420923 Beyers May 1995 A
5428252 Walker Jun 1995 A
5428606 Moskowitz Jun 1995 A
5431492 Rothschild Jul 1995 A
5440632 Bacon et al. Aug 1995 A
5444499 Saitoh Aug 1995 A
5445295 Brown Aug 1995 A
5455619 Truckenmiller et al. Oct 1995 A
5455926 Keele Oct 1995 A
5457305 Akel Oct 1995 A
5465213 Ross Nov 1995 A
5465329 Whistler Nov 1995 A
5467326 Miyashita et al. Nov 1995 A
5469370 Ostrover et al. Nov 1995 A
5469573 McGill et al. Nov 1995 A
5471576 Yee Nov 1995 A
5473746 Pritt et al. Dec 1995 A
5475835 Hickey Dec 1995 A
5481509 Knowles Jan 1996 A
5487167 Dinallo et al. Jan 1996 A
5489103 Okamoto Feb 1996 A
5495610 Shing Feb 1996 A
5496178 Back Mar 1996 A
5499921 Sone Mar 1996 A
5511000 Kaloi Apr 1996 A
5513117 Small Apr 1996 A
5515173 Mankovitz et al. May 1996 A
5519435 Anderson May 1996 A
5519457 Nishigaki et al. May 1996 A
5521631 Budow et al. May 1996 A
5521918 Kim May 1996 A
5521922 Fujinami et al. May 1996 A
5523781 Brusaw Jun 1996 A
5528732 Klotz Jun 1996 A
5532734 Goertz Jul 1996 A
5532991 Sasaki Jul 1996 A
5546039 Hewitt et al. Aug 1996 A
5548729 Akiyoshi Aug 1996 A
5550577 Verbiest Aug 1996 A
5554968 Lee Sep 1996 A
5555244 Gupta Sep 1996 A
5557515 Abbruzzese et al. Sep 1996 A
5557541 Schulhof Sep 1996 A
5557724 Sampat et al. Sep 1996 A
5559505 McNair Sep 1996 A
5559549 Hendricks Sep 1996 A
5559714 Banks et al. Sep 1996 A
5561709 Remillard Oct 1996 A
5565908 Ahmad Oct 1996 A
5566237 Dobbs Oct 1996 A
5570363 Holm Oct 1996 A
5578999 Matsuzawa et al. Nov 1996 A
5579404 Fielder et al. Nov 1996 A
5583561 Baker et al. Dec 1996 A
5583937 Ullrich et al. Dec 1996 A
5583994 Rangan Dec 1996 A
5583995 Gardner et al. Dec 1996 A
5590318 Zbikowski et al. Dec 1996 A
5592482 Abraham Jan 1997 A
5592551 Lett Jan 1997 A
5592611 Midgely et al. Jan 1997 A
5594509 Florin Jan 1997 A
5596702 Stucka et al. Jan 1997 A
5607099 Yeh et al. Mar 1997 A
5612581 Kageyama Mar 1997 A
5613909 Stelovsky Mar 1997 A
5616876 Cluts Apr 1997 A
5617565 Augenbraun et al. Apr 1997 A
5619247 Russo Apr 1997 A
5619249 Billock et al. Apr 1997 A
5619250 McClellan et al. Apr 1997 A
5619698 Lillich Apr 1997 A
5623666 Pike Apr 1997 A
5631693 Wunderlich et al. May 1997 A
5636276 Brugger Jun 1997 A
5638426 Lewis Jun 1997 A
5642337 Oskay et al. Jun 1997 A
5643831 Ochiai et al. Jul 1997 A
5644714 Kikinis Jul 1997 A
5644766 Coy Jul 1997 A
5654714 Takahashi et al. Aug 1997 A
5659466 Norris et al. Aug 1997 A
5661517 Budow et al. Aug 1997 A
5661802 Nilssen Aug 1997 A
5663756 Blahut et al. Sep 1997 A
5668592 Spaulding Sep 1997 A
5668778 Quazi Sep 1997 A
5668788 Allison Sep 1997 A
5675734 Hair Oct 1997 A
5680533 Yamato et al. Oct 1997 A
5684716 Freeman Nov 1997 A
5689641 Ludwig et al. Nov 1997 A
5691778 Song Nov 1997 A
5691964 Niederlein et al. Nov 1997 A
5696914 Nahaboo et al. Dec 1997 A
5697844 Von Kohorn Dec 1997 A
5703795 Mankowitz Dec 1997 A
5704146 Herring et al. Jan 1998 A
5708811 Arendt Jan 1998 A
5712976 Falcon et al. Jan 1998 A
5713024 Halladay Jan 1998 A
5715416 Baker Feb 1998 A
5717452 Janin et al. Feb 1998 A
5721583 Harada et al. Feb 1998 A
5721815 Ottesen et al. Feb 1998 A
5721827 Logan et al. Feb 1998 A
5721829 Dunn et al. Feb 1998 A
5724525 Beyers et al. Mar 1998 A
5726909 Krikorian Mar 1998 A
5734719 Tsevdos et al. Mar 1998 A
5734961 Castille Mar 1998 A
5739451 Winksy et al. Apr 1998 A
5743745 Reintjes Apr 1998 A
5745391 Topor Apr 1998 A
5748254 Harrison et al. May 1998 A
5748468 Notenboom et al. May 1998 A
5748954 Mauldin May 1998 A
5751336 Aggarwal et al. May 1998 A
5752232 Basore et al. May 1998 A
5757936 Lee May 1998 A
5758340 Nail May 1998 A
5761655 Hoffman Jun 1998 A
5762552 Vuong Jun 1998 A
5774527 Handelman et al. Jun 1998 A
5774668 Choqiuer Jun 1998 A
5774672 Funahashi Jun 1998 A
5778395 Whiting Jul 1998 A
5781889 Martin et al. Jul 1998 A
5786784 Gaudichon Jul 1998 A
5790172 Imanaka Aug 1998 A
5790671 Cooper Aug 1998 A
5790856 Lillich Aug 1998 A
5790935 Payton Aug 1998 A
5793364 Bolanos et al. Aug 1998 A
5793980 Glaser Aug 1998 A
5798785 Hendricks Aug 1998 A
5802283 Grady et al. Sep 1998 A
5802558 Pierce Sep 1998 A
5802599 Cabrera Sep 1998 A
5805804 Laursen et al. Sep 1998 A
5808224 Kato Sep 1998 A
5809246 Goldman Sep 1998 A
5812643 Schelberg et al. Sep 1998 A
5815146 Youden et al. Sep 1998 A
5825884 Zdepski et al. Oct 1998 A
5828343 MacDonald et al. Oct 1998 A
5831555 Yu et al. Nov 1998 A
5831663 Waterhouse et al. Nov 1998 A
5832024 Schotz et al. Nov 1998 A
5832287 Atalla Nov 1998 A
5835843 Haddad Nov 1998 A
5842869 McGregor et al. Dec 1998 A
5845104 Rao Dec 1998 A
5845256 Pescitelli et al. Dec 1998 A
5848398 Martin Dec 1998 A
5851149 Xidos et al. Dec 1998 A
5854887 Kindell Dec 1998 A
5857020 Peterson Jan 1999 A
5857707 Devlin Jan 1999 A
5862324 Collins Jan 1999 A
5864811 Tran et al. Jan 1999 A
5864868 Contois Jan 1999 A
5864870 Guck Jan 1999 A
5867714 Todd Feb 1999 A
5870721 Norris Feb 1999 A
5880386 Wachi et al. Mar 1999 A
5880769 Nemirofsky et al. Mar 1999 A
5884028 Kindell Mar 1999 A
5884298 Smith Mar 1999 A
5887139 Madison, Jr. et al. Mar 1999 A
5887193 Takahashi Mar 1999 A
5893162 Lau et al. Apr 1999 A
5895455 Bellinger et al. Apr 1999 A
5896094 Narisada et al. Apr 1999 A
5903266 Berstis et al. May 1999 A
5913040 Rakavy Jun 1999 A
5914712 Sartain et al. Jun 1999 A
5915094 Kouloheris Jun 1999 A
5915238 Tjaden Jun 1999 A
5917537 Lightfoot Jun 1999 A
5917835 Barrett Jun 1999 A
5918213 Bernard et al. Jun 1999 A
5920700 Gordon et al. Jul 1999 A
5920702 Johnson Jul 1999 A
5923885 Johnson Jul 1999 A
5926531 Petite Jul 1999 A
5926624 Katz et al. Jul 1999 A
5930765 Martin Jul 1999 A
5931908 Gerba Aug 1999 A
5933090 Christenson Aug 1999 A
5940504 Griswold Aug 1999 A
5949411 Doerr et al. Sep 1999 A
5949688 Montoya Sep 1999 A
5953005 Liu Sep 1999 A
5953429 Wakai et al. Sep 1999 A
5956716 Kenner et al. Sep 1999 A
5959869 Miller Sep 1999 A
5959945 Kleiman Sep 1999 A
5960167 Roberts et al. Sep 1999 A
5963916 Kaplan Oct 1999 A
5966495 Takahashi Oct 1999 A
5970467 Alavi Oct 1999 A
5978855 Metz et al. Nov 1999 A
5978912 Rakavy et al. Nov 1999 A
5980261 Mino et al. Nov 1999 A
5999499 Pines et al. Dec 1999 A
5999624 Hopkins Dec 1999 A
6002720 Yurt Dec 1999 A
6005599 Asai et al. Dec 1999 A
6008735 Chiloyan et al. Dec 1999 A
6009274 Fletcher Dec 1999 A
6011758 Dockes et al. Jan 2000 A
6018337 Peters Jan 2000 A
6018726 Tsumura Jan 2000 A
6021386 Davis Feb 2000 A
6023705 Bellinger et al. Feb 2000 A
6025868 Russo Feb 2000 A
6026168 Li et al. Feb 2000 A
6034925 Wehmeyer Mar 2000 A
6038591 Wolfe et al. Mar 2000 A
6040829 Croy et al. Mar 2000 A
6041354 Biliris et al. Mar 2000 A
6049891 Inamoto Apr 2000 A
6054987 Richardson Apr 2000 A
6055573 Gardenswartz et al. Apr 2000 A
6057874 Michaud May 2000 A
6067564 Urakoshi et al. May 2000 A
6069672 Claassen May 2000 A
6072982 Haddad Jun 2000 A
6107937 Hamada Aug 2000 A
6118450 Proehl et al. Sep 2000 A
6124804 Kitao et al. Sep 2000 A
6131088 Hill Oct 2000 A
6131121 Mattaway et al. Oct 2000 A
6134547 Huxley et al. Oct 2000 A
6138150 Nichols et al. Oct 2000 A
6146210 Cha et al. Nov 2000 A
6148142 Anderson Nov 2000 A
6151077 Vogel et al. Nov 2000 A
6151634 Glaser Nov 2000 A
6154207 Farris et al. Nov 2000 A
6157935 Tran et al. Dec 2000 A
6161059 Tedesco et al. Dec 2000 A
6167358 Othmer et al. Dec 2000 A
6170060 Mott et al. Jan 2001 B1
6173172 Masuda et al. Jan 2001 B1
6175861 Williams, Jr. et al. Jan 2001 B1
6182126 Nathan et al. Jan 2001 B1
6185184 Mattaway et al. Feb 2001 B1
6185619 Joffe et al. Feb 2001 B1
6191780 Martin et al. Feb 2001 B1
6192340 Abecassis Feb 2001 B1
6195732 Adams et al. Feb 2001 B1
6198408 Cohen Mar 2001 B1
6202060 Tran Mar 2001 B1
6209060 Machida Mar 2001 B1
6212138 Kalis et al. Apr 2001 B1
6216175 Sliger et al. Apr 2001 B1
6216227 Goldstein et al. Apr 2001 B1
6219692 Stiles Apr 2001 B1
6223209 Watson Apr 2001 B1
6226412 Schwab May 2001 B1
6226715 Van Der Wolf et al. May 2001 B1
6240550 Nathan et al. May 2001 B1
6243725 Hempleman et al. Jun 2001 B1
6247022 Yankowski Jun 2001 B1
6256773 Bowman-Amuah Jul 2001 B1
6262569 Carr et al. Jul 2001 B1
6280327 Leifer et al. Aug 2001 B1
6282709 Reha et al. Aug 2001 B1
6288688 Hughes et al. Sep 2001 B1
6288991 Kajiyama et al. Sep 2001 B1
6289382 Bowman-Amuah Sep 2001 B1
6292443 Awazu et al. Sep 2001 B1
6298373 Burns et al. Oct 2001 B1
6301710 Fujiwara Oct 2001 B1
6302793 Fertitta et al. Oct 2001 B1
6308204 Nathan et al. Oct 2001 B1
6311214 Rhoads Oct 2001 B1
6315572 Glaser Nov 2001 B1
6323911 Schein et al. Nov 2001 B1
6332025 Takahashi et al. Dec 2001 B2
6336219 Nathan Jan 2002 B1
6337698 Keely, Jr. Jan 2002 B1
6341166 Basel Jan 2002 B1
6344862 Williams et al. Feb 2002 B1
6346951 Mastronardi Feb 2002 B1
6353820 Edwards et al. Mar 2002 B1
6356971 Katz et al. Mar 2002 B1
6359616 Ogura et al. Mar 2002 B1
6359661 Nickum Mar 2002 B1
6370580 Kriegsman Apr 2002 B2
6379187 Nishimatsu Apr 2002 B2
6381575 Martin et al. Apr 2002 B1
6384737 Hsu et al. May 2002 B1
6393584 McLaren et al. May 2002 B1
6396480 Schindler et al. May 2002 B1
6397189 Martin et al. May 2002 B1
6407987 Abraham Jun 2002 B1
6408435 Sato Jun 2002 B1
6408437 Hendricks et al. Jun 2002 B1
6421651 Tedesco et al. Jul 2002 B1
6425125 Fries et al. Jul 2002 B1
6430537 Tedesco et al. Aug 2002 B1
6430738 Gross et al. Aug 2002 B1
6434678 Menzel Aug 2002 B1
6438450 DiLorenzo Aug 2002 B1
6442549 Schneider Aug 2002 B1
6446080 Van Ryzin et al. Sep 2002 B1
6446130 Grapes Sep 2002 B1
6449688 Peters et al. Sep 2002 B1
6470496 Kato et al. Oct 2002 B1
6473794 Guheen et al. Oct 2002 B1
6488508 Okamoto Dec 2002 B2
6490570 Numaoka Dec 2002 B1
6493871 McGuire et al. Dec 2002 B1
6496927 McGrane et al. Dec 2002 B1
6498855 Kokkosoulis et al. Dec 2002 B1
6522707 Brandstetter et al. Feb 2003 B1
6535911 Miller et al. Mar 2003 B1
6538558 Sakazume et al. Mar 2003 B2
6543052 Ogasawara Apr 2003 B1
6544122 Araki et al. Apr 2003 B2
6549719 Mankovitz Apr 2003 B2
D475029 Nathan et al. May 2003 S
6560651 Katz et al. May 2003 B2
6570507 Lee et al. May 2003 B1
6571282 Bowman-Amuah May 2003 B1
6577735 Bharat Jun 2003 B1
6578051 Mastronardi et al. Jun 2003 B1
6587403 Keller et al. Jul 2003 B1
6590838 Gerlings et al. Jul 2003 B1
6598230 Ballhorn Jul 2003 B1
6622148 Noble Sep 2003 B1
6622307 Ho Sep 2003 B1
6628939 Paulsen Sep 2003 B2
6629318 Radha et al. Sep 2003 B1
6643620 Contolini et al. Nov 2003 B1
6643690 Duursma et al. Nov 2003 B2
6654801 Mann et al. Nov 2003 B2
6658090 Harjunen et al. Dec 2003 B1
6662231 Drosset et al. Dec 2003 B1
6702585 Okamoto Mar 2004 B2
6724974 Naruto et al. Apr 2004 B2
6728824 Chen Apr 2004 B1
6728956 Ono Apr 2004 B2
6728966 Arsenault et al. Apr 2004 B1
6744882 Gupta et al. Jun 2004 B1
6751794 McCaleb et al. Jun 2004 B1
6755744 Nathan et al. Jun 2004 B1
6762585 Liao Jul 2004 B2
D495755 Wurz et al. Sep 2004 S
6789215 Rupp et al. Sep 2004 B1
6816578 Kredo et al. Nov 2004 B1
6850252 Hoffberg Feb 2005 B1
6898161 Nathan May 2005 B1
6904592 Johnson Jun 2005 B1
6920614 Schindler et al. Jul 2005 B1
6928653 Ellis et al. Aug 2005 B1
6934700 Ijdens et al. Aug 2005 B1
6942574 LeMay et al. Sep 2005 B1
6974076 Siegel Dec 2005 B1
7024485 Dunning et al. Apr 2006 B2
7073172 Chamberlain Jul 2006 B2
7103583 Baum et al. Sep 2006 B1
7107109 Nathan et al. Sep 2006 B1
7111129 Percival Sep 2006 B2
7114013 Bakke et al. Sep 2006 B2
7124194 Nathan et al. Oct 2006 B2
7160132 Phillips et al. Jan 2007 B2
7181458 Higashi Feb 2007 B1
7188352 Nathan et al. Mar 2007 B2
7195157 Swartz et al. Mar 2007 B2
7198571 LeMay et al. Apr 2007 B2
7205471 Looney et al. Apr 2007 B2
7206417 Nathan Apr 2007 B2
7210141 Nathan et al. Apr 2007 B1
7231656 Nathan Jun 2007 B1
7237198 Chaney Jun 2007 B1
7281652 Foss Oct 2007 B2
7293277 Nathan Nov 2007 B1
D560651 Berkheimer et al. Jan 2008 S
7347723 Daily Mar 2008 B1
D566195 Ichimura et al. Apr 2008 S
7356831 Nathan Apr 2008 B2
7406529 Reed Jul 2008 B2
7415707 Taguchi et al. Aug 2008 B2
7418474 Schwab Aug 2008 B2
7424731 Nathan et al. Sep 2008 B1
7430736 Nguyen et al. Sep 2008 B2
7433832 Bezos et al. Oct 2008 B1
7448057 Nathan Nov 2008 B1
7483958 Elabbady et al. Jan 2009 B1
7500192 Mastronardi Mar 2009 B2
7512632 Mastronardi et al. Mar 2009 B2
7519442 Nathan et al. Apr 2009 B2
7522631 Brown et al. Apr 2009 B1
7533182 Wurtzel et al. May 2009 B2
7549919 Nathan et al. Jun 2009 B1
7574727 Nathan et al. Aug 2009 B2
7634228 White et al. Dec 2009 B2
7647613 Drakoulis et al. Jan 2010 B2
7657910 McAulay et al. Feb 2010 B1
D616414 Nathan et al. May 2010 S
7749083 Nathan et al. Jul 2010 B2
7757264 Nathan Jul 2010 B2
7761538 Lin et al. Jul 2010 B2
7770165 Olson et al. Aug 2010 B2
7778879 Nathan et al. Aug 2010 B2
7783593 Espino Aug 2010 B2
7783774 Nathan et al. Aug 2010 B2
7793331 Nathan et al. Sep 2010 B2
7819734 Nathan et al. Oct 2010 B2
7822687 Brillon et al. Oct 2010 B2
D629382 Nathan et al. Dec 2010 S
7922178 Finocchio Apr 2011 B2
7937724 Clark et al. May 2011 B2
D642553 Nathan et al. Aug 2011 S
7992178 Nathan et al. Aug 2011 B1
7996873 Nathan et al. Aug 2011 B1
8015200 Seiflien et al. Sep 2011 B2
8028318 Nathan Sep 2011 B2
8032879 Nathan et al. Oct 2011 B2
8037412 Nathan et al. Oct 2011 B2
8052512 Nathan et al. Nov 2011 B2
8103589 Nathan et al. Jan 2012 B2
8151304 Nathan et al. Apr 2012 B2
8165318 Nathan et al. Apr 2012 B2
8214874 Nathan Jul 2012 B2
D665375 Garneau et al. Aug 2012 S
8292712 Nathan et al. Oct 2012 B2
8325571 Cappello et al. Dec 2012 B2
8332895 Nathan et al. Dec 2012 B2
8429530 Neuman et al. Apr 2013 B2
20010016815 Takahashi et al. Aug 2001 A1
20010023403 Martin et al. Sep 2001 A1
20010030660 Zainoulline Oct 2001 A1
20010030912 Kalis et al. Oct 2001 A1
20010037367 Iyer Nov 2001 A1
20010044725 Matsuda et al. Nov 2001 A1
20020002079 Martin et al. Jan 2002 A1
20020002483 Siegel et al. Jan 2002 A1
20020018074 Buil et al. Feb 2002 A1
20020032603 Yeiser Mar 2002 A1
20020040371 Burgess Apr 2002 A1
20020113824 Myers Aug 2002 A1
20020116476 Eyal et al. Aug 2002 A1
20020118949 Jones et al. Aug 2002 A1
20020120925 Logan Aug 2002 A1
20020123331 Lehaff et al. Sep 2002 A1
20020126141 Mastronardi Sep 2002 A1
20020129036 Ho Yuen Lok et al. Sep 2002 A1
20020162104 Raike et al. Oct 2002 A1
20030004833 Pollak et al. Jan 2003 A1
20030005099 Sven et al. Jan 2003 A1
20030006911 Smith et al. Jan 2003 A1
20030008703 Gauselmann Jan 2003 A1
20030014272 Goulet et al. Jan 2003 A1
20030018740 Sonoda et al. Jan 2003 A1
20030027120 Jean Feb 2003 A1
20030031096 Nathan et al. Feb 2003 A1
20030037010 Schmelzer Feb 2003 A1
20030041093 Yamane et al. Feb 2003 A1
20030050058 Walsh et al. Mar 2003 A1
20030064805 Wells Apr 2003 A1
20030065639 Fiennes et al. Apr 2003 A1
20030076380 Yusef et al. Apr 2003 A1
20030088538 Ballard May 2003 A1
20030093790 Logan et al. May 2003 A1
20030101450 Davidsson et al. May 2003 A1
20030104865 Itkis et al. Jun 2003 A1
20030108164 Laurin et al. Jun 2003 A1
20030135424 Davis et al. Jul 2003 A1
20030144910 Flaherty et al. Jul 2003 A1
20030163388 Beane Aug 2003 A1
20030176218 LeMay et al. Sep 2003 A1
20030191753 Hoch Oct 2003 A1
20030208586 Mastronardi et al. Nov 2003 A1
20030225834 Lee et al. Dec 2003 A1
20040010800 Goci Jan 2004 A1
20040025185 Goci et al. Feb 2004 A1
20040085334 Reaney May 2004 A1
20040103150 Ogdon et al. May 2004 A1
20040145477 Easter Jul 2004 A1
20040158555 Seedman et al. Aug 2004 A1
20040204220 Fried et al. Oct 2004 A1
20040205171 Nathan et al. Oct 2004 A1
20040220926 Lamkin et al. Nov 2004 A1
20040243482 Laut Dec 2004 A1
20050034084 Ohtsuki Feb 2005 A1
20050048816 Higgins Mar 2005 A1
20050060405 Nathan et al. Mar 2005 A1
20050073782 Nathan Apr 2005 A1
20050086172 Stefik Apr 2005 A1
20050111671 Nathan May 2005 A1
20050125833 Nathan et al. Jun 2005 A1
20050201254 Looney et al. Sep 2005 A1
20050267819 Kaplan Dec 2005 A1
20060018208 Nathan et al. Jan 2006 A1
20060031896 Pulitzer Feb 2006 A1
20060035707 Nguyen et al. Feb 2006 A1
20060062094 Nathan et al. Mar 2006 A1
20060143575 Sauermann Jun 2006 A1
20060227673 Yamashita et al. Oct 2006 A1
20060239131 Nathan Oct 2006 A1
20060293773 Nathan et al. Dec 2006 A1
20070025701 Kawasaki et al. Feb 2007 A1
20070047198 Crooijmans et al. Mar 2007 A1
20070086280 Cappello Apr 2007 A1
20070121430 Nathan May 2007 A1
20070139410 Abe et al. Jun 2007 A1
20070142022 Madonna et al. Jun 2007 A1
20070160224 Nathan Jul 2007 A1
20070204263 Nathan et al. Aug 2007 A1
20070209053 Nathan Sep 2007 A1
20070220052 Kudo et al. Sep 2007 A1
20070247979 Brillon et al. Oct 2007 A1
20080003881 Wu Jan 2008 A1
20080005698 Koskinen Jan 2008 A1
20080065925 Oliverio et al. Mar 2008 A1
20080066016 Dowdy et al. Mar 2008 A1
20080069545 Nathan et al. Mar 2008 A1
20080077962 Nathan Mar 2008 A1
20080086379 Dion et al. Apr 2008 A1
20080096659 Kreloff et al. Apr 2008 A1
20080137849 Nathan Jun 2008 A1
20080155588 Roberts et al. Jun 2008 A1
20080168807 Dion et al. Jul 2008 A1
20080171594 Fedesna et al. Jul 2008 A1
20080189155 Miloslavsky Aug 2008 A1
20080195443 Nathan et al. Aug 2008 A1
20080198271 Malki Aug 2008 A1
20080222199 Tiu et al. Sep 2008 A1
20080239887 Tooker et al. Oct 2008 A1
20080275771 Levine Nov 2008 A1
20080305738 Khedouri et al. Dec 2008 A1
20090002335 Chaudhri Jan 2009 A1
20090030802 Plotnick et al. Jan 2009 A1
20090037969 Nathan et al. Feb 2009 A1
20090042632 Guenster et al. Feb 2009 A1
20090063976 Bull et al. Mar 2009 A1
20090070341 Mastronardi et al. Mar 2009 A1
20090091087 Wasmund Apr 2009 A1
20090100092 Seiflein et al. Apr 2009 A1
20090109224 Sakurai Apr 2009 A1
20090129753 Wagenlander May 2009 A1
20090138111 Mastronardi May 2009 A1
20090165634 Mahowald Jul 2009 A1
20090168901 Yarmolich et al. Jul 2009 A1
20090172565 Jackson et al. Jul 2009 A1
20090177301 Hayes Jul 2009 A1
20090240721 Giacalone Sep 2009 A1
20090241061 Asai et al. Sep 2009 A1
20090265734 Dion et al. Oct 2009 A1
20090282491 Nathan Nov 2009 A1
20090287696 Galuten Nov 2009 A1
20090298577 Gagner et al. Dec 2009 A1
20090307314 Smith et al. Dec 2009 A1
20090328095 Vinokurov et al. Dec 2009 A1
20100042505 Straus Feb 2010 A1
20100042511 Sundaresan et al. Feb 2010 A1
20100131558 Logan et al. May 2010 A1
20100211818 Nathan et al. Aug 2010 A1
20100211872 Rolston Aug 2010 A1
20100241259 Nathan Sep 2010 A1
20100247081 Pons Sep 2010 A1
20100269066 Nathan Oct 2010 A1
20100299232 Nathan et al. Nov 2010 A1
20100306179 Lim Dec 2010 A1
20110055019 Coleman Mar 2011 A1
20110066943 Brillon et al. Mar 2011 A1
20110125599 Morin May 2011 A1
20110173521 Horton et al. Jul 2011 A1
20110246517 Nathan et al. Oct 2011 A1
20110270894 Mastronardi et al. Nov 2011 A1
20110283236 Beaumier et al. Nov 2011 A1
20110298938 Nathan et al. Dec 2011 A1
20110304685 Khedouri et al. Dec 2011 A1
20110321026 Nathan et al. Dec 2011 A1
20120009985 Nathan et al. Jan 2012 A1
20120053713 Nathan Mar 2012 A1
20120105464 Franceus May 2012 A1
20120143732 Nathan et al. Jun 2012 A1
20120150614 Dion et al. Jun 2012 A1
20120158531 Dion et al. Jun 2012 A1
20120166965 Nathan et al. Jun 2012 A1
20120240140 Nathan Sep 2012 A1
20120323652 Mastronardi et al. Dec 2012 A1
20130021281 Tse et al. Jan 2013 A1
20130040715 Nathan et al. Feb 2013 A1
20130044995 Cappello et al. Feb 2013 A1
20130070093 Rivera et al. Mar 2013 A1
20130091054 Nathan et al. Apr 2013 A1
20140026154 Nathan Jan 2014 A1
Foreign Referenced Citations (146)
Number Date Country
199954012 Apr 2000 AU
2119184 Sep 1994 CA
1340939 Mar 2002 CN
3406058 Aug 1985 DE
3723737 Jan 1988 DE
3820835 Jan 1989 DE
3815071 Nov 1989 DE
4244198 Jun 1994 DE
19539172 Sep 1996 DE
19610739 Sep 1997 DE
19904007 Aug 2000 DE
0082077 Jun 1983 EP
0140593 May 1985 EP
0256921 Feb 1988 EP
0283304 Sep 1988 EP
0283350 Sep 1988 EP
0309298 Mar 1989 EP
0313359 Apr 1989 EP
0340787 Nov 1989 EP
0363186 Apr 1990 EP
0425168 May 1991 EP
0464562 Jan 1992 EP
0480558 Apr 1992 EP
0498130 Aug 1992 EP
0507110 Oct 1992 EP
0529834 Mar 1993 EP
0538319 Apr 1993 EP
0631283 Dec 1994 EP
0632371 Jan 1995 EP
0711076 May 1996 EP
0786122 Jul 1997 EP
0817103 Jan 1998 EP
0841616 May 1998 EP
0919964 Jun 1999 EP
0959570 Nov 1999 EP
0974896 Jan 2000 EP
0974941 Jan 2000 EP
0982695 Mar 2000 EP
1001391 May 2000 EP
1170951 Jan 2002 EP
1288802 Mar 2003 EP
1408427 Apr 2004 EP
1549919 Apr 2004 EP
1962251 Aug 2008 EP
2602352 Feb 1988 FR
2808906 Nov 2001 FR
2122799 Jan 1984 GB
2166328 Apr 1986 GB
2170943 Aug 1986 GB
2193420 Feb 1988 GB
2238680 Jun 1991 GB
2254469 Oct 1992 GB
2259398 Mar 1993 GB
2262170 Jun 1993 GB
2380377 Apr 2003 GB
2505584 Aug 2014 GB
57173207 Oct 1982 JP
58-179892 Oct 1983 JP
60-253082 Dec 1985 JP
61084143 Apr 1986 JP
62-192849 Aug 1987 JP
62-284496 Dec 1987 JP
63-60634 Mar 1988 JP
2-153665 Jun 1990 JP
5-74078 Mar 1993 JP
5122282 May 1993 JP
06-127885 May 1994 JP
07281682 Oct 1995 JP
07-311587 Nov 1995 JP
08-037701 Feb 1996 JP
8274812 Oct 1996 JP
08279235 Oct 1996 JP
08289976 Nov 1996 JP
928918 Feb 1997 JP
9114470 May 1997 JP
9127964 May 1997 JP
09-244900 Sep 1997 JP
10-098344 Apr 1998 JP
10-222537 Aug 1998 JP
11-003088 Jan 1999 JP
11-024686 Jan 1999 JP
11-095768 Apr 1999 JP
2002-83640 Mar 2002 JP
2002-537584 Nov 2002 JP
2003-076380 Mar 2003 JP
2003-084903 Mar 2003 JP
2003-099072 Apr 2003 JP
2005-107267 Apr 2005 JP
2005-184237 Jul 2005 JP
2006-048076 Feb 2006 JP
2007-034253 Feb 2007 JP
2007-041722 Feb 2007 JP
2007505410 Mar 2007 JP
07504517 Mar 2007 JP
2007-102982 Apr 2007 JP
2007-104072 Apr 2007 JP
2007-128609 May 2007 JP
2007-164078 Jun 2007 JP
2007-164298 Jun 2007 JP
2007179333 Jul 2007 JP
2007-241748 Sep 2007 JP
2008-058656 Mar 2008 JP
2009-017529 Jan 2009 JP
2009-075540 Apr 2009 JP
514511 Dec 2002 TW
M274284 Sep 2005 TW
M290206 May 2006 TW
WO 8601326 Feb 1986 WO
WO 9000429 Jan 1990 WO
WO 9007843 Jul 1990 WO
WO 9108542 Jun 1991 WO
WO 9120082 Dec 1991 WO
WO 9316557 Aug 1993 WO
WO 9318465 Sep 1993 WO
WO93021732 Oct 1993 WO
WO 9403894 Feb 1994 WO
WO 9414273 Jun 1994 WO
WO 9415306 Jul 1994 WO
WO 9415416 Jul 1994 WO
WO 9503609 Feb 1995 WO
WO 9529537 Nov 1995 WO
WO 9612255 Apr 1996 WO
WO 9612256 Apr 1996 WO
WO 9612257 Apr 1996 WO
WO 9612258 Apr 1996 WO
WO 9807940 Feb 1998 WO
WO 9811487 Mar 1998 WO
WO 9845835 Oct 1998 WO
WO 9935753 Jul 1999 WO
WO 0100290 Jan 2001 WO
WO 0108148 Feb 2001 WO
WO 0171608 Sep 2001 WO
WO 02060546 Aug 2002 WO
WO 02095752 Nov 2002 WO
WO 01084353 Jan 2003 WO
WO 2003005743 Jan 2003 WO
WO 03069613 Aug 2003 WO
WO 2004029775 Apr 2004 WO
WO 2005026916 Mar 2005 WO
WO 2006014739 Feb 2006 WO
WO 2006056933 Jun 2006 WO
WO 2006138064 Dec 2006 WO
WO 2007092542 Aug 2007 WO
WO 2008-033853 Mar 2008 WO
WO 2011094330 Aug 2011 WO
WO 2013040603 Mar 2013 WO
Non-Patent Literature Citations (70)
Entry
Office Action dated Oct. 1, 2013 in corresponding Japanese Application No. 2012-500786 (with translation).
Yuki Murata, iTunes no ‘Kankyo Settei’ Catalog & Tips 10 Sen, Mac People, ASCII Corporation, Oct. 1, 2007, vol. 13, No. 10, pp. 147.
“About Ecast”, date unknown, leaflet.
Austin Cyber Limits: Name That Tune [online], [retrieved Jul. 23, 2001]. Retrieved from the Internet: <http://www.pbs.ork/klru/austin/games/namethattune.html>.
Back to the Tunes [online], [retrieved Jul. 23, 2001]. Retrieved from the Internet: <http://citc5.hispeed.com/rules.html>.
Bonczek et al, “The DSS Development System”, 1983 National Computer Conference, Anaheim, California, May 16-19, 1983, pp. 441-455.
Chan et al., “Distributed servers architectures for networked video services”, IEEE Trans on Networking, vol. 9, No. 2, pp. 125-136, 2001.
Chen et al., “Optimization of the grouped sweeping scheduling (GSS) with heterogeneous multimedia streams”, ACM Multimedia, pp. 1-7, 1993.
Crutcher et al., “The networked video Jukebox”, IEEE, Trans. on circuits and systems for video technology, vol. 4, No. 2, pp. 105-120, 1994.
“Darts Revolution Again”, Replay Magazine, Mar. 1991, pp. 146-148.
Decision of the European Patent Office to revoke the related EP Patent No. 786 125, dated Feb. 17, 2005.
Derfler et al., “How Networks Work”, Millennium Ed., Que Corporation, Jan. 2000.
Drews, C.; Pestoni, F.; “Virtual jukebox: reviving a classic,” Proceedings of the 35th Annual Hawaii International Conference System Sciences, pp. 887-893, Jan. 7-10, 2002.
“Ecast Forges Landmark International Technology Partnership”, Business Wire at www.findarticles.com/cf_0/m0EIN/2000_July_25/63663604/print.html, 2 pages, Jul. 25, 2000.
“Ecast Selects Viant to Build Siren Entertainment System (TM)”, ScreamingMedia, PR Newswire San Francisco, industry.java.sum.com/javanews/stories/story2/0,1072,17618,00.html, 3 pages, Aug. 3, 1999.
European Search Report from EP 1 993 079.
European Search Report issued for European Application No. 08000845.1-1238/1962251, dated Apr. 3, 2009.
Fachbuch, “Unterhaltungselektronic von A-Z” gfu 1, VDE—Verlag GmbH, pp. 12-13, 1983-1984.
Gallardo et al., “Tangible Jukebox: back to palpable music”, ACM TEI, pp. 199-202, 2010.
Gralla, “How the Internet Works”, Millennium Ed., Que Corporation, Aug. 1999.
Grimes, Chapter 18, “Taking Advantage of Web-based Audio”.
Hicks et al., “Dynamic software updating”, ACM PLDI, pp. 13-23, 2001.
IBM Technical Disclosure Bulletin, vol. 30, No. 5, Oct. 1987, “Method for Automated Assembly of Software Versions”, pp. 353-355.
IBM Technical Disclosure Bulletin, vol. 32, No. 9A, Feb. 1990, “Robotic Wafer Handling System for Class 10 Environments” pp. 141-143.
IBM Technical Disclosure Bulletin, vol. 33, No. 12, May 1991, “High-speed Opens and Shorts Substrate Tester”, pp. 251-259.
IBM Technical Disclosure Bulletin, vol. 41, No. 1, Jan. 1998, “Safe Mechanism for Installing Operating System Updates with Applications,” pp. 557-559.
International Search Report for PCT/US2009/003998, dated Aug. 17, 2009.
International Search Report for PCT/US2011/22598, dated Mar. 29, 2011.
iTouch 8 Plus brochure, JVL Corporation, 2005, 2 pages.
iTOUCH 27 New Games brochure, JVL Corporation, 2005, 2 pages.
Johnny Rockets Name That Tune [online], [retrieved Mar. 7, 2002]. Retrieved from the Internet: <http://www.johnnyrockets.com/docs/funstuff.html>.
Koskelainem, “Report on Streamworks™”.
Kozierok, The PC Guide, Site Version 2.2.0, http://www.pcguide.com, Apr. 17, 2001.
Kraiss et al., “Integrated document caching and prefetching in storage hierarchies based on Markov chain predictions”, The VLDB Journal, vol. 7, issue 3, pp. 141-162, 1998.
Liang et al., “Dynamic class loading in the Java virtual machine”, ACM OOPSLA, pp. 36-44, 1998.
Look and iTouch brochure, JVL Corporation, 2004, 2 pages.
Ludescher et al., “File Storage Management for TFTF physics data”, IEEE, pp. 856-859, 1992.
Megatouch Champ brochure, Merit Industries, Inc., 2005, 2 pages.
Melnik et al., “A mediation infrastructure for digital library services”, ACM DL, pp. 123-132, 2000.
Merriam Webster's Collegiate Dictionary, Tenth Edition, Merriam-Webster, Inc., p. 361 (definition of dynamically).
Mickey B's Jukebox Revue—Name That Tune! [online], [retrieved Jul. 23, 2001]. Retrieved from the Internet: <http://mickeyb.com/tune/>.
Mod Box Internet brochure, Merit Entertainment, 2006, 2 pages.
Newsome et al., “Proxy compilation of dynamically loaded java classes with MoJo”, ACM LCTES, pp. 204-212, Jun. 2002.
Outlaw, Computer Technology Review, “Virtual Servers Offer Performance Benefits for Network Imaging”, 1993.
Patent Abstract of Japan vol. 95, No. 010 & JP 07 281682 A (Naguo Yuasa), Oct. 27 1 JP 07 281682, figure 1-6 abrége.
Pohlmann, “Principles of Digital Audio”, Third Edition, 1995.
PR Newswire, Press Release, “MusicMatch Announces Commercial Availability of Meta Trust Certified MusicMatch jukebox”, New York; Nov. 15, 1999, extracted from Internet, http://proquest.umi.com on Sep. 17, 2002.
Rollins et al., “Pixie: A jukebox architecture to support efficient peer content exchange”, ACM Multimedia, pp. 179-188, 2002.
Schneier, “Applied Cryptography”, Second Edition, John Wiley & Sons, Inc. New York, 1996.
Sprague et al., “Music selection using the partyvote democratic Jukebox”, ACM AVI, pp. 433-436, 2008.
Stevens, “TCP/IP Illustrated: vol. 1, the Protocols”.
Stewart, “Ecast Deploys Marimba's Castanet to Power an Internet-Based, Entertainment Management System for the Out-of-Home Market”, Marimba, Press Release, 3 pages, www.marimba.com/news/releases/ecast.dec13.html, Dec. 13, 1999.
Strauss et al., “Information Jukebox A semipublic device for presenting multimedia information content”, Pers. Ubiquit Comput, 7, pp. 217-220, 2003.
Summary of the oral proceedings regarding EP 786 125 before the Opposition Division of the European Patent Office, dated Feb. 17, 2005.
Tom & Liz's Name That Tune [online], [retrieved Jul. 23, 2001]. Retrieved from the Internet: <http://home.att.net/˜tomnliz/Music.html>.
U.S. Appl. No. 29/371,355, Garneau et al., filed Dec. 14, 2010.
U.S. Appl. No. 61/129,637, Dion, filed Jul. 9, 2008.
U.S. Appl. No. 61/202,617, Dion, filed Mar. 18, 2009.
Vortex Brochure, JVL Corporation, 2005, 2 pages.
Waingrow, “Unix Hints & Hacks”, Que Corporation, Indianapolis, IN, 1999.
White, “How Computers Work”, Millennium Ed., Que Corporation, Indianapolis, IN, Sep. 1999 (Sep. 22, 1999).
Search Report and Written Opinion dated Sep. 19, 2014 in corresponding International Application No. PCT/US14/24648.
Ahanger et al.; A Digital On-Demand Video Service Supporting Content-Based Queries; 1993; 9 pages.
“Foobar 2000 Evaluation Updated,” MonkeyBiz, Aug. 3, 2008, 4 pages (with partial English translation). http://monkeybizinfo.blogspot.jp/2008/08/foobar2000.html.
Hewlett-Packard Development Co; HP Open View Storage Data Protector Admin's Guideline Manual Edition; May 2003; Copyright 2003, 60 pages, http://h20000.www2.hp.com/bc/docs/support/SupportManual/c006637931/c00663793.pdf.
Peter Pawlowski, “Basic Player Whose Appearance and Functions can be Customized Freely ‘Foobar 2000’ v1.0 is Unveiled,” Windows Forest, Japan, Jan. 12, 2010, 3 pages (with partial English translation). http://forest.impress.co.jp/docs/news/20100112_341870.html.
Merriam Webster's Collegiate Dictionary, Ninth Edition, Merriam Webster, Inc., p. 1148, 1986 (definition of “Stand”).
Patent Examination Report No. 3 Australian Patent Application No. 2015203639 dated Jun. 10, 2016.
Office Action in related U.S. Appl. No. 12/929,466 dated Apr. 14, 2016.
Office Action in related U.S. Appl. No. 14/729,392 dated May 25, 2017.
Related Publications (1)
Number Date Country
20180239503 A1 Aug 2018 US
Provisional Applications (5)
Number Date Country
61584750 Jan 2012 US
61536015 Sep 2011 US
61202617 Mar 2009 US
61431036 Jan 2011 US
61298509 Jan 2010 US
Continuations (3)
Number Date Country
Parent 15074863 Mar 2016 US
Child 15962176 US
Parent 13833173 Mar 2013 US
Child 15074863 US
Parent 12929466 Jan 2011 US
Child 13833173 US
Continuation in Parts (2)
Number Date Country
Parent 13621922 Sep 2012 US
Child 13833173 US
Parent 13138660 US
Child 13621922 US