The current invention is related to searching media content, including video files with audio tracks, audio tracks, and other types of media content that include data that can be fully or partially transcribed to produce a text transcript, to identify portions of the media content related to search-query terms and phrases, and, in particular, to a concept-service component of a content-search-service system that employs an ontology and a text transcript to score the transcript for use by a search-service component of the content-search-service to render search results to a client of the content-search-service system.
In the early days of computing, information was generally encoded as formatted alphanumeric character strings or as unformatted ordered sequences of information-storage units, typically bytes. As computing hardware, operating systems, and computer applications have together evolved, many different types of information encodings have been developed, and a very large variety of different types of information are now routinely electronically encoded, stored, exchanged, and rendered for access by users, including text files, specially formatted application-specific files, audio recordings, video recordings, and multimedia presentations. While, in early days of computing, data was primarily rendered as character strings displayed on monochromatic, 24-line terminals, the many different types of information currently electronically encoded and distributed by computer systems are rendered for presentation to human users by a variety of different application programs, including text and image editors, video players, audio players, and web browsers.
An important class of information comprises information encoded as an ordered sequence of information units that are sequentially rendered for display or presentation to a human user. An MPEG-encoded video file is one example of a sequentially ordered information encoding. MPEG encoding employs a number of rather complex layers of different types of encoding methods to compactly encode a video stream and/or audio stream. In general, video frames are reconstructed from an MPEG-encoded video file frame-by-frame, in sequence. Rendering of an MPEG-encoded video file provides a stream of video frames and an audio stream. Rendering applications and devices generally allow a user to start or resume rendering of the video file, to stop rendering of the video file, and to skip forward or backward to selected positions within a video stream.
In many cases, a user may only be interested in a certain portion of a video presentation. For example, a particular user may be interested only in a weather report included in a local television news broadcast that includes reviews of current local and national events, reviews of sporting events, and presentations of human-interest stories in addition to the weather report. In many cases, video presentations may not be indexed by sections, in order to facilitate direct access to portions of the video presentation of interest to a user, or may be indexed at a very coarse topic granularity, requiring a user to employ a hit-or-miss strategy of starting, stopping, advancing, and reversing the video stream by relatively crude techniques in order to locate portions of interest. In addition to encoded video, there are many other types of sequentially ordered information encodings that are sequentially rendered for human perception, including pure audio recordings, various types of multimedia presentations, images of pages within books and text documents, and other such information encodings. In many cases, searching for portions of the encoded information of interest to human users is currently limited to the above-described stop/start/advance/and reverse operations familiar to users of video-rendering applications and many video-signal-rendering devices.
Designers and manufacturers of computers and other electronic devices that render sequentially ordered information encodings for presentation to human users, designers, implementers, vendors and users of information-rendering applications, including media players, web browsers, and control programs, and many others involved in recording, disseminating, and rendering information have recognized the need for more effective searching tools to allow users to identify and efficiently access portions of an information encoding of interest to those to whom the information is rendered. In response to these needs, a content-search-service system has been developed. Various embodiments of the present invention comprise components and subsystems of this content-search-service system. The content-search-service system receives and/or locates and retrieves various content items electronically available to clients of the content-search-service system and prepares internal representations of the content items, or portions of the content items, to enable the content-search-service to graphically render search results generated by the content-search-service system in response to search requests made by clients of the content-search-service system. Designers, developers, and manufacturers of content-search-service systems, as well as content-search-service providers and users of content-search-service systems and services provided by content-search services, have all recognized the need for efficient and accurate content-search-service components to facilitate rapid and accurate responses to search requests directed to content items received from clients of content-search services that employ content-search-service systems.
Various embodiments of the present invention include concept-service components of content-search-service systems which employ ontologies and vocabularies prepared for particular categories of content at particular times in order to score transcripts prepared from content items to enable a search-service component of a content-search-service system to assign estimates of the relatedness of portions of a content item to search criteria in order to render search results to clients of the content-search-service system. The concept-service component processes a search request to generate lists of related terms, and then employs the lists of related terms to process transcripts in order to score transcripts based on information contained in the ontologies.
Embodiments of the present invention are employed within content-search-service systems that are used to provide content-search services to clients.
In general, the content rendered to the user is a type of content that, unlike a text file, cannot be easily searched using commonly available search tools, such as search facilities provided within text editors, in order to find portions of the content of particular interest to the user. In the discussion that follows, the content is assumed to be a video file with an accompanying audio track, such as a news broadcast or sports broadcast provided by a news service or sports service through the Internet to accessing users. However, content-search-service systems, in which embodiments of the current invention may be implemented and employed, may provide search services for a variety of different types of content, from multi-media presentations to various types of images, graphics, and musical selections.
In general, the content provider 106 provides content items to the content-search-service system 112 for preprocessing, to facilitate rapid responses to subsequent client search requests directed to the content items. However, in alternative implementations, the content-search-service system may concurrently receive the content item and search phrase or search term from a client, process the content item to prepare for searching the content, carry out the search request, and render results to the user in real time. In general, client systems are distinct from both content-provider systems and content-search-service systems, although it is possible that the content provider, client, and content-search service may all execute concurrently or simultaneously within a single computer system or distributed computer system.
In some embodiments, a determination of the “hot spots” in the video clip can be made. The hot spots can include positions in which content related to a search criterion are predominant. In some instances, the hot spots are identified by generating ontology-based scores for each portion of a set of portions of the video clip (e.g., based on an associated transcript), and determining portions associated with high ontology-based scores (e.g., scores above a threshold). A second video may be transcoded from the original video which is solely inclusive of the “hot spots”. Subsequently, an automatic summary of the video into a condensed abstract form that includes all the snippets that are considered to be relevant may be generated. Furthermore, fade of, for example, 10, seconds before each snippet and 10, seconds after each snippet may be used when putting the snippets in the abridged video file together. Alternatively, an increase or decrease of time may be used depending on the length of the snippets and the number of snippets being used. Furthermore, fade can be skipped for certain snippets to produce larger snippets. Additionally, multiple videos may be used to create an abridged video compiled from the multiple videos. Further details about creation of an abridged video are disclosed in U.S. application Ser. No. 13/245,843,, U.S. application Ser. No. 13/245,855,, U.S. application Ser. No. 13/312,373,, and U.S. application Ser. No. 13/245,850,, each of which is incorporated by reference in its entirety for all purposes.
In some embodiments, a heat-map display is generated, not in response to a user's search specific criterion but instead may be based on words or phrases, e.g., frequently occurring in a transcript, that are determined to be important (e.g., based on an analysis of how rare and spread out words and group of words are), and/or candidate concepts for a video dip (e.g., based on intersecting a transcript and ontology for a category of the video clip. Further details about these types of words or phrases are disclosed in U.S. application Ser. No. 13/245,843,, U.S. application Ser. No. 13/245,855,, U.S. application Ser. No. 13/312,373,, and U.S. application Ser. No. 13/245,850,, each of which is incorporated by reference in its entirety for all purposes.
In some embodiments, a heat-map display is generated in response to a search criterion entered while searching across a set of videos. For example, a keyword and related concepts can be used to obtain an initial list of videos where either a) a search of the keyword matches the transcript of a video file or b) the keyword concept matches top concepts associated the video file. Further, use of a “search inside” technique on each video in the initial list can be used to generate a heat map vector. In one embodiment, the “search inside” techniques allows for the searching of words and terms within a video file to produce a display of the locations of each occurrence of the searched words or terms. Furthermore, use of the vector to generate a new search score for each video can be accomplished. As such, a re-ranking and return of videos according to the new search score occurs. Then, a compilation of all the hot spots from multiple videos can be generated, and could then be added to the transcript OCR from video such as credits, signs, title, etc.
In some embodiments, the heat map is generated based on a targeted transcript associated with the video clip. A targeted transcript may be generated by enhancing a transcript based, e.g., on category of the video file to produce a more targeted transcript. For example, because of variations in speech, inflection, accents, etc., producing an accurate transcript of an audio file can be difficult. However, if the category of the video file was known to the system prior to generating a transcript, a more accurate and more intelligent transcript would be produced. In one example, if the category is “a football telecast”, then the system can be tuned to be listening for terms associated with football (i.e., touchdown, pass, quarterback, etc.). Whereas, if the system identified ‘touch’ and ‘down’ in succession, then the system could make the determination that the speaker is actually saying ‘touchdown’ as opposed to two separate words, thus producing a more accurate transcript (hence a ‘targeted’ transcript).
The CSS service additionally includes a content-service component 430 that receives content items from remote content providers, or that searches for, finds, and retrieves content items provided by content providers, and furnishes the content items to a content-processor component 432 that prepares and stores one or more transcripts 426 for each processed content item in the extracted-content-data component 416. The content processor 432 accesses a language model, such as language model 434, stored in a language-model store 436, in order to process a given content item. The content-processor component 432 also deposits additional information about content items in the media-data-service component 418. In the described embodiments of the present invention, transcripts are text-based transcripts of audio tracks and audio files, carried out by automatic-speech-recognition subcomponents of the content processor component. In alternative embodiments of the present invention, text transcripts may be prepared from other types of media content, including descriptive transcripts of still or moving images prepared by computer-vision subcomponents of the content processor component.
An information-aggregator-and-classifier component 440 continuously, or at intervals, searches through information available on the Internet and other information sources for documents, text files, and other information items related to various different categories to which content items may be assigned. The information-aggregator-and-classifier component 440 classifies those information items deemed to be useful to the CSS system by category, and stores the information items, for each category and for particular ranges of dates and times, into a categorized-information-storage component 442. These information items are processed by the information-aggregator-and-classifier component to remove unnecessary information, linguistically normalize terms and phrases, and compute various parameters and values associated with the information items that are used both b the information-aggregator-and-classifier component to classify the items as well as by the language-model-builder component 444 and ontology-builder component 446, which use the information items stored in the categorized-information-storage component 442 to build language models and ontologies, respectively.
Each ontology is physically or conceptually associated with a vocabulary. The vocabulary is also prepared from information items collected by the information-aggregator-and-classifier component (440 in
A vocabulary comprises a list of nouns, or noun phrases, in one embodiment of the present invention, that commonly occur in information items related to a particular information category. For example, a sports category for content items might be expected to include nouns such as “bat,” “base,” “pitcher,” “quarterback,” “goalpost,” “football,” “javelin,” “roller derby,” and other such nouns and noun phrases. Because it is inefficient to programmatically manipulate strings of symbols, such as character strings, when implementing components of the CSS system, each term or phrase in a vocabulary is represented by an integer value.
While it is convenient to represent an ontology as a graph that includes term and phrase nodes interconnected by arcs, as shown in
The m×m representation of an ontology, shown in
In practice, even the list representation of an ontology, shown in
In the current discussion, content items are video files that include audio tracks. In one embodiment of the present invention, searching is carried out by the CSS system exclusively on the audio track of a video file, using terms and phrases input by a user to find those terms or phrases, or related terms and phrases, which occur at points in time in the audio track. Thus, portions of the audio track can be identified as being related to search terms and of particular interest to a user. Those portions of the audio track can, in turn, be related to the video images that are displayed in the interval of time in which the portions of the audio track are rendered, when the video file is presented to the user by a video-file-rendering application. In these embodiments, a transcript (426 in
Although the present invention has been described in terms of particular embodiments, it is not intended that the invention be limited to these embodiments. Modifications within the spirit of the invention will be apparent to those skilled in the art. For example, the CS component of a CSS system can be implemented in any number of different programming languages for execution on any number of different operating systems running on different hardware platforms within many different types of CSS systems. Implementations of the CS component may vary according to variations in familiar programming parameters and characteristics, including control structures, data structures, modular organization, and other familiar parameters and characteristics. As discussed above, many different types of ontologies and ontology representations, and many different types of transcripts, and transcript representations, may be employed by various embodiments of the CS component to prepare scored transcripts. Co-occurrence metrics and other numeric values may have different ranges and representations, in alternative embodiments.
The foregoing description, for purposes of explanation, used specific nomenclature to provide a thorough understanding of the invention. However, it will be apparent to one skilled the art that the specific details are not required in order to practice the invention. The foregoing descriptions of specific embodiments of the present invention are presented for purpose of illustration and description. They are not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The embodiments are shown and described in order to best explain the principles of the invention and its practical applications, to thereby enable others skilled in the art to best utilize the invention and various embodiments with various modifications as are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the following claims and their equivalents:
This application is a continuation-in-part of application Ser. No. 12/077,590,, filed Oct. 5, 2009,, which is a continuation-in-part of U.S. application Ser. No. 11/903,279,, filed Sep. 21, 2007. This application is also a continuation-in-part of U.S. application Ser. No. 13/048,268,, filed Mar. 15, 2011,, which is a continuation of U.S. application Ser. No. 12/077,591,, filed Mar. 19, 2008,, which is a continuation-in-part of U.S. application Ser. No. 11/903,279,, filed Sep. 21, 2007. This application is further a continuation-in-part of U.S. application Ser. No. 11/903,279,, filed Sep. 21, 2007,, which claims the benefit and priority of U.S. Provisional Application No. 60/846,484,, filed Sep. 22, 2006. This application is also a continuation-in-part of U.S. application Ser. No. 13/245,840,, filed Sep. 26, 2011,, which is a continuation-in-part of U.S. application Ser. No. 11/903,279,, filed Sep. 21, 2007,, which claims the benefit and priority of U.S. Provisional Application No. 60/846,484. This application is also a continuation-in-part of U.S. application Ser. No. 13/245,843,, filed Sep. 26, 2011. This application is also a continuation in part of U.S. application Ser. No. 13/245,855,, filed Sep. 26, 2011. This application is also a continuation-in-part of U.S. application Ser. No. 13/312,373,, filed Dec. 6, 2011,, which is a continuation of U.S. application Ser. No. 13/245,850,, filed Sep. 26, 2011. Each of the above-listed applications is hereby incorporated by reference in its entirety for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
5892536 | Logan et al. | Apr 1999 | A |
6070158 | Kirsch et al. | May 2000 | A |
6192360 | Dumais et al. | Feb 2001 | B1 |
6286006 | Bharat et al. | Sep 2001 | B1 |
6411725 | Rhoads | Jun 2002 | B1 |
6523000 | Ando et al. | Feb 2003 | B1 |
6546556 | Kataoka et al. | Apr 2003 | B1 |
6553365 | Summerlin et al. | Apr 2003 | B1 |
6636238 | Amir et al. | Oct 2003 | B1 |
6665640 | Bennett et al. | Dec 2003 | B1 |
6665659 | Logan | Dec 2003 | B1 |
6675159 | Lin et al. | Jan 2004 | B1 |
6681227 | Kojima et al. | Jan 2004 | B1 |
6700932 | Shen et al. | Mar 2004 | B2 |
6714909 | Gibbon et al. | Mar 2004 | B1 |
6721361 | Covell et al. | Apr 2004 | B1 |
6782394 | Landeck et al. | Aug 2004 | B1 |
6877002 | Prince | Apr 2005 | B2 |
6901398 | Horvitz et al. | May 2005 | B1 |
7073193 | Marsh | Jul 2006 | B2 |
7082162 | Prakash et al. | Jul 2006 | B2 |
7096208 | Zaragoza et al. | Aug 2006 | B2 |
7185008 | Kawatani | Feb 2007 | B2 |
7206303 | Karas et al. | Apr 2007 | B2 |
7210100 | Berger et al. | Apr 2007 | B2 |
7257577 | Fagin et al. | Aug 2007 | B2 |
7260773 | Zernik | Aug 2007 | B2 |
7281260 | Puente et al. | Oct 2007 | B2 |
7490092 | Sibley et al. | Feb 2009 | B2 |
7519565 | Prakash et al. | Apr 2009 | B2 |
7542969 | Rappaport et al. | Jun 2009 | B1 |
7606444 | Erol et al. | Oct 2009 | B1 |
7650572 | Olander et al. | Jan 2010 | B2 |
7673234 | Kao et al. | Mar 2010 | B2 |
7689578 | Albornoz et al. | Mar 2010 | B2 |
7711193 | Sanford | May 2010 | B2 |
7743061 | Jones et al. | Jun 2010 | B2 |
7752186 | Abajian | Jul 2010 | B2 |
7917492 | Bargeron et al. | Mar 2011 | B2 |
7930714 | Konig et al. | Apr 2011 | B2 |
8065604 | Blankinship | Nov 2011 | B2 |
8073194 | Lienhart et al. | Dec 2011 | B2 |
8098976 | Acharya | Jan 2012 | B1 |
8117206 | Sibley et al. | Feb 2012 | B2 |
20020022956 | Ukrainczyk et al. | Feb 2002 | A1 |
20020120925 | Logan | Aug 2002 | A1 |
20020164136 | Fukuda et al. | Nov 2002 | A1 |
20030030752 | Begeja et al. | Feb 2003 | A1 |
20030037073 | Tokuda et al. | Feb 2003 | A1 |
20030061028 | Dey et al. | Mar 2003 | A1 |
20030103645 | Levy et al. | Jun 2003 | A1 |
20030126214 | Oliszewski | Jul 2003 | A1 |
20030229604 | Zaragoza et al. | Dec 2003 | A1 |
20040093564 | Kumhyr et al. | May 2004 | A1 |
20040102958 | Anderson | May 2004 | A1 |
20040162827 | Nakano | Aug 2004 | A1 |
20040186743 | Cordero | Sep 2004 | A1 |
20040199547 | Winter et al. | Oct 2004 | A1 |
20040230892 | Horton | Nov 2004 | A1 |
20050004897 | Lipson et al. | Jan 2005 | A1 |
20050021502 | Chen et al. | Jan 2005 | A1 |
20050022106 | Kawai | Jan 2005 | A1 |
20050028194 | Elenbaas et al. | Feb 2005 | A1 |
20050086224 | Franciosa et al. | Apr 2005 | A1 |
20050138540 | Baltus et al. | Jun 2005 | A1 |
20050198006 | Boicey et al. | Sep 2005 | A1 |
20050222975 | Nayak et al. | Oct 2005 | A1 |
20050246739 | Davidson | Nov 2005 | A1 |
20050262050 | Fagin et al. | Nov 2005 | A1 |
20050262432 | Wagner | Nov 2005 | A1 |
20050268221 | Shur et al. | Dec 2005 | A1 |
20060031217 | Smith et al. | Feb 2006 | A1 |
20060047632 | Zhang | Mar 2006 | A1 |
20060075205 | Martin et al. | Apr 2006 | A1 |
20060110057 | Liu et al. | May 2006 | A1 |
20060114992 | Shibata et al. | Jun 2006 | A1 |
20060155693 | Chowdhury et al. | Jul 2006 | A1 |
20060183089 | Gleissner et al. | Aug 2006 | A1 |
20060184512 | Kohanim et al. | Aug 2006 | A1 |
20060195859 | Konig et al. | Aug 2006 | A1 |
20060248458 | Li | Nov 2006 | A1 |
20060288015 | Schirripa et al. | Dec 2006 | A1 |
20070061728 | Sitomer et al. | Mar 2007 | A1 |
20070118391 | Malaney et al. | May 2007 | A1 |
20070130007 | Haberman et al. | Jun 2007 | A1 |
20070168413 | Barletta et al. | Jul 2007 | A1 |
20070192192 | Haberman et al. | Aug 2007 | A1 |
20070201558 | Xu et al. | Aug 2007 | A1 |
20070203885 | Kim et al. | Aug 2007 | A1 |
20070219778 | Whittaker et al. | Sep 2007 | A1 |
20070226204 | Feldman | Sep 2007 | A1 |
20070245242 | Yagnik | Oct 2007 | A1 |
20080086453 | Fabian | Apr 2008 | A1 |
20080120290 | Delgo et al. | May 2008 | A1 |
20080126319 | Bukai et al. | May 2008 | A1 |
20090083256 | Thompson et al. | Mar 2009 | A1 |
20090083257 | Bargeron et al. | Mar 2009 | A1 |
20090094189 | Stephens | Apr 2009 | A1 |
20090094520 | Kulas | Apr 2009 | A1 |
20090116736 | Neogi et al. | May 2009 | A1 |
20090174526 | Howard et al. | Jul 2009 | A1 |
20090182644 | Panagopulos et al. | Jul 2009 | A1 |
20090210395 | Sedam | Aug 2009 | A1 |
20090222442 | Houh | Sep 2009 | A1 |
20090254540 | Musgrove et al. | Oct 2009 | A1 |
20090271380 | Julia et al. | Oct 2009 | A1 |
20090287676 | Dasdan | Nov 2009 | A1 |
20100131336 | Steelberg et al. | May 2010 | A1 |
20100249976 | Aharoni et al. | Sep 2010 | A1 |
20100287473 | Recesso et al. | Nov 2010 | A1 |
20100293174 | Bennett et al. | Nov 2010 | A1 |
20110225136 | Tu et al. | Sep 2011 | A1 |
20110282879 | Bargeron et al. | Nov 2011 | A1 |
Number | Date | Country |
---|---|---|
1452098 | Oct 2003 | CN |
101128818 | Feb 2008 | CN |
101059806 | Jul 2009 | CN |
20030014804 | Feb 2003 | KR |
Entry |
---|
Milic-Frayling, “Text Processing and information retrieval,” Chapter 1, Text Mining and its Applications to Intelligence, CRM and Knowledge Management, Jan. 1, 2005, 44 pages, Microsoft Corporation. |
Natsev, et al., “Semantic Concept-Based Query Expansion and Re-ranking for Multimedia Retrieval,” Proceedings of the 15th International Conference on Multimedia, Jan. 1, 2007, 10 pages, New York, New York. |
Pooch, et al., “A survey of Indexing Techniques for Sparse Matrices,” Comparing Surveys, Jun. 1973, 26 pages, vol. 5, No. 2, New York, NY. |
European Search Report of 09723190.6 mailed on Aug. 23, 2012, 6 pages. |
Crystal Fibre AJS, “Nonlinear Photonic Crystal Fibers Selected Datasheets 800 NM Fibers NL-800 List”, [online], [search on Sep. 26, 2007], internet (URL:http://www.crystal-fibre.com/products/nonlinear.shtm). |
Knight, J. C. et al., “Anomalous Dispersion in Photonic Crystal Fiber”, IEEE Photonics Technology Letters, vol. 12, No. 7, Jul. 2000, pp. 807-809. |
Ranka, J. K. et al., “Optical properties of high-delta air-silca microstructure optical fibers”, Optics Letters. vol. 25, No. 11, Jun. 1, 2000, pp. 796-798. |
U.S. Appl. No. 13/048,268 Final Office Action mailed Sep. 13, 2013, 23 pages. |
Extracting and Preparing Metadata to Make Video Files Searchable, Telestream, Retrieved on Mar. 15, 2012, from http://www.telestream.net/pdfs/whitepapers/wp-preparing-video-metadata.pdf, 6 pages. |
Natasa Milic-Frayling: “Text processing and information retrieval” In: “Text Mining and its Applications to Intelligence, CRM and Knowledge Management”, Jan. 1, 2005, 44 pgs. (Microsoft Corporation, XP55035433). |
U.S. Appl. No. 13/048,268 First Action Interview Office Action mailed Apr. 17, 2013, 15 pages. |
U.S Appl. No. 11/903,279, filed Sep. 21, 2007, mailed on Jun. 6, 2014, 26 pages. |
Number | Date | Country | |
---|---|---|---|
20130013305 A1 | Jan 2013 | US |
Number | Date | Country | |
---|---|---|---|
60846484 | Sep 2006 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12077591 | Mar 2008 | US |
Child | 13048268 | US | |
Parent | 13245850 | Sep 2011 | US |
Child | 13312373 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12077590 | Mar 2008 | US |
Child | 13524178 | US | |
Parent | 11903279 | Sep 2007 | US |
Child | 12077590 | US | |
Parent | 13524178 | US | |
Child | 12077590 | US | |
Parent | 13048268 | Mar 2011 | US |
Child | 13524178 | US | |
Parent | 11903279 | Sep 2007 | US |
Child | 12077591 | US | |
Parent | 13524178 | US | |
Child | 12077591 | US | |
Parent | 11903279 | Sep 2007 | US |
Child | 13524178 | US | |
Parent | 13524178 | US | |
Child | 13524178 | US | |
Parent | 13245840 | Sep 2011 | US |
Child | 13524178 | US | |
Parent | 11903279 | Sep 2007 | US |
Child | 13245840 | US | |
Parent | 13524178 | US | |
Child | 13245840 | US | |
Parent | 13245843 | Sep 2011 | US |
Child | 13524178 | US | |
Parent | 13245855 | Sep 2011 | US |
Child | 13245843 | US | |
Parent | 13312373 | Dec 2011 | US |
Child | 13245855 | US |