There is a need to speak thoughts, but no current technological method for recording, saving, multipoint sharing, and searching audio and speech-to-text on a single platform, or analyzing or monetizing content on a single platform.
Disclosed is a system, method, and platform that enables users to record audio, transcribe speech to text, save and view audio and speech-to-text on a timeline, identify audio and speech-to-text content using a set of parameters, share audio and speech-to-text with multiple users, view analytics of content across a user's timeline, search audio by keyword, and monetize content.
The platform features an audio recordation component, audio transcription component, analytics identification component, a sharing component, timeline component, an analytics aggregate component, a feed component, channel subscription component, search component, and an export component. The components interact with and communicate data to one another through a system logic, which is embedded in software and/or a server. The platform may additionally comprise a user interface enabling a user to interact with the platform components. The platform may comprise a set of computing devices interacting over a network, exchanging data and data requests. Ideally, the platform is designed to operate on a mobile device, but can be extended to other devices, including computers or dedicated device/hardware systems.
The audio recordation component may feature a microphone built into a mobile device or computer, an audio-to-digital converter, and one or more databases in which an audio signal captured by the microphone and converted by the audio-to-digital converter may be saved. Information relevant to the audio recording, such as the chronological entry number, time, date, and user information, may be automatically created by other components of the platform, associated with the audio recording, and saved with it in a database. The audio recordation component may be coupled to the user interface, enabling the user to communicate to the platform that the user wishes to record audio by selecting a button. The audio transcription component analyzes the audio waves and converts speech to text. This text may then be associated and stored with the audio recording.
The analytics identification component assigns parameter tags to content. These parameter identifiers may include mood, type, characters, and location. The mood parameter may describe an emotion or attitude related to the content, i.e., positive, neutral, or negative, or more specifically, i.e., happy, sad, frightened, or bored. The type parameter may describe the nature of the content, particularly its frequency and/or conscious state. For example, content created by the user may be ordinary, refer to a recurrent thought, or it may be lucid—having occurred during lucid dreaming. The character parameter may identify individuals, groups, or entities, real or fictional, that are subjects of or otherwise relating to the content. One or more individuals, groups, or entities may be inputted and included as character parameters. The location parameter may describe a location relevant to the content or the place in which the content was created or experienced. The analytics identification component may also be coupled to the user interface, allowing the user to select or input parameters. The user may select the parameters by touching or clicking one or more buttons disposed on the user interface or input text, speech, and/or speech-to-text into a field.
The sharing component identifies whether the audio and speech-to-text post can be accessed by just the user, individuals selected by the user, or all users who subscribe to the user. Selecting the friends or public button shares the audio and speech-to-text post with the feed component.
The timeline component organizes posts by chronological entry number. In a preferential version, this entry number is created automatically by the platform and applied consecutively to content posts created by the user. Audio recordings are displayed as posts on a user interface, with interactive buttons and disposed adjacent to text transcriptions of the audio recordings, entry number, and timestamp. Users may select a play button to hear the original audio recording, press again to pause or stop the recording, a comment button to view previous comments and/or leave a comment, and a clap button to like the audio and speech-to-text post (and/or to increase the number of users who have liked the post). Selecting the play button sends a request to the platform to access a database and retrieve the audio clip information, which is then processed into sound by the requesting the user's processor and speakers. Similarly, selecting the comment button may cause the platform to access the database and retrieve and display comments associated with the audio and speech-to-text post. A comment field may be displayed, permitting the user to comment on the audio and speech-to-text post via text, speech, and/or speech-to-text. This comment may then be saved by the platform to the database and associated with the post.
The analytics aggregate component shows parameters including the mood, type, characters, and/or location by percentage for a user's content posts. It executes calculations using information stored in the database relating to the data on mood, type, characters, and location. These calculations may include identifying averages for any given instantiation of mood among all instantiations of mood. These calculations may be similarly executed for type, characters, and location. Calculations may also include identifying the instantiations as percentages of the total instantiations for the parameter class. When the analytics aggregate component is coupled to the user interface, these averages and/or percentages may be displayed. For example, an analytics screen may display the percent of a user's content that are identified as positive, neutral, and negative.
The feed component may feature audio and speech-to-text information of content created by users who have accepted friend requests or sent accepted friend requests, or channel users they have subscribed to. When coupled to the user interface, content may be displayed in the order in which the posts were created, with more recent posts appearing more prominently and/or above less recent posts. Similarly to the way in which interaction options are displayed and accessed on the timeline, a user may press to play the audio recording, press again to stop the audio recording, view and leave comments via displays and input fields, like and increase user engagement, and view text transcriptions of the audio recordings, the chronological entry number for the posting user, and timestamp. These posts may also identify the user adjacent to the text transcriptions, along with the location identified by the posting user or recorded by the platform.
The search component features a search engine and input field of text, speech, and/or speech-to-text where a user can input one or more keyword(s). The search engine then searches all posts that are saved publicly on the platform or displayed in the feed and timeline, for speech-to-text transcriptions comprising the keyword(s). The platform may then display the results in a similar format to posts on the feed screen, where a user may press to play the audio recording, press again to stop the audio recording, view and leave comments via displays and input fields, like and increase user engagement, and view text transcriptions of the audio recordings, the chronological entry number for the posting user, and timestamp.
The channel subscription component enables users to publicize their content, thereby giving subscribers access to view their content that are saved as public, in the feed screen. A user may search for channel users. The platform may receive a subscription request from a user for the account of a channel user. The channel subscription component may be coupled to a payment component. The platform may require receipt of user payment or an agreement to pay before enabling a subscription event. The payment component may be coupled to a user interface in which the user may enter payment details, such as credit card information. Alternatively, the payment component may be connected to an SDK of a payment processor. Payment received from the user may be transmitted to a bank account managed by the platform. Payment may then be divided and transmitted monthly or regularly to the email of the user being subscribed to via a payment processor such as PayPal, for example, based on the monthly difference in the amount of users who have subscribed to a channel user.
The export component enables users to transfer audio and speech-to-text posts to an inputted email for other use independent from the platform. The export component may be coupled to the user interface that displays an input field where a user can input an email address. The user interface may have an anonymous button that can be selected to replace the user's identification information, such as name, in the export. Once the request is received by the platform, audio and speech-to-text data are sent in the body of an email, in which case audio information may be embedded therein as a link to the audio file stored in the platform's database that can be downloaded and/or an audio player. Alternately, the posts are copied as several distinct files, including separate audio and text files. These files are then send via the export component, which may have access to an email server or engage with an API from an email platform, to the address identified.
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
As shown in
Number | Date | Country | |
---|---|---|---|
62456604 | Feb 2017 | US | |
62517879 | Jun 2017 | US | |
62534421 | Jul 2017 | US |