The present specification relates to a method for generating content using user sentences using artificial intelligence.
There may be several approaches that utilize artificial intelligence in producing experiential content. For example, there may be an interactive storytelling technology in which users become main characters of a specific story, a method of developing a virtual character that interacts with the user in an environment such as virtual reality or augmented reality, or the like.
A combination of various technologies and tools is required to produce completed experiential content in which users directly participate using artificial intelligence. For example, natural language processing technologies and machine learning algorithms may be used to construct systems that understand user input and generate content accordingly.
To this end, it is necessary to further develop artificial intelligence models to meet complex interactions and various user needs, and to construct systems that process users' actions and responses in real time and provide results accordingly.
In addition, development is required in a direction of providing a personalized experience by collecting user preferences or feedback in a process of producing content according to user needs.
An object of the present specification is to generate information necessary for an asset search capable of generating content through artificial intelligence using sentences written by a user.
In addition, an object of the present specification is to generate content using sentences written by a user through artificial intelligence.
Objects of the present specification are not limited to the above-mentioned objects. That is, other objects that are not mentioned may be obviously understood by those skilled in the art to which the present specification pertains from the following detailed description.
According to an embodiment of the present invention, a method of generating, by a recommendation apparatus, content to a user and recommending an asset to be displayed in the content, the method comprising: receiving a sentence for generating the content from a user; transforming the sentence into a story type text through a language model; transforming the story type text into a storyline including 1) a background, 2) a main character, and 3) a main element through the language model; transforming the storyline into sentence data for generating the content or recommending the asset through the language model; and generating the content or recommending the asset based on the sentence data.
The method may further include: transforming the sentence into English; and transforming the storyline into Hangul.
The sentence data may include a source word, an English word corresponding to the source word, parts of speech, and importance.
The story type text may be transformed through the language model based on a command for writing a synopsis.
The method may further include: extracting a keyword by analyzing an element file including element information; transforming the keyword into the story type text through the language model; and updating meta information of the asset based on the sentence data.
The method may include the meta information of the asset is used to search for the asset.
According to another embodiment of the present invention, a recommendation apparatus for recommending an asset and content to a user include a communication unit, a memory including a language model, and a processor configured to functionally control the communication unit and the memory, in which the processor receives a sentence for generating the content from a user, transforms the sentence into a story type text through a language model, transforms the story type text into a storyline including 1) a background, 2) a main character, and 3) a main element through the language model, transforms the storyline into sentence data for generating the content or recommending the asset through the language model, and generates the content or recommending the asset based on the sentence data.
According to an embodiment of the present specification, it is possible to generate information necessary for an asset search capable of generating content through artificial intelligence using sentences written by a user.
In addition, according to an embodiment of the present specification, it is possible to generate content using sentences written by a user through artificial intelligence.
Effects which can be achieved by the present specification are not limited to the above-mentioned effects. That is, other objects that are not mentioned may be obviously understood by those skilled in the art to which the present specification pertains from the following description.
The accompanying drawings, which are included as part of the detailed description to assist understanding of the present disclosure, illustrate embodiments of the present disclosure and explain the technical features of the present disclosure together with the detailed description.
Hereinafter, embodiments disclosed in this specification are described in detail with reference to the accompanying drawings. The same or similar element is assigned the same reference numeral regardless of its reference numeral, and a redundant description thereof is omitted. It is to be noted that the suffixes of elements used in the following description, such as a “module” and a “unit”, are assigned or interchangeable with each other by taking into consideration only the ease of writing this specification, but in themselves are not particularly given distinct meanings and roles. Furthermore, in describing an embodiment disclosed in this specification, when it is determined that a detailed description of a related known technology may obscure the subject matter of an embodiment disclosed in this specification, the detailed description will be omitted. Furthermore, it is to be understood that the accompanying drawings are merely intended to make easily understood the embodiments disclosed in this specification, and the technical spirit disclosed in this specification is not restricted by the accompanying drawings and includes all changes, equivalents, and substitutions which fall within the spirit and technical scope of this specification.
Terms including ordinal numbers, such as a “first” and a “second”, may be used to describe various components, but the components are not restricted by the terms. The terms are used to only distinguish one element from the other elements.
When it is described that one element is “connected” or “coupled” to another element, it should be understood that one element may be directly connected or coupled to the other element, but a third element may exist between the two elements. In contrast, when it is described that one element is “directly connected to” or “directly coupled to” the other element, it should be understood that a third element does not exist between the two elements.
An expression of the singular number includes an expression of the plural number unless clearly defined otherwise in the context.
In this specification, it is to be understood that a term, such as “include” or “have”, is intended to designate that a characteristic, a number, a step, an operation, an element, a part or a combination of them described in the specification is present, and does not exclude the presence or addition possibility of one or more other characteristics, numbers, steps, operations, elements, parts, or combinations of them in advance.
An electronic device 100 may include a wireless communication unit 110, an input unit 120, a sensing unit 140, an output unit 150, an interface unit 160, memory 170, a controller 180, a power supply unit 190, etc. The components illustrated in
More specifically, among the components, the wireless communication unit 110 may include one or more modules that enable wireless communication between the electronic device 100 and a wireless communication system, between the electronic device 100 and another electronic device 100, or between the electronic device 100 and an external server. Furthermore, the wireless communication unit 110 may include one or more modules that connect the electronic device 100 to one or more networks.
The wireless communication unit 110 may include at least one of a broadcasting reception module 111, a mobile communication module 112, a wireless Internet module 113, a short-distance communication module 114, and a position information module 115.
The input unit 120 may include a camera 121 or an image input unit for receiving an image signal, a microphone 122 or an audio input unit for receiving an audio signal, and a user input unit 123 (e.g., a touch key or a mechanical key) for receiving information from a user. Voice data or image data that is collected by the input unit 120 may be analyzed and processed as a control command of the user.
The sensing unit 140 may include one or more sensors for sensing at least one of information within the electronic device, surrounding environment information around the electronic device, and user information. For example, the sensing unit 140 may include at least one of a proximity sensor 141, an illumination sensor 142, a touch sensor, an acceleration sensor, a magnetic sensor, a G-sensor, a gyroscope sensor, a motion sensor, an RGB sensor, an infrared (IR) sensor, a finger scan sensor, an ultrasonic sensor, an optical sensor, for example, a camera (refer to 121), a microphone (refer to 122), a battery gauge, an environment sensor (e.g., a barometer, a hygrometer, a thermometer, a radioactivity sensor, a thermal sensor, or a gas sensor), a chemical sensor (e.g., an electronic nose, a healthcare sensor, a bio recognition sensor). The electronic device disclosed in this specification may combine and use pieces of information that are sensed by at least two of these sensors.
The output unit 150 is for generating an output related to a visual, auditory, or tactile sense, and may include at least one of a display unit 151, an acoustic output unit 152, a haptic module 153, and an optical output unit 154. The display unit 151 may implement a touch screen by forming a mutual layer structure along with a touch sensor or being integrally formed along with a touch sensor. The touch screen may function as the user input unit 123 that provides an input interface between the electronic device 100 and a user, and may also provide an output interface between the electronic device 100 and a user.
The interface unit 160 may serve as a passage with various types of external devices that are connected to the electronic device 100. The interface unit 160 may include at least one of a wired/wireless headset port, an external charger port, a wired/wireless data port, a memory card port, a port that connects a device equipped with an identification module, an audio input/output (I/O) port, a video I/O port, and an earphone port. The electronic device 100 may perform proper control related to an external device connected thereto, in response to the connection of the external device with the interface unit 160.
Furthermore, the memory 170 may store data that supports various functions of the electronic device 100. The memory 170 may store multiple application programs or applications that are driven in the electronic device 100, data for an operation of the electronic device 100, and instructions. At least some of the application programs may be downloaded from an external server through wireless communication. Furthermore, at least some of the application programs may be present on the electronic device 100 from the time of release for basic functions (e.g., call incoming and outgoing functions and message reception and sending functions) of the electronic device 100. The application program may be stored in the memory 170, may be installed in the electronic device 100, and may be driven to perform an operation (or function) of the electronic device by the controller 180.
The controller 180 commonly controls an overall operation of the electronic device 100 in addition to an operation related to the application program. The controller 180 may provide information or a function suitable for a user or process the information or function by processing a signal, data, or information that is input or output through the aforementioned components or driving an application program stored in the memory 170.
Furthermore, the controller 180 may control at least some of the components that have been described with reference to
The power supply unit 190 may be supplied with external power and internal power under the control of the controller 180, and may supply power to each of the components included in the electronic device 100. The power supply unit 190 includes a battery. The battery may be an embedded type battery or a replaceable type battery.
At least some of the components may cooperatively operate in order to implement an operation, control, or a control method of the electronic device according to various embodiments described hereinafter. Furthermore, the operation, control, or control method of the electronic device may be implemented on the electronic device by the driving of at least one application program stored in the memory 170.
In this specification, the electronic device 100 may include a recommendation apparatus, a terminal, a visual coding apparatus, and a convergence type production apparatus.
Referring to
The recommendation apparatus includes an AI device 20 to be described later. The recommendation apparatus is connected to a user terminal and a convergence type production apparatus to analyze a user's propensity using meta information generated from the terminal, the platform, and the convergence type production apparatus, and recommend content and an asset suitable for the user using the analysis result.
For example, the recommendation apparatus may generate JSON sentence data using a large-scale language model framework based on sentences written in an input window by a user through a terminal. In this way, the recommendation apparatus recommends the content generation and asset. By analyzing the uploaded file, JSON sentence data may be generated through a language model framework and stored in asset meta information.
The convergence type production apparatus receives an instruction to produce content from the terminal through the Web (S2010). For example, the content may include a 2-D object and/or a 3-D object.
The convergence type production apparatus generates a page for the production of the content (S2020). For example, the page includes a page which may be represented in a 2-D or 3-D form. The convergence type production apparatus may construct a screen by adding and disposing a predefined asset or template to and in each page. More specifically, an event may be registered with the added asset, so that an interaction with a content user may be added to the added asset. Accordingly, the user may produce immersive and creative content.
Referring to
Referring back to
Referring to
Referring back to
Referring to
For example, a page may include (1) the attribute of the page, (2) an event list, and (3) a resource list. More specifically, the event list includes information of events assigned to an asset. The resource list includes information of assets added to the page.
Referring to
Referring back to
For example, when the user selects an asset disposed on the page display screen 300, the convergence type production apparatus displays the attribute window 410 corresponding to the selected asset. The user may modify an attribute value of the asset by using a mouse, a touch, or the input of numerical values through the registered controller 400 based on a page.
The convergence type production apparatus registers an event corresponding to the asset (S2060). For example, the event may include a set of “actions” and “results”. More specifically, the action may define a condition in which the event occurs. For example, the “action” may include various forms of events or invoking, such as a keyboard event, a mouse/touch event, a gesture event, an area event, a value event, and an invoking event, as a condition in which an “action” function is performed.
Furthermore, the results may include a “function” and a “target”.
More specifically, the “function” may define an attribute change and specific behaviors to be performed on a “target”, that is, the purpose of the function when an event is activated, and may include a control function for the default attributes of an asset, such as a location, a size, rotation, and transparency, and media assets, such as view, hiding, playback, a stop, and a pause. Furthermore, the “function” may include a camera of a user terminal, a GPS, or an accelerometer for using information of an external environment.
The convergence type production apparatus registers an action, function and/or target corresponding to the event (S2070). For example, the convergence type production apparatus may register the action, function and/or target based on the attribute value of the asset.
The convergence type production apparatus executes the results based on the registered action when an event for the asset is detected (S2080).
Referring to
Referring to
Referring to
Referring to
Referring to
Referring to
For example, if an uploaded asset file is a video file, the convergence type production apparatus may generate HTMLVideoElement for the video file, may add the HTMLVideoElement to a screen, may register an object of the HTMLVideoElement as a “use asset”, and may control the object.
One piece of content may include several pages. Each of the pages may include a resource manager “resourceManager” that manages assets and an event manager “eventManager” that manages an operation event.
Assets disposed in a page have separate depths. Accordingly, the convergence type production apparatus may adjust the order of assets that are exposed to a screen by adjusting the depths of the assets depending on a user's need.
For example, when an asset is disposed in a page, the convergence type production apparatus registers the asset with the resource manager of the page, and manages the asset through a change, or the registration or deletion of an attribute, a state, etc. of the asset. The resource manager may manage the disposed asset, and may invoke a connected function according to a change in the resource based on a resource list of the page by generating an event for a change.
When an asset is added, the convergence type production apparatus may add the asset to a management list in order to control the asset through a controller registered with a page in which the asset is disposed so that the asset can be recognized and managed as an asset which may be controlled by the controller. An attribute of the asset registered with the controller may be modified or controlled through a mouse, a gesture, a touch, an external controller, etc. of a user.
Furthermore, a Redo/Undo function may be required in a process of a user controlling an asset. To this end, the convergence type production apparatus may construct a history manager of a page in which the asset is disposed so that the history manager may record a change in the state of the asset by storing a change in the controller, may invoke the attribute of the asset from a list including the change, which is kept in the history manager in response to a request, and may restore or update again the state of the asset by updating a current attribute of the asset. The history manager may have a memory capacity problem because a change is stored for each asset. Accordingly, in order to prevent the memory capacity problem, the convergence type production apparatus may restrict the number of lists of changes that are stored depending on circumstances.
Furthermore, the convergence type production apparatus may designate a user with which content will be shared when the content is produced, and may perform the task of producing the content simultaneously with the user. For example, when content is initially produced, the convergence type production apparatus may generate a unique channel corresponding to the content. The convergence type production apparatus may designate a sharing user who may use the unique channel, and may add the sharing user to the same channel when the user accesses the content.
The user in the same channel may synchronize authoring data in real time while exchanging a change in the authoring data in real time. Communication for the synchronization may be performed in real time through websocket or webrtc.
However, when a new user accesses content that has not been stored, a difference between a version of the content that has not been stored and a previous version of the content that has been stored may occur because the user is synchronized with a state of the content having the previous version. In order to prevent such a problem, when a new user is added to a sharing channel and first accesses content, initial synchronization may be performed by performing an overall update on a task change through a specific user among users who are performing the existing task. After the initial synchronization is performed as described above, the convergence type production apparatus may solve a synchronization issue for content that has not been stored through a method of sharing data having a change in real time.
The AI device 20 may include an electronic device including an AI module capable of performing AI processing, a server including the AI module, or the like. In addition, the AI device 20 may be included in at least a part of the electronic device 100 illustrated in
The AI device 20 may include an AI processor 21, a memory 25, and/or a communication unit 27.
The AI device 20 is a computing device capable of learning neural networks, and may be implemented as various electronic devices such as a server, a desktop PC, a notebook PC, and a tablet PC.
The AI processor 21 may train a neural network using a program stored in the memory 25. For example, the AI processor 21 may generate a Tensorflow-utilized recurrent neural networks model on the memory 25, and use data that may be collected from the user terminal, the convergence production apparatus, and the like to train these AI models.
For example, the trained artificial intelligence model may include a language model, and generate sub-models having the following tasks:
Referring to
For example, the platform may be provided to a terminal through the WEB, and may be connected to the convergence type production device to provide a content production environment to a user. In addition, content produced by other users may be displayed, and a search function for the content may be provided. In addition, the platform may provide a user with a template that may be used in a content production environment, and may also provide a search function for the template.
The trained artificial intelligence model may generate 1) a user characteristic analysis model, 2) a content recommendation model, and 3) an element recommendation model. The user characteristics output by the user characteristic analysis model may be used for training of the artificial intelligence model 1300.
In addition, the content recommendation model may receive sentences from the user, transform the sentences into JSON type sentence data, and provide automatically generated content to a user through the convergence type production apparatus.
In addition, the element recommendation model may recommend an asset to a user through the convergence type production device using the JSON type sentence data.
Referring to
Furthermore, the terminal may be connected to the visual coding apparatus through the Web even without a separate application. A user may produce content through visual coding by using the terminal.
The terminal generates a page for visual coding, and disposes an asset in the page (S1410). For example, the terminal may provide the user with a list of assets and/or a template including assets. The user may select, from the template, an asset to be disposed in the page.
The terminal sets a target asset, that is, the target of visual coding, based on the disposed asset (S1420). For example, a user may select the target asset, among assets disposed in the page, by clicking on the target asset.
The terminal sets a user behavior related to an interaction with the user (S1430). For example, the user behavior may be a condition in which an event of the target asset occurs.
The terminal sets results related to the target asset, based on the user behavior (S1440). For example, the results may mean an operation which is performed in relation to the target asset, when the terminal receives, from the user, an input corresponding to the user behavior. The results may include a “function” for controlling the size, location, and state value of the target asset, an “operation” for operating variables related to the target asset, and a “function page” indicative of a movement of a page.
The terminal displays the results of the target asset based on the user behavior being input (S1450). For example, when one click on the target asset is received from a user, the terminal may move the target asset for 1 second or display a screen that has been moved to another page, as the results.
Referring to
The recommendation apparatus receives sentences from the user through the convergence type production apparatus and/or platform (S1510). For example, a user may input a sentence about a specific topic or story to the convergence type production apparatus to produce content. In more detail, the user can input the sentence “blue sky and grass with a smiley sun” into the convergence type production apparatus. In this case, the input language may be mainly Korean, and the recommendation apparatus may store the input sentence for the follow-up process.
Referring to
Referring back to
For example, the sentence “Blue sky and grass with a smiley sun” may be converted into an English sentence “Blue sky and grass with a smile.”
The recommendation apparatus transforms a sentence into a story type text through a language model (S1530). In more detail, the template of the language model may be configured to issue the command Write a synopsis about the “blue sky and grass with a smile” for the sentence “Blue sky and grass with a smile”.
For example, “Blue sky and grass with a smile” may be transformed into the story type text as follows.
In the enchanting world of “Blue Sky and Grass with a Smile,” we are transported to a serene and picturesque landscape where the beauty of the blue sky and lush green grass becomes a catalyst for transformation. The story revolves around Emily, a young girl grappling with the hardships of life and burdened by sorrow. However, one fateful day, she stumbles upon a hidden meadow, bathed in the golden sunlight, where the sky is a vivid shade of blue and the grass dances with an infectious joy.
As Emily immerses herself in this magical haven, she starts to notice how the vibrant hues of the sky and grass affect her mood. The brilliant blue sky acts as a gentle reminder to look beyond her troubles and find hope, while the cheerful green grass inspires her to embrace the present moment and seek happiness. Slowly but surely, a smile starts to grace Emily's face as she discovers the power of perspective and the healing balm of nature.
With each visit to the meadow, Emily's connection to the blue sky and grass deepens. She begins to cultivate a profound sense of gratitude for the beauty surrounding her, finding solace and strength in the simplicity of life's wonders. The radiant smile that forms on her lips becomes a beacon of optimism, influencing those around her and spreading contagious joy.
Through this poignant journey, “Blue Sky and Grass with a Smile” highlights the transformative power of nature and the profound impact it can have on our emotional well-being. It reminds us to find solace in the smallest of moments, to appreciate the beauty that surrounds us, and to nurture a smile that can illuminate even the darkest of days.
This is to compensate for the fact that the sentence input by a first user has insufficient information for content generation and/or asset recommendation.
The recommendation apparatus transforms the story type text into a storyline including a background, a main character, and main elements through a language model (S1540). The recommendation apparatus analyzes the story type text by utilizing the language model again, and may be summarized as a storyline by condensing the story type text into one scene including the background, main characters, and main elements. This is to extract key elements of content, and the recommendation apparatus may generate a storyline by extracting the key elements such as the background, main characters, and main elements among many data included in the story type text.
To this end, the template of the language model may be set to Summarize the storyline in three lines, including the background, main characters, and main elements.
For example, the story type text exemplified in S1530 described above may be transformed into a storyline as follows.
In “Blue Sky and Grass with a Smile,” Emily, a young girl burdened by sorrow, discovers a hidden meadow where the vibrant blue sky and dancing green grass become catalysts for transformation. Immersed in this magical haven, she learns to find hope, embrace the present, and cultivate gratitude. Through the healing power of nature, Emily's radiant smile becomes a symbol of optimism, inspiring those around her and reminding us to appreciate life's simple wonders.
The recommendation apparatus transforms the storyline into the sentence data for the content generation and/or asset recommendation (S1550). If necessary, the recommendation apparatus may translate the storyline back into Korean. For example, the storyline exemplified in S1540 described above may be translated as follows.
“In “With the Smile of the Sky and Grass,” Emily, a grieving girl, discovers a hidden meadow where vibrant blue sky and dancing green grass are a starting point for change. As she immerses herself in this magical place, she learns to find her hope, accept the present, and cultivate gratitude. Through her natural healing powers, Emily's radiant smile has become a symbol of optimism, inspires those around her and reminds us to realize the simple beauty of life.” In addition, the recommendation apparatus may transform the storyline into JSON type sentence data based on source words, English words, classification, zero classification, importance, and parts of speech using the language model.
Table 1 below illustrates sentence data.
Referring to Table 1, the recommendation apparatus may convert the storyline type text for the first input sentence into the JSON type data. The JSON type data may be used as data for automatic content generation and/or asset recommendation. The JSON type data may include sentence data information such as background, main characters, and main elements.
The recommendation apparatus recommends content generation and/or assets based on sentence data (S1560).
Referring to
For example, content 1710 in the form of a smiley sun included in the blue sky and grass laid at the bottom may be generated for the sentence “Blue sky and grass with a smiley sun.”
In addition, the recommendation apparatus may add an animation 1720 to the smile element included in the generated content 1710 based on sentence data.
Referring to
A user may select an asset from the recommended asset list 1810 and utilize the selected asset as an element displayed in the content 1820.
Referring to
The recommendation apparatus analyzes an element file and extracts a keyword (S1910). For example, a file uploaded by a user is analyzed using an artificial intelligence model, and important keywords related to content and assets may be extracted through a language model.
The recommendation apparatus transforms the extracted keywords into English (S1920). When the extracted keyword is languages other than English, the extracted Keyword may be transformed into English for accuracy of the language model included in the recommendation apparatus.
The recommendation apparatus transforms the keyword into the story type text through the language model (S1930). For example, the keyword transformed into English may be further decorated with a story type text using a language model template. The language model may serve to generate and structure sentences based on previously trained text data.
The recommendation apparatus transforms the story type text into the storyline that includes the background, main characters, and main elements (S1940).
For example, the recommendation apparatus may analyze the returned story by utilizing the language model again, and write the analyzed story as the storyline including the background, main characters, and main elements among those.
The recommendation apparatus transforms the storyline into the sentence data for the content generation and/or asset recommendation (S1950). For example, the recommendation apparatus may transform information such as the background, main characters, and main elements included in the story into the JSON type sentence data.
The recommendation apparatus updates meta information of an asset based on sentence data (S1960). For example, the recommendation apparatus may include the asset meta information. Through the meta information update, asset attributes, features, related information, etc., may be enriched, and used for search and classification of assets in the future.
For example, the platform may utilize the added meta information to increase the search range of assets. In more detail, users may search for and select a more accurate and desired asset by utilizing this meta information. Through this, users may quickly find necessary assets and improve efficiency in the content production process.
The aforementioned present disclosure may be implemented in a medium on which a program has been recorded as a computer-readable code. The computer-readable medium includes all types of recording media in which data readable by a computer system is stored. Examples of the computer-readable medium include a hard disk drive (HDD), a solid state disk (SDD), a silicon disk drive (SDD), ROM, RAM, CD-ROM, a magnetic tape, a floppy disk, an optical data storage device, and also include an implementation having the form of carrier waves (e.g., transmission through the Internet). Accordingly, the detailed description should not be construed as being limitative, but should be considered to be illustrative from all aspects. The scope of the present disclosure should be determined by reasonable analysis of the attached claims, and all changes within the equivalent scope of the present disclosure are included in the scope of the present disclosure.
Furthermore, although the services and embodiments have been chiefly described, they are only illustrative and are not intended to limit the present disclosure. A person having ordinary knowledge in the art to which the present disclosure pertains may understand that various modifications and applications not illustrated above are possible without departing from the essential characteristics of the present services and embodiments. For example, each of the components described in the embodiments may be modified and implemented. Furthermore, differences related to such modifications and applications should be construed as belonging to the scope of the present disclosure defined in the appended claims.
| Number | Date | Country | Kind |
|---|---|---|---|
| 10-2023-0080604 | Jun 2023 | KR | national |