WO2023220068A1 - Systèmes et procédés de création de contenu sur la base d'une préférence de public et de facteurs contextuels - Google Patents

Systèmes et procédés de création de contenu sur la base d'une préférence de public et de facteurs contextuels Download PDF

Info

Publication number
WO2023220068A1
WO2023220068A1 PCT/US2023/021559 US2023021559W WO2023220068A1 WO 2023220068 A1 WO2023220068 A1 WO 2023220068A1 US 2023021559 W US2023021559 W US 2023021559W WO 2023220068 A1 WO2023220068 A1 WO 2023220068A1
Authority
WO
WIPO (PCT)
Prior art keywords
content
audience
user
preferences
parameter
Prior art date
Application number
PCT/US2023/021559
Other languages
English (en)
Inventor
Amod Jain
Rajat Mishra
Shreos Roychowdhury
Original Assignee
Prezent Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US18/156,649 external-priority patent/US11775748B1/en
Application filed by Prezent Inc. filed Critical Prezent Inc.
Publication of WO2023220068A1 publication Critical patent/WO2023220068A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/166Editing, e.g. inserting or deleting
    • G06F40/186Templates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/103Formatting, i.e. changing of presentation of documents
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management

Definitions

  • This specification relates to content creation and, in particular, to content creation based on audience preferences and contextual factors.
  • At least one aspect of the present disclosure is directed to a method for creating a template for content.
  • the method includes receiving at least one audience parameter representing at least one preference of a target audience when receiving the content, receiving at least one context parameter representing a context of the content; determining a content storyline based on the at least one audience parameter and the at least one context parameter, and generating a content template based on the content storyline.
  • the target audience is a single individual. In some embodiments, the target audience is two or more individuals.
  • the at least one audience parameter represents one or more of image preferences, color preferences, illustration preferences, presentation preferences for logical arguments, presentation preferences for data, and preferences for visualizing data of the target audience.
  • the method includes presenting at least one preference inquiry to the target audience, deriving, based on a response to the at least one preference injury, the at least one preference of the target audience when receiving the content, and generating at least one preference fingerprint for the target audience based on the at least one preference of the target audience.
  • the at least one preference fingerprint includes the at least one audience parameter.
  • the at least one context parameter corresponds to one or more of a type, purpose, and objective of the content.
  • the content storyline corresponds to one or more of a sequence, order, and flow of the content.
  • the method includes selecting a style profile based on the at least one audience parameter and the at least one context parameter.
  • generating the content template includes generating the content template based on the content storyline and the style profile.
  • the style profile corresponds to one or more of a content theme, a font theme, a color theme, and an organizational theme of the content.
  • the content is a presentation and the content template corresponds to a slide deck for the presentation.
  • the method includes exporting the content template to at least one content editing application.
  • the system includes at least one memory for storing computer-executable instructions and at least one processor for executing the instructions stored on the memory. Execution of the instructions programs the at least one processor to perform operations that include receiving at least one audience parameter representing at least one preference of a target audience when receiving the content, receiving at least one context parameter representing a context of the content, determining a content storyline based on the at least one audience parameter and the at least one context parameter, and generating a content template based on the content storyline.
  • the target audience is a single individual. In some embodiments, the target audience is two or more individuals.
  • the at least one audience parameter represents one or more of image preferences, color preferences, illustration preferences, presentation preferences for logical arguments, presentation preferences for data, and preferences for visualizing data of the target audience.
  • execution of the instructions programs the at least one processor to perform operations that include presenting at least one preference inquiry to the target audience, deriving, based on a response to the at least one preference injury, the at least one preference of the target audience when receiving the content, and generating at least one preference fingerprint for the target audience based on the at least one preference of the target audience.
  • the at least one preference fingerprint includes the at least one audience parameter.
  • the at least one context parameter corresponds to one or more of a type, purpose, and objective of the content.
  • the content storyline corresponds to one or more of a sequence, order, and flow of the content.
  • execution of the instructions programs the at least one processor to perform operations that include selecting a style profile based on the at least one audience parameter and the at least one context parameter.
  • generating the content template includes generating the content template based on the content storyline and the style profile.
  • the style profile corresponds to one or more of a content theme, a font theme, a color theme, and an organizational theme of the content.
  • the content is a presentation and the content template corresponds to a slide deck for the presentation.
  • execution of the instructions programs the at least one processor to perform operations that include exporting the content template to at least one content editing application.
  • FIG. 1 illustrates a block diagram of a system for creating content templates in accordance with aspects described herein;
  • FIG. 2 illustrates a user interface of a client application in accordance with aspects described
  • FIGS. 3A-3F illustrate a user interface of a client application used to create a user fingerprint in accordance with aspects described
  • FIG. 4 illustrates a flow diagram of a method for creating a content template
  • FIG. 5 illustrates a user interface of a client application in accordance with aspects described
  • FIG. 6 illustrates a user interface of a client application in accordance with aspects described
  • FIG. 7 illustrates a user interface of a client application in accordance with aspects described
  • FIG. 8 illustrates a flow diagram of a method for transforming user-uploaded content
  • FIG. 9 illustrates a slide library in accordance with aspects described
  • FIG. 10 illustrates a template management interface in accordance with aspects described.
  • FIG. 11 illustrates an example computing device.
  • business professionals often create content to communicate information with a target audience. For example, such professionals may give a presentation to an audience to pitch an idea, report findings, or provide knowledge on particular topics.
  • these presentations take hours to construct and refine, and can be manually laborious to put together.
  • professionals and their teams find it difficult to standardize content, distribute presentations, and establish a common language.
  • presentations or content are often developed based on the presenter’s preferences, rather than the audience’s preferences.
  • a content storyline is generated based on parameters of the target audience and the context of the content.
  • the parameters of the target audience correspond to a preference fingerprint of the audience member(s).
  • a style profile is selected or created for the content template based on the parameters of the target audience, the context of the content, and the storyline.
  • FIG. 1 is a block diagram of a system 100 for creating content templates in accordance with aspects described herein.
  • the system 100 is implemented by an application server 102.
  • the application server 102 provides functionality for creating and providing one or more content templates.
  • the application server 102 comprises software components and databases that can be deployed at one or more data centers (not shown) in one or more geographic locations, for example.
  • the application server 102 software components may include a fingerprint engine 106, a context engine 107, an outline engine 108, a style engine 109, and a user interface (UI) engine 110.
  • the software components can comprise subcomponents that can execute on the same or on a different individual data processing apparatus.
  • the application server 102 databases may include an application database 112a and a user database 112b.
  • the databases can reside in one or more physical storage systems. Example features of the software components and data processing apparatus will be further described below.
  • the application server 102 is configured to send and receive data to and from users’ client devices (e.g., user device 116) through one or more data communication networks 104 such as the Internet, for example.
  • client devices e.g., user device 116
  • the user can access a user interface of a client application 118.
  • the client application 118 is configured to run in a web browser or a special -purpose software application executing on the user device 116.
  • the client application 118 can communicate with the application server 102 over the network(s) 104 using Hypertext Transfer Protocol (HTTP), another standard protocol, or a proprietary protocol, for example.
  • HTTP Hypertext Transfer Protocol
  • the user device 116 can be a mobile phone, a smart watch, a tablet computer, a personal computer, or a game console. Other types of user devices are possible.
  • the system 100 can enable the creation of content templates associated with different types of content.
  • the fingerprint engine 106 of the application server 102 is configured to generate a fingerprint for each user.
  • the fingerprint of each user represents the preferences of the user when receiving or viewing content.
  • the fingerprint represents the preferences of the user when the user is the target audience (or included in the target audience) for content.
  • the fingerprint engine 106 is configured to collect data from the user (e.g., via the client application 118) to generate the fingerprint.
  • the fingerprint engine 106 may store the fingerprint of each user in the user database 112b.
  • the context engine 107 is configured to manage the context of the content template and the corresponding content.
  • the context engine 107 collects information from the user(s) creating the content template (e g., via the client application 118) to determine the context of the content.
  • the context includes a type, purpose, and/or objective of the content. For example, if the content is a presentation, the context may represent the type or tone of the presentation.
  • the application database 112a includes a library of context information that is accessed by the context engine 107.
  • the context engine 107 may be configured to store the context information for the content template in the application database 112a.
  • the outline engine 108 is configured to generate an outline (or storyline) for the content template.
  • the outline may correspond to sequence, order, and/or flow of the content template. For example, if the content is a presentation, the outline may represent the type, number, and/or order of slides in a slideshow.
  • the application database 112a includes an outline library (e.g., a slide library) that is accessed by the outline engine 108.
  • the outline engine 108 may be configured to store the outline (or storyline) for the content template in the application database 112a.
  • the style engine 109 is configured to generate a style profile for the content template.
  • the style profile incudes one or more stylistic attributes of the content template and/or the content.
  • the style profile may include one or more of a template theme, a font theme, a color theme, and an organizational theme.
  • the template theme represents the stylistic preferences (e.g., graphics, logos, etc.) of a company or organization associated with the target audience of the content.
  • the font theme may represent one or more font attributes (e g., type, size, color, etc.) included in the content template and/or to be used for the generation of the content in a content editing tool.
  • the color theme may represent one or more colors (or color schemes) included in the content template and/or to be used for the generation of the content.
  • the organizational theme may represent one or more approaches or techniques for presenting information (e.g., data, ideas, concepts, plans, etc.) included in the content template and/or to be used for the generation of the content.
  • the style profde selected or created by the style engine 109 includes a work area for the content template.
  • the work area may correspond to a region or area of a presentation slide where the content is located.
  • the location and/or size of the work area may be determined by the style profde.
  • the style profde may also include a node count that is used to create various diagrams included in the content template (or the content). For example, the node count may the define the number of nodes (e.g., boxes) included in a flow chart or flow diagram.
  • the application database 112a includes a library of stylistic resources (e.g., fonts, colors, graphics, charts, tables, etc.) that is accessed by the style engine 109 to generate the style profde.
  • the style engine 109 may receive an uploaded presentation and extract various stylistic attributes. For example, a presentation associated with a particular company or organization may be provided to the style engine 109 to generate a corresponding style profde.
  • the style engine 109 may analyze the uploaded presentation and create (of find) a style profde that matches the stylistic attributes of the company or organization.
  • the style engine 109 may analyze XML data associated with the uploaded presentation to identify the corresponding stylistic attributes.
  • the style engine 109 may be configured to store each style profde in the application database 112a.
  • the UI engine 110 is configured to manage the UI of the client application 118.
  • the UI engine 110 may collect or gather information from users (e.g., content template creators) via the client application 118.
  • the UI engine 110 provides the collected information to the engines 106-109 and/or stores the collected information in the databases 112a, 112b.
  • the UI engine 110 may provide and present information from the engines 106-109 or the databases 112a, 112b to users via the client application 1 18.
  • the UT engine 1 10 provides an export service that allows the content template to be exported in various formats (e.g., PDF, Microsoft Word, Microsoft PowerPoint, etc.).
  • the UI engine 102 is configured to provide additional services such as user login, user account setup, user account management, etc.
  • FIG. 2 is an example view 200 of the client application 118 in accordance with aspects described herein.
  • the view 200 corresponds to a home screen of the client application 118.
  • FIG. 2 and other figures presenting user interfaces in this application include icons and labels and refer to various features displayed by the user interface (e.g., search, account, notifications, etc.). While such icons and labels will be used to reference and describe such features in this application, the features may be presented with different icons and labels as well.
  • the client application 118 includes a new presentation button 202 that enables the user (e.g., content template creator) to create a new presentation template.
  • the client application 118 includes a presentation button 204 that allows the user to view previously created templates.
  • the client application 118 may include a slides button 206 that allows the user to view, download, and/or a modify a plurality of presentation slides (e g., in a slide library of application database 112a).
  • the client application 118 includes a user button 208 that enables the user to manage a user account and/or various settings associated with the system 100.
  • the client application 118 includes a search bar 210 that provides a search function enabling the user to search for specific slides (e.g., from the slide library) and/or specific users (e.g., potential audience members).
  • the user may be prompted by the client application 118 to generate a fingerprint.
  • the user may initiate the fingerprint process via the user button 208 to create a new fingerprint or modify/update an existing fingerprint.
  • the user can generate a fingerprint for someone other than themselves (e.g., a member of the same company as the user, a friend of the user, etc.).
  • FIG. 3 A is an example view 300 of the client application 118 in accordance with aspects described herein.
  • the view 300 corresponds to the start of the fingerprint generation process.
  • the fingerprint generation process is configured to determine the visual preferences, the story choice (e g., storyline) preferences, and the data style preferences of the user when viewing or receiving content. The user may select the “Discover My Fingerprint” button 312 to start the fingerprint generation process.
  • the fingerprint generation process includes presenting a plurality of questions (or inquires) to the user.
  • FIG. 3B illustrates a view 320 of the client application 118 for determining the image preferences of the user.
  • the view 320 includes a first question 322 (e.g., “Which slide do you like more?”) that is presented to the user.
  • an explanation is provided with the first question 322 to provide context to the user (e.g., “This helps us understand your image preferences”).
  • At least two options are presented to the user for selection in response to the first question 322.
  • the user can select one option; however, in other examples, the user may select more than one option.
  • a first option 324a and a second option 324b are presented to the user for selection.
  • the options 324a, 324b correspond to example presentation slides.
  • the user may select an option by clicking (or tapping) on the option directly.
  • the user may select an option by selecting a button or other UI element.
  • FIG. 3C illustrates a view 330 of the client application 118 for determining the presentation preferences of the user for logical arguments.
  • the view 330 includes a second question 332 (e.g., “Which slide makes a better argument?”) that is presented to the user.
  • a second question 332 e.g., “Which slide makes a better argument?”
  • an explanation is provided with the second question 332 to provide context to the user (e.g., “This helps us understand your preference for logical arguments”).
  • At least two options are presented to the user for selection in response to the second question 332.
  • the user can select one option; however, in other examples, the user may select more than one option.
  • a first option 334a and a second option 334b are presented to the user for selection.
  • FIG. 3D illustrates a view 340 of the client application 118 for determining the data presentation preferences of the user.
  • the view 340 includes a third question 342 (e.g., “Which slide is more convincing to you?”) that is presented to the user.
  • a third question 342 e.g., “Which slide is more convincing to you?”
  • an explanation is provided with the third question 342 to provide context to the user (e g., “This choice helps us learn your data preferences”).
  • At least two options are presented to the user for selection in response to the third question 342. Tn some examples, the user can select one option; however, in other examples, the user may select more than one option.
  • a first option 344a and a second option 344b are presented to the user for selection.
  • FIG. 3E illustrates a view 350 of the client application 118 for determining the top (or most important) preferences of the user.
  • the view 350 includes a fourth question 352 (e.g., “When others present to you, what are some of your top preferences they should know?”) that is presented to the user.
  • a fourth question 352 e.g., “When others present to you, what are some of your top preferences they should know?”
  • an explanation is provided with the fourth question 352 to provide context to the user (e.g., “We will share this information with your colleagues, so they can build presentations that you will love!”).
  • a plurality of preference options 354 are presented to the user for selection in response to the third question 352.
  • the preference options 354 represent “pet peeves” of the user
  • the options included in the plurality of preference options 354 correspond to the top (or most popular) options of all users (or a subset of users) across the platform.
  • the fingerprint engine 106 may be configured to monitor and analyze the top preferences of users across the platform to derive the plurality of preference options 354.
  • the user can select one or more of the options 354.
  • the user can select option 356 to provide (e.g., write or type) their own specific preference(s).
  • the fingerprint generation process can include any number of questions or inquires.
  • the fingerprint generation process may include questions directed to image preferences, color preferences, illustration preferences, logical arguments preferences, data preferences, data visualization preferences, or any other type of relevant preferences of the user.
  • the various options e.g., 324a, 324b, etc.
  • the questions and/or corresponding options may be presented in a different format (e.g., text only, video, etc.).
  • more than two options can be presented with each question.
  • the fingerprint engine 106 is configured to generate a fingerprint for the user based on the user’s selections (or inputs). In some examples, the fingerprint engine 106 uses one or more algorithms to generate the fingerprint. In some examples, the fingerprint engine 106 uses one or more lookup tables to generate the fingerprint (e.g., stored in the application database 1 12b). Tn some examples, the fingerprint engine 106 includes a machine learning model to generate the fingerprint. In general, any suitable machine learning technique can be used, such as, for example: a gradient boosted random forest, a regression, a neural network, a decision tree, a support vector machine, a Bayesian network, other type of technique.
  • the fingerprint represents the visual preferences, story choice (e.g., storyline) preferences, data style preferences, and other personal preferences of the user.
  • the fingerprint may indicate one or more characteristics of the user.
  • FIG. 3F illustrates a view 360 of the client application 118 corresponding to a user fingerprint.
  • the fingerprint is classified with a fingerprint type 362 (e.g., “Performer”).
  • the fingerprint type 362 may be selected by the fingerprint engine 106 from a plurality of fingerprint types, such as, for example: architect, director, navigator, performer, producer, scholar, scientist, and surgeon.
  • the fingerprint type 362 has a corresponding symbol 364.
  • the fingerprint type 362 is selected based on one or more fingerprint attributes 366.
  • the user has a visual preference attribute 366a of “Vibrant” that is described as being “open to bold imagery and colors”, a story choice attribute 366b of “Deductive” that is described as “logic and facts first, punchline later”, and a data style attribute 366c of “Narrative” that is described as “support argument with data.”
  • the fingerprint includes personal preferences 368 of the user.
  • the personal preferences 368 may correspond to the user’s selections from the plurality of preference options 354 and/or user provided preferences.
  • the fingerprint for each user is stored in the user database 112b.
  • the fingerprint engine 106 can process and analyze existing content (e.g., a slideshow presentation) to derive a fingerprint. For example, a user may upload an example presentation via the client application 118. The fingerprint engine 106 can analyze the presentation to extract the visual preferences, story choice preferences, data style preferences, and other personal preferences of the user. In such examples, the fingerprint engine 106 may include a topic model (e.g., a Latent Dirichlet allocation (LDA) model) to process and analyze the user-provided content. In some examples, the fingerprint engine 109 may be configured to process and analyze XML data associated with the user-provided content.
  • LDA Latent Dirichlet allocation
  • FIG. 4 is a flow diagram of a method 400 for creating a content template in accordance with aspects described herein.
  • the method 400 corresponds to a process carried out by the application server 102 and the client application 118.
  • the fingerprint engine 106 receives at least one target audience parameter.
  • the “target audience” corresponds to the intended or expected audience that the content template is being used to deliver content to.
  • the at least one target audience parameter includes the names of members in the target audience.
  • the at least one target audience parameter includes the name of a group, team, club, organization, company, or another type of group including one or more individuals.
  • the at least one target audience parameter includes at least one fingerprint type (e.g., Architect, Surgeon, etc.).
  • the fingerprint engine 106 is configured to access, retrieve, or download the corresponding user fingerprints from the application database 112a.
  • the target audience parameters are provided to the fingerprint engine 106 from the client application 118.
  • the client application 118 may be configured to collect the various parameters from the user(s) creating the content template.
  • FIG. 5 illustrates a view 500 of the client application 118 for collecting template information from the user.
  • a text entry box is provided enabling the user to enter a title 502 for the content template (or the content).
  • the client application 118 may provide the title 502 to the fingerprint engine 106 or another engine of the application server 102 (e.g., the context engine 107).
  • a search function is provided such that the user can add one or more audience members 504 to the target audience.
  • a drop down menu or other UI element may be provided for the user to add audience members.
  • the user can add individuals who are users of the platform (e.g., have a fingerprint).
  • the user can add individuals who are not members of the platform (e.g., do not have a fingerprint).
  • the user may be prompted to provide contact information (e g., an email address) for each non-member such that an invite or fingerprint inquiry can be sent.
  • the client application 118 is configured to provide the one or more audience members 504 to the fingerprint engine 106 as the at least one audience parameter.
  • the context engine 107 receives at least one context parameter.
  • “context” corresponds to the intended type, purpose, and/or objective of the content that the content template is being created for.
  • the “context” of the content corresponds to the tone of the content (e g., good news, bad news, etc.).
  • the at least one context parameter includes a content type.
  • the content type may be one of: all-hands meeting, board of directors meeting, brand performance review, brand plan, budget overview, business case, business plan, communication plan, company introduction, company policy, competitor analysis, corporate responsibility update, customer deep dive, customer onboarding, customer pitch, employee performance review, financial performance, forecast update, go-to-market strategy, internal audit update, investor pitch, IT deep dive, journey line, knowledge transfer, leadership meeting, leadership review, M&A proposal, managers meeting, market research findings, monthly review, new employee orientation, new market entry, operational deep dive, organizational updates, partner deep dive, portfolio review, pricing plan, product development plan, product launch plan, product performance review, project update, purchasing decisions, R&D findings, recruitment update, request for proposal, rewards and recognition, roadshow, salesforce contests, staff meeting, staffing update, steering committee meeting, strategy plan, tactical plan, team meeting, town hall, training collateral, webinar, or any other suitable presentation type.
  • the context parameters are provided to the context engine 107 from the client application 118.
  • a drop down menu is provided such that the user can select a content type 506 for the content template.
  • a search function or other UI element may be provided for the user to select a content type 506.
  • the client application 118 is configured to provide the selected content type 506 to the context engine 107 as the at least one context parameter.
  • the content types displayed to the user are provided by the UI engine 110.
  • the context engine 107 may be configured to maintain and/or update the list of content types and provide a refreshed content type list to the UI engine 110.
  • users may upload or create their own custom content types.
  • custom content types may be stored in the application database 112a or the user database 112b.
  • the custom content types may be “private” and displayed only to one or more users (e.g., the creating user(s), users in the same organization, etc.).
  • the custom content types may be “public” and shared with all users once created.
  • the outline engine 108 generates an outline (or storyline) for the content template.
  • the outline engine 108 is configured to generate the storyline based on the at least one audience parameter and the at least one context parameter.
  • the fingerprint engine 106 may provide the at least one audience parameter and/or the corresponding user fingerprint(s) to the outline engine 108.
  • the context engine 107 may provide the at least one context parameter to the outline engine 108.
  • the outline engine 108 generates a storyline that is optimized relative to the content type (e g., all-hands meeting, brand plan, etc.) and the preferences of the target audience members (e.g., the fingerprints).
  • the outline engine 108 is configured to generate a storyline including a plurality of sections. Each section of the plurality of sections may include one or more content items (e.g., slides).
  • FIG. 6 illustrates a view 600 of the client application 118 for displaying a presentation storyline generated by the outline engine 108.
  • the storyline includes a plurality of sections 602. For example, a first section 602a represents a “Beginning” of the presentation, a second section 602b represents a “Middle” of the presentation, and a third section 602c represents an “End” of the presentation. It should be appreciated that in other examples, the storyline can include more or less sections and/or different types of sections.
  • each section 602 includes a set of slides.
  • the first section 602a includes a first set of slides (e.g., Title, Agenda, Executive Summary, Vision, Objectives, and Teams)
  • the second section 602b includes a second set of slides (e.g., Strategy, Success Factors, Insights, Messages, Recommendation)
  • the third section 602c includes a third set of slides (e.g., Milestones, What’s Next).
  • each section 602 includes a unique set of slides.
  • each slide type e.g., Insights, Messages, etc.
  • each slide type can be used multiple times throughout the presentation.
  • the first and third sections 602a, 602c may each include an “Insights” slide.
  • the user can modify the generated storyline. For example, the user may change the slide order within each section 602 to adjust the order generated by the outline engine 108. Likewise, the user can drag and drop slides to move slides between different sections. In some examples, the user can delete or remove slides added by the outline engine 108. In addition, the user may add slides that were not included in the storyline generated by the outline engine 108. For example, the user may drag and drop slides from an idea gallery 604. In some examples, the idea gallery 604 includes a search function that allows the user to search for new slides to add.
  • the outline engine 108 may provide a list of slides based on the at least one audience parameter and the at least one context parameter. For example, when a “custom” content type is selected, the outline engine may provide a slide list that the user can then organize and/or modify the storyline on their own. In some examples, the user can save the storyline as a custom storyline.
  • the style engine 109 selects (or creates) a style profde for the content template.
  • the style engine 109 is configured to select (or create) the style profile based on the at least one audience parameter (e.g., user fingerprints), the at least one context parameter, and the content template storyline.
  • the user fingerprints can be used to determine the image preferences, color preferences, illustration preferences, presentation preferences for logical arguments, presentation preferences for data, and/or preferences for visualizing data for each audience member.
  • the style profile incudes one or more stylistic attributes of the content template and/or the content.
  • the style engine 109 is configured to select a pre-rendered style profile from a plurality of style profiles (e.g., a style profile library stored in the application database 112a).
  • the style engine 109 is configured to create a style profile including one or more of a template theme, a font theme, a color theme, and an organizational theme.
  • the style profile includes XML data that can be used to implement the corresponding style attributes in various content editing tools (e.g., Microsoft PowerPoint, Google Slides, etc.).
  • a weighting system may be used when the target audience includes two or more individuals. For example, a weight (or ranking) may be assigned to each audience member. When two or more audience members have conflicting preferences, the weights (or ranks) of each member can be used to determine which preference is adopted by the content template (e.g., use preference of highest rank). In some examples, the weights (or ranks) are assigned by the user creating the content template. In other examples, the weights (or ranks) may be assigned by client application 118 or the application server 102. For example, a title (e.g., CEO, CFO, etc.) or description (e.g., Boss, Manager, etc.) associated with each audience member may be used to determine the weighting/ranking. Alternatively, conflicting preferences may be split across the target audience such that each member’s preferences are equally represented in the content template.
  • a title e.g., CEO, CFO, etc.
  • description e.g., Boss, Manager, etc.
  • conflicting preferences may be split
  • the user may change the at least one target audience parameter and/or the at least one context parameter. For example, the user may add members to the target audience and/or remove members from the target audience to change the at least one target audience parameter.
  • the fingerprint engine 106 may detect a change to the fingerprint of one or more audience members and update the at least one target audience parameter accordingly.
  • the user may modify the at least one context parameter by changing the context type.
  • the changes to the at least one target audience parameter and/or the at least one context parameter are provided to the outline engine 108.
  • the outline engine 108 is configured to dynamically update the storyline of the content template (step 406).
  • the style engine 109 is configured to dynamically modify the style profile for the content template based on the updated storyline, the changes to the at least one audience parameter, and/or the changes to the at least one context parameter (step 408).
  • the content template is generated by the style engine 109.
  • the content template is generated by applying the style profile to the storyline generated by the outline engine 108.
  • the style profile is applied by selecting one or more slides included in the storyline from a slide library with style attributes matching the style profile.
  • the style profile is applied by creating one or more slides included in the storyline with style attributes matching the style profile.
  • FIG. 7 illustrates a view 700 of the client application 118 for displaying the content template.
  • the user may view (or preview) the content template before downloading.
  • the user can download the content template by selecting the download button 702.
  • the user is presented with a prompt to select a desired fde format (or type) for the content template.
  • the UI engine 110 may provide an export service that allows the content template to be downloaded in various formats (e g., PDF, Microsoft Word, Microsoft PowerPoint, etc.).
  • the content template can be directly imported to a content editing tool (e g., Microsoft PowerPoint, Google Slides, etc.).
  • the user may utilize one or more content editing tools to “fill in” the content template with the corresponding content.
  • the user may modify one or more features or characteristics of the content template. For example, the user may review the content template and decide to change the template’s node count, layout, imagery, font, color, work area, shapes, charts, graphs, icons, text, brand, or any combination thereof.
  • the changes (or change requests) identified by the user are received via the client application 118 and directed to the corresponding engine of the application server 102 (e.g., the style engine 109).
  • the style engine 109 is configured to auto-transform the content template based on the user’s changes and modifications.
  • the style engine 109 is configured to apply the user’s changes by dynamically generating a new content template (step 412).
  • FIG. 8 is a flow diagram of a method 800 for transforming content in accordance with aspects described herein.
  • the method 800 corresponds to a process carried out by the application server 102 and the client application 118.
  • the application server 102 receives a user-uploaded presentation.
  • the presentation corresponds to a slide deck.
  • the presentation may be uploaded by the user via the client application 118.
  • the various engines of the application server 102 may analyze the presentation to identify and classify content included in the presentation.
  • the outline engine 108 and/or the style engine 109 may deconstruct the presentation and analyze the content of the presentation.
  • the engines 108, 109 can include a topic model (e g., an LDA model) to process and analyze the content of the presentation.
  • the engines 108, 109 are configured to process and analyze XML data associated with the content of the presentation.
  • the engines 108, 109 include a machine learning model configured to analyze and process the content of the presentation.
  • a machine learning model configured to analyze and process the content of the presentation.
  • any suitable machine learning technique can be used, such as, for example: a gradient boosted random forest, a regression, a neural network, a decision tree, a support vector machine, a Bayesian network, other type of technique.
  • the content of the presentation is extracted from the presentation and stored in one or more content data structures (e.g., in the application database 112a).
  • the data from a chart included in the user-uploaded presentation may be extracted and saved as a content data structure having a content tag of “Chart”.
  • the data (or content) from an executive summary included in the user-uploaded presentation may be extracted and saved as a data structure having a content tag of “Executive Summary”, and so on.
  • the engines 108, 109 can add tags to the slides and/or objects of the user-uploaded presentation.
  • the tags may indicate that the slides and/or objects include transformable content that will be tailored to the target audience.
  • the tags may provide a link (or location) to the corresponding content data structure saved in the application database 112a.
  • the outline engine 108 and/or the style engine 109 may “sanitize” the user-uploaded presentation.
  • the engines 108, 109 are configured to sanitize the presentation by removing any references to confidential, proprietary, private, and/or nonpublic information from the presentation (or the content of the presentation).
  • the sanitization process may generalize the presentation.
  • the level of sanitization provided may be adjusted by the user via the client application 118. In some examples, the user may be prompted to review the results of the sanitization process before proceeding with the content transformation.
  • the fingerprint engine 106 receives at least one target audience parameter.
  • the “target audience” corresponds to the intended or expected audience that the content is being tailored to.
  • the at least one target audience parameter includes the names of members in the target audience.
  • the at least one target audience parameter includes the name of a group, team, club, organization, company, or another type of group including one or more individuals.
  • the at least one target audience parameter includes at least one fingerprint type (e.g., Architect, Surgeon, etc.).
  • the fingerprint engine 106 is configured to access, retrieve, or download the corresponding user fingerprints from the application database 112a.
  • the target audience parameters are provided to the fingerprint engine 106 from the client application 118.
  • the client application 118 may be configured to collect the various parameters from the user(s) transforming the content.
  • the context engine 107 receives at least one context parameter.
  • “context” corresponds to the intended type, purpose, and/or objective of the content.
  • the “context” of the content corresponds to the tone of the content (e.g., good news, bad news, etc.).
  • the at least one context parameter includes a content type.
  • the context parameters are provided to the context engine 107 from the client application 118. In other examples, the context engine 107 may be configured to derive the context type from the user-uploaded presentation (or the included content).
  • the outline engine 108 generates an outline (or storyline) for the content.
  • the outline engine 108 is configured to generate the storyline based on the at least one audience parameter and the at least one context parameter.
  • the fingerprint engine 106 may provide the at least one audience parameter and/or the corresponding user fingerprint(s) to the outline engine 108.
  • the context engine 107 may provide the at least one context parameter to the outline engine 108.
  • the outline engine 108 generates a storyline that is optimized relative to the content type (e.g., all-hands meeting, brand plan, etc.) and the preferences of the target audience members (e.g., the fingerprints).
  • the outline engine 108 is configured to generate a storyline including a plurality of sections. Each section of the plurality of sections may include one or more content items (e.g., slides).
  • the outline engine 108 may provide a list of slides based on the at least one audience parameter and the at least one context parameter. For example, when a “custom” content type is selected, the outline engine may provide a slide list that the user can then organize and/or modify the storyline on their own. In some examples, the user can save the storyline as a custom storyline.
  • the style engine 109 selects (or creates) a style profile for the content.
  • the style engine 109 is configured to select (or create) the style profile based on the at least one audience parameter (e.g., user fingerprints), the at least one context parameter, and the content storyline.
  • the user fingerprints can be used to determine the image preferences, color preferences, illustration preferences, presentation preferences for logical arguments, presentation preferences for data, and/or preferences for visualizing data for each audience member.
  • the style profile incudes one or more stylistic attributes of the content.
  • the style engine 109 is configured to select a pre-rendered style profile from a plurality of style profiles (e.g., a style profile library stored in the application database 112a).
  • the style engine 109 is configured to create a style profile including one or more of a template theme, a font theme, a color theme, and an organizational theme.
  • the style profile includes XML data that can be used to implement the corresponding style attributes in various content editing tools (e.g., Microsoft PowerPoint, Google Slides, etc.).
  • a weighting system may be used when the target audience includes two or more individuals. For example, a weight (or ranking) may be assigned to each audience member. When two or more audience members have conflicting preferences, the weights (or ranks) of each member can be used to determine which preference is adopted by the content (e.g., use preference of highest rank). In some examples, the weights (or ranks) are assigned by the user transforming the content. In other examples, the weights (or ranks) may be assigned by client application 118 or the application server 102. For example, a title (e g., CEO, CFO, etc.) or description (e.g., Boss, Manager, etc.) associated with each audience member may be used to determine the weighting/ranking. Alternatively, conflicting preferences may be split across the target audience such that each member’s preferences are equally represented in the content.
  • a title e g., CEO, CFO, etc.
  • description e.g., Boss, Manager, etc.
  • conflicting preferences may be split across the
  • the content is auto-transformed by the style engine 109.
  • the content is transformed by applying the style profile to the storyline generated by the outline engine 108.
  • the style profile is applied by selecting one or more slides included in the storyline from a slide library with style attributes matching the style profile.
  • the style profile is applied by creating one or more slides included in the storyline with style attributes matching the style profile.
  • the style engine 109 is configured to access the tagged content data stored in the application database 112a and integrate the content data into the presentation based on the content storyline and the style profile.
  • the transformed presentation is analyzed to find the slides and/or objects corresponding to the tagged content data.
  • the tagged content data is updated (e g., in the application database 112a) based on the representation of the content in the transformed presentation.
  • the tagged content data is saved and used to generate subsequent presentations (e g., tailored to a plurality of target audiences).
  • the tagged content data is used to generate different forms of content. For example, the tagged content data from a presentation may be used to generate a memo, email, etc.
  • the user may change the at least one target audience parameter and/or the at least one context parameter. For example, the user may add members to the target audience and/or remove members from the target audience to change the at least one target parameter.
  • the fingerprint engine 106 may detect a change to the fingerprint of one or more audience members and may update the at least one target audience parameter accordingly.
  • the user may modify the at least one context parameter by changing the context type.
  • the changes to the at least one target audience parameter and/or the at least one context parameter are provided to the outline engine 108.
  • the outline engine 108 is configured to dynamically update the storyline of the content (step 810).
  • the style engine 109 is configured to dynamically modify the style profile for the content based on the updated storyline, the changes to the at least one audience parameter, and/or the changes to the at least one context parameter (step 812).
  • the UI engine 110 may provide an export service that allows the content to be downloaded in various formats (e g., PDF, Microsoft Word, Microsoft PowerPoint, etc.).
  • the content can be directly imported to a content editing tool (e.g., Microsoft PowerPoint, Google Slides, etc.).
  • FIG. 9 illustrates a slide library 900 in accordance with aspects described herein.
  • the user can access the slide library 900 by selecting the slides button 206 of FIG. 2.
  • the slide library 900 may be stored in the application database 112a and managed by the engines 106-1 10 of the application server 102.
  • the user may access the slide library 900 to preview slides corresponding to different user fingerprint, context types, and style profiles.
  • the user may use the interface to assist in selecting the appropriate context type when creating a new content template.
  • the user may use the interface to check the slides associated with their own fingerprint.
  • FIG. 10 illustrates a template management interface 1000 of the client application 118.
  • the user can access the template management interface 1000 by selecting the presentation button 204 of FIG. 2.
  • the contents of the template management interface 1000 may be stored in the application database 112a and/or the user database 112b and managed by the UI engine 110 of the application server 102.
  • the user may use the interface 1000 to view content templates they have created (e.g., via tab 1002).
  • the user may use the interface 1000 to view content templates that have been assigned to them for review (e.g., via tab 1004).
  • the user may use the interface 1000 to view content templates they have created or are creating in collaboration with other users (e.g., via tab 1006).
  • the system 100 can be adapted for other types of content.
  • the system 100 can be configured to generate templates for emails, memos, spreadsheets, reports, letters, or any other type of content or media.
  • the system 100 can generate content templates that are compatible with various content editing tools and platforms, such as, for example: Microsoft PowerPoint, Microsoft Office, Microsoft Excel, Google Slides, Microsoft Outlook, Gmail, Adobe Acrobat, etc.
  • a content storyline is generated based on parameters of the target audience and the context of the content.
  • the parameters of the target audience correspond to a preference fingerprint of the audience member(s).
  • a style profile is selected or created for the content template based on the parameters of the target audience, the context of the content, and the storyline.
  • Hardware and Software Implementations FTG. 1 1 shows an example of a generic computing device 1 100, which may be used with some of the techniques described in this disclosure (e g., as user device 116 or application server 102).
  • Computing device 1100 includes a processor 1102, memory 1104, an input/output device such as a display 1106, a communication interface 1108, and a transceiver 1110, among other components.
  • the device 1100 may also be provided with a storage device, such as a micro-drive or other device, to provide additional storage.
  • a storage device such as a micro-drive or other device, to provide additional storage.
  • Each of the components 1100, 1102, 1104, 1106, 1108, and 1110 are interconnected using various buses, and several of the components may be mounted on a common motherboard or in other manners as appropriate.
  • the processor 1102 can execute instructions within the computing device 1100, including instructions stored in the memory 1104.
  • the processor 1102 may be implemented as a chipset of chips that include separate and multiple analog and digital processors.
  • the processor 1102 may provide, for example, for coordination of the other components of the device 1100, such as control of user interfaces, applications run by device 1100, and wireless communication by device 1100.
  • Processor 1102 may communicate with a user through control interface 1112 and display interface 1114 coupled to a display 1106.
  • the display 1106 may be, for example, a TFT LCD (Thin-Film-Transistor Liquid Crystal Display) or an OLED (Organic Light Emitting Diode) display, or other appropriate display technology.
  • the display interface 1114 may comprise appropriate circuitry for driving the display 1106 to present graphical and other information to a user.
  • the control interface 1112 may receive commands from a user and convert them for submission to the processor 1102.
  • an external interface 1116 may be provided in communication with processor 1102, so as to enable near area communication of device 1100 with other devices.
  • External interface 1116 may provide, for example, for wired communication in some implementations, or for wireless communication in other implementations, and multiple interfaces may also be used.
  • the memory 1104 stores information within the computing device 1100.
  • the memory 1104 can be implemented as one or more of a computer-readable medium or media, a volatile memory unit or units, or a non-volatile memory unit or units.
  • Expansion memory 1118 may also be provided and connected to device 1100 through expansion interface 1120, which may include, for example, a SIMM (Single In Line Memory Module) card interface.
  • SIMM Single In Line Memory Module
  • expansion memory 1 118 may provide extra storage space for device 1 100, or may also store applications or other information for device 1100.
  • expansion memory 1118 may include instructions to carry out or supplement the processes described above, and may include secure information also.
  • expansion memory 1118 may be provided as a security module for device 1100, and may be programmed with instructions that permit secure use of device 1100.
  • secure applications may be provided via the SIMM cards, along with additional information, such as placing identifying information on the SIMM card in a non-hackable manner.
  • the memory may include, for example, flash memory and/or NVRAM memory, as discussed below.
  • a computer program product is tangibly embodied in an information carrier.
  • the computer program product contains instructions that, when executed, perform one or more methods, such as those described above.
  • the information carrier is a computer- or machine-readable medium, such as the memory 1104, expansion memory 1118, memory on processor 1102, or a propagated signal that may be received, for example, over transceiver 1110 or external interface 1116.
  • Device 1100 may communicate wirelessly through communication interface 1108, which may include digital signal processing circuitry where necessary.
  • Communication interface 1108 may in some cases be a cellular modem.
  • Communication interface 1108 may provide for communications under various modes or protocols, such as GSM voice calls, SMS, EMS, or MMS messaging, CDMA, TDMA, PDC, WCDMA, CDMA2000, or GPRS, among others.
  • GSM voice calls SMS, EMS, or MMS messaging
  • CDMA Code Division Multiple Access
  • TDMA Time Division Multiple Access
  • PDC Wideband Code Division Multiple Access
  • WCDMA Code Division Multiple Access 2000
  • GPRS Global System
  • Device 1100 may also communicate audibly using audio codec 1124, which may receive spoken information from a user and convert it to usable digital information. Audio codec 1124 may likewise generate audible sound for a user, such as through a speaker, e.g., in a handset of device 1100. Such sound may include sound from voice telephone calls, may include recorded sound (e g., voice messages, music fdes, etc.) and may also include sound generated by applications operating on device 1100. Tn some examples, the device 1100 includes a microphone to collect audio (e.g., speech) from a user. Likewise, the device 1100 may include an input to receive a connection from an external microphone.
  • audio codec 1124 may receive spoken information from a user and convert it to usable digital information. Audio codec 1124 may likewise generate audible sound for a user, such as through a speaker, e.g., in a handset of device 1100. Such sound may include sound from voice telephone calls, may include recorded sound (e g., voice
  • the computing device 1100 may be implemented in a number of different forms, as shown in FIG. 11. For example, it may be implemented as a computer (e.g., laptop) 1126. It may also be implemented as part of a smartphone 1128, smart watch, tablet, personal digital assistant, or other similar mobile device.
  • a computer e.g., laptop
  • smartphone e.g., smart watch, tablet, personal digital assistant, or other similar mobile device.
  • implementations of the subject matter and the operations described in this specification can be implemented in digital electronic circuitry, or in computer software, firmware, or hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them.
  • Implementations of the subject matter described in this specification can be implemented as one or more computer programs, i.e., one or more modules of computer program instructions, encoded on computer storage medium for execution by, or to control the operation of, data processing apparatus.
  • the program instructions can be encoded on an artificially-generated propagated signal, e.g., a machine-generated electrical, optical, or electromagnetic signal, that is generated to encode information for transmission to suitable receiver apparatus for execution by a data processing apparatus.
  • a computer storage medium can be, or be included in, a computer-readable storage device, a computer-readable storage substrate, a random or serial access memory array or device, or a combination of one or more of them.
  • a computer storage medium is not a propagated signal, a computer storage medium can be a source or destination of computer program instructions encoded in an artificially-generated propagated signal.
  • the computer storage medium can also be, or be included in, one or more separate physical components or media (e.g., multiple CDs, disks, or other storage devices).
  • the operations described in this specification can be implemented as operations performed by a data processing apparatus on data stored on one or more computer-readable storage devices or received from other sources.
  • the term “data processing apparatus” encompasses all kinds of apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, a system on a chip, or multiple ones, or combinations, of the foregoing.
  • the apparatus can include special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application-specific integrated circuit).
  • the apparatus can also include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, a cross-platform runtime environment, a virtual machine, or a combination of one or more of them.
  • the apparatus and execution environment can realize various different computing model infrastructures, such as web services, distributed computing and grid computing infrastructures.
  • a computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, declarative or procedural languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, object, or other unit suitable for use in a computing environment.
  • a computer program may, but need not, correspond to a file in a file system.
  • a program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language resource), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub-programs, or portions of code).
  • a computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
  • the processes and logic flows described in this specification can be performed by one or more programmable processors executing one or more computer programs to perform actions by operating on input data and generating output.
  • the processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application-specific integrated circuit).
  • processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer.
  • a processor will receive instructions and data from a read-only memory or a random access memory or both.
  • the essential elements of a computer are a processor for performing actions in accordance with instructions and one or more memory devices for storing instructions and data.
  • a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto-optical disks, or optical disks.
  • mass storage devices for storing data, e.g., magnetic, magneto-optical disks, or optical disks.
  • a computer need not have such devices.
  • a computer can be embedded in another device, e.g., a mobile telephone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a Global Positioning System (GPS) receiver, or a portable storage device (e.g., a universal serial bus (USB) flash drive), to name just a few.
  • Devices suitable for storing computer program instructions and data include all forms of non-volatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks.
  • the processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
  • implementations of the subject matter described in this specification can be implemented on a computer having a display device, e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor, for displaying information to the user and a keyboard and a pointing device, e.g., a mouse or a trackball, by which the user can provide input to the computer.
  • a display device e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor
  • keyboard and a pointing device e.g., a mouse or a trackball
  • Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback, e.g., visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input.
  • a computer can interact with a user by sending resources to and receiving resources from a device that is used
  • Implementations of the subject matter described in this specification can be implemented in a computing system that includes a back-end component, e.g., as a data server, or that includes a middleware component, e.g., an application server, or that includes a front-end component, e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back-end, middleware, or front-end components.
  • the components of the system can be interconnected by any form or medium of digital data communication, e.g., a communication network.
  • Examples of communication networks include a local area network (“LAN”) and a wide area network (“WAN”), an inter-network (e.g., the Internet), and peer-to-peer networks (e.g., ad hoc peer-to-peer networks).
  • LAN local area network
  • WAN wide area network
  • inter-network e.g., the Internet
  • peer-to-peer networks e.g., ad hoc peer-to-peer networks.
  • the computing system can include clients and servers.
  • a client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
  • a server transmits data (e.g., an HTML page) to a client device (e.g., for purposes of displaying data to and receiving user input from a user interacting with the client device).
  • client device e.g., for purposes of displaying data to and receiving user input from a user interacting with the client device.
  • Data generated at the client device e., a result of the user interaction
  • a system of one or more computers can be configured to perform particular operations or actions by virtue of having software, firmware, hardware, or a combination of them installed on the system that in operation causes or cause the system to perform the actions.
  • One or more computer programs can be configured to perform particular operations or actions by virtue of including instructions that, when executed by data processing apparatus, cause the apparatus to perform the actions.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Strategic Management (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Human Resources & Organizations (AREA)
  • Data Mining & Analysis (AREA)
  • Economics (AREA)
  • Marketing (AREA)
  • Operations Research (AREA)
  • Quality & Reliability (AREA)
  • Tourism & Hospitality (AREA)
  • General Business, Economics & Management (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

Procédé de création d'un modèle de contenu consistant à recevoir au moins un paramètre de public représentant au moins une préférence d'un public cible lors de la réception du contenu, à recevoir au moins un paramètre de contexte représentant un contexte du contenu, à déterminer un scénario de contenu sur la base du ou des paramètres d'audience et du ou des paramètres de contexte, et à générer un modèle de contenu sur la base du scénario de contenu.
PCT/US2023/021559 2022-05-09 2023-05-09 Systèmes et procédés de création de contenu sur la base d'une préférence de public et de facteurs contextuels WO2023220068A1 (fr)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US202263339827P 2022-05-09 2022-05-09
US63/339,827 2022-05-09
US18/156,649 US11775748B1 (en) 2022-05-09 2023-01-19 Systems and methods for content creation based on audience preference and contextual factors
US18/156,649 2023-01-19

Publications (1)

Publication Number Publication Date
WO2023220068A1 true WO2023220068A1 (fr) 2023-11-16

Family

ID=86710754

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2023/021559 WO2023220068A1 (fr) 2022-05-09 2023-05-09 Systèmes et procédés de création de contenu sur la base d'une préférence de public et de facteurs contextuels

Country Status (1)

Country Link
WO (1) WO2023220068A1 (fr)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3014862A1 (fr) * 2013-06-24 2016-05-04 Microsoft Technology Licensing, LLC Présentation automatique de suggestions de conception de diapositive
US20180196784A1 (en) * 2017-01-10 2018-07-12 Microsoft Technology Licensing, Llc Dynamic content generation

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3014862A1 (fr) * 2013-06-24 2016-05-04 Microsoft Technology Licensing, LLC Présentation automatique de suggestions de conception de diapositive
US20180196784A1 (en) * 2017-01-10 2018-07-12 Microsoft Technology Licensing, Llc Dynamic content generation

Similar Documents

Publication Publication Date Title
Zhan et al. Unlocking the power of big data in new product development
US10984180B2 (en) Electronic document supplementation with online social networking information
US20170337261A1 (en) Decision Making and Planning/Prediction System for Human Intention Resolution
CN107533670B (zh) 数字实体的预测热门
US20180233057A1 (en) Proactive content recommendation in teaching space
Conole et al. Visualising learning design to foster and support good practice and creativity
Rego et al. Metadata and knowledge management driven web-based learning information system towards web/e-learning 3.0
WO2006016208A2 (fr) Infrastructure de soutien personnel pour developpement d'applications et d'interfaces utilisateur
Owoseni et al. Evaluating mobile app usage by service sector micro and small enterprises in Nigeria: an abductive approach
KR20180039019A (ko) 상황-특정 기록 프레임워크와 관련된 방법 및 시스템
Conlan et al. The narrative approach to personalisation
US11714967B1 (en) College admissions and career mentorship platform
US11775748B1 (en) Systems and methods for content creation based on audience preference and contextual factors
Tapia et al. A process to support the remote tree testing technique for evaluating the information architecture of user interfaces in software projects
Shivakumar Build a Next-Generation Digital Workplace: Transform Legacy Intranets to Employee Experience Platforms
US20220343233A1 (en) Systems and Methods for Data Analytics
Tram Firebase
WO2023220068A1 (fr) Systèmes et procédés de création de contenu sur la base d'une préférence de public et de facteurs contextuels
Vashisth et al. Hype cycle for data science and machine learning, 2019
Das et al. Improving E-Governance Through Application of Hyperautomation
US20170076009A1 (en) Systems and methods for presenting events
Christman et al. BIBFRAME Basics: A Crash Course
Bettinson Enabling Large-Scale Collaboration in Language Conservation
Petri et al. Surmounting Obstacles for Academic Resilience: A Dynamic Portal for Supporting an Alliance of Students with Disabilities
KR102509501B1 (ko) 사업계획서 자동작성장치 및 방법

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23729232

Country of ref document: EP

Kind code of ref document: A1