EP3972709A1 - Empfehlungen von inhaltselementen - Google Patents

Empfehlungen von inhaltselementen

Info

Publication number
EP3972709A1
EP3972709A1 EP20836901.7A EP20836901A EP3972709A1 EP 3972709 A1 EP3972709 A1 EP 3972709A1 EP 20836901 A EP20836901 A EP 20836901A EP 3972709 A1 EP3972709 A1 EP 3972709A1
Authority
EP
European Patent Office
Prior art keywords
score data
user
popularity
data
client device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
EP20836901.7A
Other languages
English (en)
French (fr)
Other versions
EP3972709A4 (de
Inventor
Anthony John COX
Christian Carollo
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Valve Corp
Original Assignee
Valve Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Valve Corp filed Critical Valve Corp
Publication of EP3972709A1 publication Critical patent/EP3972709A1/de
Publication of EP3972709A4 publication Critical patent/EP3972709A4/de
Pending legal-status Critical Current

Links

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/70Game security or game management aspects
    • A63F13/79Game security or game management aspects involving player-related data, e.g. identities, accounts, preferences or play histories
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/60Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor
    • A63F13/67Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor adaptively or by learning from player actions, e.g. skill level adjustment or by storing successful combat sequences for re-use
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/60Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor
    • A63F13/69Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor by enabling or updating specific game elements, e.g. unlocking hidden features, items, levels or versions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/435Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/903Querying
    • G06F16/9035Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9535Search customisation based on user profiles and personalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9536Search customisation based on social or collaborative filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning

Definitions

  • FIG. 1 is a diagram illustrating an example environment that includes a remote computing system configured to train and use machine-learning model(s) for recommending content items to different users.
  • the remote computing system trains multiple models and, thereafter, inputs a consumption history of a user into each of the models for generating score data indicating a correlation between each of multiple content- item titles and the consumption history.
  • the computing system may use one or more biasing factors to generate result data, which the computing system may send to a client device of a user. The client device may use this result data, as well as different user inputs, to determine order(s) in which to present content-item recommendations to the user.
  • FIG. 2 illustrates an example process of operations for generating machine -learning models, applying them to history data of a particular user, sending generated result data to a client device of the user, and using this result data and received user input to determine order(s) in which to present content-item recommendations to the user.
  • FIG. 3 illustrates example components, and interplay thereof, for generating the models, biasing the resulting scores, applying the models to history data of a particular user, and using this result data and received user input to determine order(s) in which to present content-item recommendations to the user.
  • FIG. 4 illustrates example result data that the remote computing system may generate and send to a client device of a user.
  • the result data may comprise an MxN matrix of scores, which the client device may interpolate between as needed in response to receiving varying user input.
  • FIG. 5 illustrates an example user interface (UI) that a client device may present in response to receiving a request from a user for content-item recommendations.
  • UI user interface
  • FIG. 6 illustrates an example UI that the client device may present in response to the user altering a value of a first parameter (in this example, popularity of the content items) and a value of a second parameter (in this example, recency of the content items).
  • a first parameter in this example, popularity of the content items
  • a second parameter in this example, recency of the content items
  • FIG. 7 illustrates a flow diagram of an example process that the remote computing system may employ for generating one or more trained machine-learning models and generating result data for use by a client device in providing content-item recommendations to a user.
  • FIG. 8 illustrates a flow diagram of an example process that a client device of a user may employ for determining an order in which to present one or more content items to the user.
  • Described herein are, among other things, techniques, devices, and systems for generating one or more trained machine-learning models used for generating content-item recommendations. Also described herein are techniques, devices, and systems for applying a consumption history of a particular user to the trained model(s) to generate score data indicating a correlation between each content-item title and the consumption history, as well as modifying this score data using one or more biasing factors for generating result data. In addition, the techniques, devices, and systems may use this result data, along with received user input, for determining an order in which to present one or more content items to the user. For example, this may include determining which content items to recommend to a user and in which order to do so.
  • the disclosed techniques may be implemented, at least in part, by a remote computing system that distributes content items (e.g., video games, movies, television shows, songs, etc.) to client devices of a user community (e.g., as part of a video game service, movie service, song service, etc.). These client devices may individually install a client application that is configured to execute the content items received (e.g., downloaded, streamed, etc.) from the remote computing system.
  • the video-game platform enables registered users of the community to play video games as“players.” For example, a user can load the client application, login with a registered user account, select a desired video game, and execute the video game on his/her client machine via the client application.
  • content items e.g., video games, movies, television shows, songs, etc.
  • client devices of a user community e.g., as part of a video game service, movie service, song service, etc.
  • client devices may individually install a client application that is configured to execute the content items received
  • the computing system which is typically remote from client devices of individual users, generates multiple trained models, each associated with a particular value of a parameter. For example, the computing system may generate a first trained model for use in recommending content items that have been released within a first time window (e.g., the prior six months), a second trained model for use in recommending content items that have been released in a second time window (e.g., the prior year), and so forth. As will be described below, these models may applying to individual user data for generating tailored content-item recommendations.
  • a first time window e.g., the prior six months
  • a second trained model for use in recommending content items that have been released in a second time window
  • these models may applying to individual user data for generating tailored content-item recommendations.
  • the computing system may train these models in any suitable manner, in one example the computing system begins by accessing respective consumption histories (e.g., gameplay histories, moving histories, etc.) of multiple users associated with the computing-system community. For example, the computing system may access a gameplay history of a first user of a video-game-service community and may use this history for training the first model. For example, if the first model is for use in recommending game titles that have been released within the past six months, the computing system may select, from the gameplay history of the user, one game title that has been released within the past six months and may indicate to the first model that this selected game title is to be the output of the classifier.
  • respective consumption histories e.g., gameplay histories, moving histories, etc.
  • the computing system may access a gameplay history of a first user of a video-game-service community and may use this history for training the first model.
  • the computing system may select, from the gameplay history of the user, one game title that has been released within the past six months and may indicate to the first model that this
  • the computing system may provide the entire gameplay history of the user as input to the first model and the selected game title as output of the first model for training one or more internal layers of the first model.
  • the first model comprises an artificial neural network
  • selecting the game title as output and providing information regarding the gameplay history of the user as input may be effective to train one or more internal layers of the neural network.
  • the input and output of the neural network, and other types of models that may be used may comprise features of the corresponding game titles or other content items, such as release data, genre, game-type, and/or the like.
  • the input to the models may include information associated with the user, such as a geographical location associated with the user, demographic information associated with the user, and/or the like.
  • any other type of input may be provided to the models.
  • the computing system may take into account the amount of time that the user has played each game in her history in selecting the game title to be used as output for the machine-learning model. For example, the computing system may initially determine an amount of time that the user has played each individual game title referenced in the gameplay history of the user. This may be determined either absolutely (e.g., in terms of time played) or relatively (e.g., in terms of time played relative to other users that have also played the game title). This information may be applied as a weight to the individual game title or may be passed into the model as an input.
  • the computing system may then select one (or more) game title to be used as output by the first model. By doing so, the computing system increases the chances that the games played the most by the user are selected as the output of the first model, thus training the internal layer(s) of the first model in a way that results in the model being more likely to select games of interest to the user. It is to be appreciated, however, that while the above description illustrates one manner in which to train the models, other models for outputting content-item recommendations may be trained in any other number of ways.
  • the computing device may also train multiple other models, such as a model for recommending games released within the last year, a model for recommending games released within the last three years, and so forth.
  • the computing system may access a gameplay history of a user and select, as the output to the particular model, a game title that is both indicated in the gameplay history and that was released within the amount of time associated with the particular model (e.g., within one year, three years, etc.).
  • the computing system may analyze thousands or more of the gameplay histories for training the models.
  • the computing system may then use the models for generating data that may be used by client devices for surfacing content-item (e.g., video game) recommendations. For example, envision that a particular user sends a request, via a client device, for video-game recommendations.
  • the computing system may identify the user (e.g., via a device or user identifier) and may access a gameplay history of the user. The computing system may then input the gameplay history into each of the multiple machine -learning models along with information regarding available games associated with each model.
  • the computing system may input the gameplay history of the user and information regarding games available to be recommended
  • the computing system may input the gameplay history of the user and information regarding those games available to be recommended, and so forth. That is, while the entire catalog of available games may be input to the first model, the second model, and so forth, the first model may be more likely to recommend games having been released within the last six months, the second model may be more likely to recommend games that have been released within the last year, and so forth.
  • Each trained machine-learning model may output score data that indicates, for each of multiple game titles, a correlation between the gameplay history of the user and the particular game title. As the reader will appreciate, those games that are highly correlated to the gameplay history will have a relatively high score, and vice versa. Further, if the computing system implements“N” number of models, then the computing system may generate“N” number of score data. It is also to be appreciated that the“correlation” might not be a statistical correlation, but rather simply a representation of a likelihood that a user with the same gameplay history of the subject user would choose to play this game.
  • the computing system may be configured to bias the score data based on different values of one or more biasing factors.
  • the biasing factors may comprise any number of factors, such as a popularity of each game, a cost of each game, a genre of each game, or the like.
  • the biasing factor may comprise a popularity of each game, as determined by sales of the game, an amount of gameplay of the game, and/or the like.
  • each score data (determined by an individual trained model) may be biased at different levels of popularity such that the resulting biased data will more closely match that particular level of popularity if so desired by the user.
  • the computing system may bias the score for each game title (and for each model) at a number of“M” target levels of popularity ranging from very popular to very unpopular (or very“niche”).
  • the computed scores of those game titles that are very popular will be boosted for a target level of popularity of“very high” and will be penalized for a target level of popularity of“very low”.
  • the computed scores of more niche game titles will be penalized for a target level of popularity of“very high” and will be boosted for a target level of popularity of“very low”.
  • computing system may now have generated an amount of scores for the user in an amount of“M x N” .
  • the computing system may have generated a total of thirty scores for any particular user (given that each of the six score data generated by the model will biased five times).
  • the computed score data may have any other number of values and may be sampled in any other (e.g., nonlinear) manner.
  • any other number of parameters may be applied as input to the model and/or as part of the postprocessing described with reference to the popularity data.
  • the postprocessing parameters are described as biasing factors, these biasing factor(s) represent parameter(s) that may be applied as inputs to the models or as part of postprocessing.
  • the computing system may then send this“result data” to the client device of the user.
  • the client device of the user may then use this result data, as well as input from the user, for determining an order in which to present one or more of the game titles available to the user. That is, the client device may use the result data and input data from the user for determining which video games to recommend to the user, and how to change this recommendation as the user provides different input data.
  • the client device may enable the user to provide, as input, a selection of a desired release recency and a selection of a desired level of popularity.
  • the user may indicate, to the client device, whether the user prefers to see, as game recommendations, games that have been more recently released or games that are older.
  • the user may specify whether she would like to receive recommendations for games that are in the mainstream (that is, popular), or games that are more niche (that is, less popular). While the examples below illustrate and describe the user as providing this input via one or more sliders, it is to be appreciated that this input may be provided in any other manner.
  • the client device may use the result data to determine which games to recommend to the user and in which order.
  • the client device may identify, from the input, which one or more of the computed score data that the client device has received for the user is closest to the input data of the user. For instance, in the example of the user specifying relatively new games that are less popular, the client device may identify which of the scores from the computed scored data correspond to: (i) the model(s) that have been trained for this time range, and (2) the scores that have been biased to boost less popular titles.
  • the client device may use interpolation for determining scores associated with individual games and may then present the game recommendations to the user based on these scores. For example, the client device may utilize bilinear interpolation or any other form of interpolation to compute scores for the game titles and may sort these game titles according to the computed scores, with those game titles having highest scores being placed at the top of the list. Further, if the user provides different input data—such as by moving one or both sliders described above—the client device may again identify the most pertinent score data from the matrix or other representation of scores and may interpolate this score data for re-ranking the game titles.
  • the described techniques greatly improve the fluidity of the process for ranking and re-ranking content-item recommendations at the client device. That is, because the techniques involve sending multiple score data to the client device rather than a single result, and because the client device is configured to perform real-time interpolation of these score data in response to receiving different user input, the techniques enable the ranking and re-ranking of recommendations without requiring additional roundtrip requests to the remote computing system.
  • the techniques enable unlimited reranking of content-item recommendations at the cost of a single request to and response from the remote computing system, as opposed to a requirement of making a call to the system and awaiting a response from the system each time the user provides different input.
  • the techniques described herein provide for a very fine-grain level of content recommendation.
  • a desired value of a first parameter e.g., cost, release date, etc.
  • a desired value of a second parameter e.g., popularity, cost, etc.
  • FIG. 1 is a diagram illustrating an example environment that includes a client device 102, associated with a display 104, and a remote computing system 106 configured to train and use machine-learning model(s) for recommending content items to different users.
  • the remote computing system 106 trains multiple models and, thereafter, inputs a consumption history of a user into each of the models for generating score data indicating a correlation between each of multiple content-item titles and the consumption history. Further, the computing system 106 may use one or more biasing factors to generate result data, which the computing system may send to the client device 102 of a user. The client device 102 may use this result data, as well as different user inputs, to determine order(s) in which to present content-item recommendations to the user, on the display 104 or otherwise.
  • the client device 102 may be implemented as any suitable type of computing device configured to execute content items, such as video games, movies, songs, and/or the like.
  • the client device 102 may comprise, without limitation, a personal computer (PC), a desktop computer, a laptop computer, a mobile phone (e.g., a smart phone), a tablet computer, a portable digital assistant (PDA), a wearable computer (e.g., virtual reality (VR) headset, augmented reality (AR) headset, smart glasses, etc.), an in-vehicle (e.g., in-car) computer, a television (smart television), a set-top-box (STB), a game console, a music player, a voice-controlled assistant, and/or any similar computing device.
  • PC personal computer
  • a desktop computer e.g., a smart phone
  • PDA portable digital assistant
  • AR augmented reality
  • STB set-top-box
  • game console e.g., a music player, a voice-controlled
  • the client device 102 may communicate with the remote computing system 106 (sometimes shortened herein to“computing system 106”) over a computer network 108.
  • the computer network 108 may represent and/or include, without limitation, the Internet, other types of data and/or voice networks, a wired infrastructure (e.g., coaxial cable, fiber optic cable, etc.), a wireless infrastructure (e.g., radio frequencies (RF), cellular, satellite, etc.), and/or other connection technologies.
  • the computing system 106 may, in some instances be part of a network-accessible computing platform that is maintained and accessible via the computer network 108.
  • Network-accessible computing platforms such as this may be referred to using terms such as“on-demand computing”, “software as a service (SaaS)”, “platform computing”, “network-accessible platform”, “cloud services”,“data centers”, and so forth.
  • the computing system 106 acts as, or has access to, a video game platform that implements a video game service to distribute (e.g., download, stream, etc.) video games (or any other type of content item) to the client device 102.
  • the client device 102 may each install a client application thereon.
  • the installed client application may be a video-game client (e.g., gaming software to play video games).
  • a client device 102 with an installed client application may be configured to download, stream, or otherwise receive programs (e.g., video games, and content thereof) from the computing system 106 over the computer network 108.
  • Any type of content-distribution model can be utilized for this purpose, such as a direct purchase model where programs (e.g., video games) are individually purchasable for download and execution on a client device 102, a subscription-based model, a content-distribution model where programs are rented or leased for a period of time, streamed, or otherwise made available to the client devices 102.
  • an individual client device 102 may include one or more installed video games that are executable by loading the client application.
  • the client device 102 may be used to register with, and thereafter login to, a video game service.
  • a user may create a user account for this purpose and specify/set credentials (e.g., passwords, PINs, biometric IDs, etc.) tied to the registered user account.
  • credentials e.g., passwords, PINs, biometric IDs, etc.
  • the client devices 102 send data to the remote computing system 106.
  • the data sent to the remote computing system 106, for a given client machine 104 may include, without limitation, user input data, video game data (e.g., game performance statistics uploaded to the remote system), social networking messages and related activity, identifiers (IDs) of the video games played on the client device 102, and so on.
  • This data can be streamed in real- time (or substantially real-time), sent the remote system 106 at defined intervals, and/or uploaded in response to events (e.g., exiting a video game).
  • this data may be used to determine a gameplay history of the user of the client device 102, which may be used for determining scores for recommending game titles to the user.
  • FIG. 1 illustrates that the client device 102 may present, on the display 104, a UI 110 for recommending one or more content items (e.g., video games) to the user of the client device 102.
  • the UI 110 includes one or more UI controls 112(l) and 112(2) for providing two different user inputs.
  • the first UI control 112(1) may enable a user of the client device 102 to provide a desired value of a first parameter
  • the second UI control 112(2) may enable the user to provide a desired value of a second parameter.
  • the client device may then use these values for generating a custom list 114 of recommended content items 114(1), 114(2), 114(3), and 114(4).
  • the first UI control 112(1) may enable the user to specify a desired level of popularity
  • the second UI control 112(2) may enable the user to specify a desired level of release-date recency of the content items.
  • the client device 102 may update the list 114 of recommended content items in response to the user provides input via the UI control(s) and without interacting with the computing system 106.
  • FIG. 1 further illustrates that the computing system 106 may include one or more processors 116 (e.g., central processing unit(s) (CPU(s)) and computer-readable media 118.
  • the computer-readable media 118 may include volatile and nonvolatile memory, removable and non-removable media implemented in any method or technology for storage of information, such as computer-readable instructions, data structures, program modules, or other data.
  • Such memory includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, RAID storage systems, or any other medium which can be used to store the desired information and which can be accessed by a computing device.
  • the computer-readable media 118 may be implemented as computer-readable storage media (“CRSM”), which may be any available physical media accessible by the processor(s) 116 to execute instructions stored on the computer-readable media 118.
  • CRSM may include random access memory (“RAM”) and Flash memory.
  • CRSM may include, but is not limited to, read-only memory (“ROM”), electrically erasable programmable read-only memory (“EEPROM”), or any other tangible medium which can be used to store the desired information and which can be accessed by the processor(s) 116.
  • ROM read-only memory
  • EEPROM electrically erasable programmable read-only memory
  • the computer-readable media 118 may store or otherwise have access to a training component 120, a correlation component 122, and a biasing component 124.
  • the media 118 may store popularity data 126 indicating popularity (e.g., sales, gameplay time, etc.) of respective content items, content items 128 available for acquisition at the client device, and history data 130 indicating respective consumption histories of different users.
  • the training component 120 may use the history data 130 to train one or more machine-learning models 132(1), 132(2), ... , 132(N). For example, the training component 120 may train each of multiple machine learning models using a portion of the history data 130 that is associated with a sampled set of user accounts as training data to obtain trained machine-learning models 132(1)-(N).
  • the trained machine-learning model(s) 132 may represent a single model or an ensemble of base-level machine-learning models, and may be implemented as any type of machine-learning model 132.
  • suitable machine-learning models 132 for use with the techniques and systems described herein include, without limitation, neural networks, tree-based models, support vector machines (SVMs), kernel methods, random forests, splines (e.g., multivariate adaptive regression splines), hidden Markov model (HMMs), Kalman filters (or enhanced Kalman fdters), Bayesian networks (or Bayesian belief networks), expectation maximization, genetic algorithms, linear regression algorithms, nonlinear regression algorithms, logistic regression-based classification models, or an ensemble thereof.
  • An“ensemble” can comprise a collection of machine -learning models 132 whose outputs (predictions) are combined, such as by using weighted averaging or voting.
  • the individual machine-learning models of an ensemble can differ in their expertise, and the ensemble can operate as a committee of individual machine-learning models that is collectively “smarter” than any individual machine-learning model of the ensemble.
  • the training data that is used to train each machine-learning model 132 may include various types of data.
  • training data for machine learning can include two components: features and labels.
  • the training data used to train the machine-learning model(s) 132 may be unlabeled, in some embodiments.
  • the machine-learning model(s) 216 may be trainable using any suitable learning technique, such as supervised learning, unsupervised learning, semi-supervised learning, reinforcement learning, and so on.
  • the features included in the training data can be represented by a set of features, such as in the form of an n-dimensional feature vector of quantifiable information about an attribute of the training data.
  • Example features included in the training data may include, without limitation, a release date a content item, a genre of a content item, a title of a content item, a length of a content item, a rating of a content item, users that have consumed the content item, and/or the like. Further, as part of the training process, the training component 120 may set weights for machine learning.
  • weights may apply to a set of features included in the training data, as derived from the historical data 130.
  • the weights that are set during the training process may apply to parameters that are internal to the machine-learning model(s) (e.g., weights for neurons in a hidden-layer of a neural network). These internal parameters of the machine-learning model(s) may or may not map one-to-one with individual input features of the set of features.
  • the weights can indicate the influence that any given feature or parameter has on the score that is output by the trained machine learning model 132.
  • the training component 120 may train individual models by accessing individual consumption histories of users, selecting a particular content item indicated by the respective history as the output of the model, inputting the consumption history as input to the model, and training one or more internal layers of the model. For example, and as described above, the training component may access a consumption history of a first user and may use this history for training the first model 132(1). For example, if the first model 132(1) is for use in recommending content-item titles that have been released within the past six months, the training component 120 may select, from the history of the user, one game title that has been released within the past six months and may indicate to the first model 132(1) that this selected title is to be the output of the model.
  • the training component 120 may provide the entire consumption history of the user as input to the first model 132(1) and the selected content-item title as output of the first model 132(1) for training one or more internal layers of the first model 132(1).
  • the first model 132(1) comprises an artificial neural network
  • selecting the content-item title as output and providing information regarding the consumption history of the user as input may be effective to train one or more internal layers of the neural network.
  • the input and output of the neural network, and other types of models that may be used may comprise features of the corresponding game titles or other content items, such as release data, genre, game-type, and/or the like.
  • the training component 120 may take into account the amount of time that the user has consumed each content item in her history in selecting the title to be used as output for the machine-learning model 132(1). For example, the training component 120 may initially determine an amount of time that the user has consumed each individual title referenced in the consumption history of the user. This may be determined either absolutely (e.g., in terms of time) or relatively (e.g., in terms of time relative to other users that have also consumed the title). This information may be applied as a weight to the individual content-item title.
  • the training component 120 may then select one (or more) content-item title to be used as output by the first model 132(1). By doing so, the training component 120 increases the chances that the content items consumed the most by the user are selected as the output of the first model 132(1), thus training the internal layer(s) of the first model in a way that results in the model being more likely to select content items of interest to the user.
  • the training component 120 may also train multiple other models, such as a model for recommending games released within the last year 132(2), a model for recommending games released within the last three years 132(3), and so forth.
  • the training component 120 may access a consumption history of a user and select, as the output to the particular model, a title that is both indicated in the history and that was released within the amount of time associated with the particular model (e.g., within one year, three years, etc.).
  • the training component 120 may analyze thousands or more of the histories for training the models 132.
  • the correlation component 122 may function to generate correlation scores for individual users in response to receiving recommendation requests from these individual users.
  • the client device 102 may include one or more processors 134 and computer-readable media 136, which may store or otherwise having access to one or more client applications 138 and an interpolation component 140.
  • a user of the client device 102 may utilize one of the client applications 138 (e.g., a browser, a dedicated application, etc.) to interact with the computing system 106, such as to request data for rendering the UI 110.
  • the correlation component may determine an identifier associated with a user of the client device 102 for generating score data that may be used to recommend one or more content items to the user.
  • the user correlation component 122 access, from the history data 130, a consumption history of the user of the client device 102 after identifying the user.
  • the correlation component may then input this consumption history into each of the“N” trained models 132.
  • the models 132 may each output score data indicating, for each of multiple content items associated with a time range corresponding to the individual model 132, a correlation between the content item and the consumption history of the user. As described above, a higher score may indicate a greater correlation.
  • each time range associated with a particular model is exclusive, such as in the example where a first model is associated with content items released within the last six months, a second model is associated with content items released between six and twelve months prior, and so forth.
  • the time ranges are not exclusive, as in the example where a first model is associated with content released within the last six months, a second model is associated with content items released within the last year (including the prior six months), and so forth.
  • the biasing component 124 may function to bias each of the“N” score data generated by the correlation component 122 according to“M” number of values of a particular parameter.
  • This parameter may comprise a level of a popularity, a cost, a release date, or the like.
  • the parameter comprises a level of popularity, as indicated by the popularity data 126.
  • the biasing component may determine, for each content item associated with each score data output by each model, whether to boost or penalize the individual score for“M” different values of the popularity.
  • the biasing component 124 boosts or penalizes the score for each content item according to five different values of popularity, ranging from very popular to very unpopular (or, rather, very“niche”) ⁇
  • the biasing component may have generated a matrix or other representation of scores, one implementation of which is illustrated and described with reference to FIG. 3.
  • the computing system 106 may send this“result data” back to the client device 102 that initially requested the recommendations.
  • the interpolation component 140 may then determine an order in which to present recommended content items based at least in part on the received result data, as well as the values of the UI controls 112(1) and 112(2). For example, for any given set of values of these two controls, the interpolation component 140 may identify the nearest one or more score data from the result data and may interpolate scores for content items(s) based on the identified score data.
  • the interpolation component 140 may employ bilinear interpolation and/or any other method of interpolation.
  • the interpolation component 140 may perform a new interpolation and update the UI 110 to indicate the new recommendation(s).
  • FIG. 2 illustrates an example process 200 of operations for generating machine -learning models, applying them to history data of a particular user, sending generated result data to a client device of the user, and using this result data and received user input to determine order(s) in which to present content-item recommendations to the user. While this process 200 is described with reference to the environment 100 of FIG. 1, it may apply to other environments.
  • the computing system 106 may train one or more machine-learning models using any of the techniques described above.
  • the operation 202 includes training“N” number of models, each associated with a particular value of a parameter, such as a particular time window dining which a content item was released.
  • the computing system 106 may begin the training operation by accessing respective consumption histories (e.g., gameplay histories, moving histories, etc.) of multiple users associated with the computing-system community. For example, the computing system may access a gameplay history of a first user of a video-game-service community and may use this history for training the first model. For example, if the first model is for use in recommending game titles that have been released within the past six months, the computing system may select, from the gameplay history of the user, one game title that has been released within the past six months and may indicate to the first model that this selected game title is to be the output of the classifier. That is, the computing system 106 may identify a game title or other content item having the value of the parameter for which the model is being trained.
  • respective consumption histories e.g., gameplay histories, moving histories, etc.
  • the computing system 106 may provide the entire gameplay history of the user as input to the first model and the selected game title as output of the first model for training one or more internal layers of the first model.
  • the first model comprises an artificial neural network
  • selecting the game title as output and providing information regarding the gameplay history of the user as input may be effective to train one or more internal layers of the neural network.
  • the input and output of the neural network, and other types of models that may be used may comprise features of the corresponding game titles or other content items, such as release data, genre, game-type, and/or the like.
  • the computing system 106 may take into account the amount of time that the user has played each game in her history in selecting the game title to be used as output for the machine-learning model. For example, the computing system may initially determine an amount of time that the user has played each individual game title referenced in the gameplay history of the user. This may be determined either absolutely (e.g., in terms of time played) or relatively (e.g., in terms of time played relative to other users that have also played the game title). This information may be applied as a weight to the individual game title or inputted as information for the models.
  • the computing system may then select one (or more) game title to be used as output by the first model. By doing so, the computing system increases the chances that the games played the most by the user are selected as the output of the first model, thus training the internal layer(s) of the first model in a way that results in the model being more likely to select games of interest to the user.
  • the computing device may also train multiple other models each associated with a different value of the parameter, such as a model for recommending games released within the last year, a model for recommending games released within the last three years, and so forth.
  • the computing system may access a gameplay or other consumption history of a user and select, as the output to the particular model, a game title that is both indicated in the gameplay history and that was released within the amount of time associated with the particular model (e.g., within one year, three years, etc.).
  • the computing system may analyze thousands or more of the gameplay histories for training the models.
  • the client device 102 receives a request from a user for one or more content-item recommendations. This may include a user using the client device 102 to navigate to a site provided by the computing system 106, interacting with a dedicated application generated by the computing system 106, or the like.
  • the computing system 106 receives the request and, at an operation 208, generates a set of“N” correlation scores using the“N” trained models. As described above, this may include identifying a consumption history of the requesting user and inputting the consumption history into each of the“N” trained models to receives“N” sets of correlation scores.
  • the computing system 106 modifies each score within the“N” sets of correlation scores using one or more biasing factors. For example, for each biasing factor, the computing system 106 may use “M” number of values of the biasing factor to generate MxN number of sets of now-biased correlation scores. Of course, if multiple biasing factors are used, then a greater number of sets may be generated. For example, if two biasing factors are used, then the computing system 106 may generate MxMxN number of biased scores. As described above, in some instances, the biasing factor may correspond to a level of popularity of each content item. Further, it is also to be appreciated that the computing system 106 may generate the correlation scores and the biased correlation scores nonlinearly. Further, any number of values may be computed.
  • the computing system sends this“result data” to the client device 102, which receives the result data at an operation 214.
  • the client device 102 presents one or more content-item recommendations to the user based on the received result data.
  • the initial recommendations may be based on the current setting of the values of any parameters having values that may be set by the user.
  • the client device 102 receives input from the user. This may include altering a value of one or more parameters, such as a desired release data, a desired popularity level, or the like.
  • the client device 102 using the input data and the result data to interpolate scores for the content items. That is, the client device 102 generates a recommendation order using the result data and the received input.
  • the client device 102 may re-present the recommendations, albeit in the newly computed order.
  • FIG. 3 illustrates example components, and interplay thereof, for generating the models, biasing the resulting scores, applying the models to history data of a particular user, and using this result data and received user input to determine order(s) in which to present content-item recommendations to the user. It is to be appreciated that while FIG. 3 illustrates one example set of components for performing the described techniques, other components may be used in other instances.
  • some of the components reside on a server-side 302 and other component(s) reside on a client-side 104.
  • the components on the server-side 302 may reside on the computing system 106, while the component(s) on the client-side may reside on the client device 102.
  • history data 130 and one or parameters 306(1), 306(2), ... , 306( ) may be input into the training component 120 for generating the respective models 132(1)-(N). That is, the training component 120 may receive an indication of particular values of the parameter for which the training component 120 is to train a model.
  • the parameter 306(1) may indicate a value of“release date within last six months”
  • the parameter 306(2) may indicate a value of“release date within last year”, and so forth.
  • the parameters may correspond to a price of a content item, a popularity of a content item, and so forth.
  • the history data 130 may comprise consumption history of a population of users as described above.
  • the training component 120 may input individual consumption histories into a respective machine-learning model for training the model. For example, for the model 132(1) having the parameter value of“release date within last six months”, the training component may input individual consumption histories as input to the model and may designate, from each respective history, a content-item title having been released within the previous six months as output for the model 132(1). The training component 120 may then train one or more internal layers of this model 132(1) using this input and output. The training component 120 may perform this operation for an array of different consumption histories from the history data 130 and for each of the models 132(1)-(N).
  • the input to the models 32 may include information associated with the user, such as a geographical location associated with the user, demographic information associated with the user, and/or the like.
  • information associated with the user such as a geographical location associated with the user, demographic information associated with the user, and/or the like.
  • these models may be used for generating content-item recommendations for users.
  • a consumption history (stored in the history data 130) associated with the user may input into the correlation component 122, along with a library of available content items 128.
  • the correlation component may input this data to each of the models 132(1)-(N), which may output respective correlation scores 308(1), 308(2), ... , 308(N).
  • the model 132(1) may output recommendations for content-items that have been released within the last six month and that are tailored to the user based on the consumption history of the user, and so forth.
  • these scores 308(1)-( ) may input into the biasing component 124, which may generate a set of biased scores 310(1), 310(2), ... , 310( ) for each of the correlation scores 308(1)-(N). That is, the biasing component may select one or more values of a particular biasing factor and may boost or penalize the score of each content item within each of the scores 308(1)-(N) based on how well the content item corresponds to the respective value of the biasing factor.
  • the biasing factor may comprise a popularity of a content item, as indicated by the popularity data 126 which may be input into the biasing component. As described above, this popularity data 126 may comprise sales data, consumption time, or the like.
  • the biasing component select“M” number of values of the biasing factor, such as M values of popularity ranging from very popular (e.g., having a very high sales number) to very unpopular or niche (e.g., having a very low sales number).
  • the biasing component 124 may boost the scores of very popular titles for the former value of the biasing factor and penalize very niche titles, and may do the opposite for the latter value of the biasing.
  • the biasing component 124 may output a number of biased scores of“MxN” or any other number, given that the sampling may occur non-linearly.
  • the computing system 106 sends these biased scores 310(1)-(N) to the client device 102 associated with the user that initiated the request for the recommendations.
  • the sets of biased scores 310(1)-( ) may be provided as result data 312 to the client device 102.
  • This result data 312 may be input into the interpolation component 140, potentially along with input data 314 generated in response to user input received at the client device 102.
  • the input data 314 may indicate a specific value of the parameter on which the models 132(1)-( ) and/or a specific value of the biasing factor on which the scores were biased.
  • the interpolation component 140 may determine an order 316 in which to present the recommendations to the user. For example, upon receiving the input data 314, the interpolation component 140 may identify one or more nearest biased scores from the result data and may interpolate between these values if need be.
  • the interpolation component 140 may perform bilinear interpolation or any other type of interpolation for determining the order 316 in which to present the recommendations to the user. Furthermore, as different input data 314 is received (e.g., a different value of the parameter or biasing factor), the interpolation component 140 may re -interpolate the order 316 using the input data 314 and the result data 312 and may present the recommendations in the new order 316 to the user. Furthermore, and as described above, the interpolation component 140 may re-compute different orders using the result data 312 and the input data 314 without interacting with the computing system 106, thus enabling real-time resorting on the recommendations on the display of the client device 102.
  • the interpolation component 140 may perform bilinear interpolation or any other type of interpolation for determining the order 316 in which to present the recommendations to the user. Furthermore, as different input data 314 is received (e.g., a different value of the parameter or biasing factor), the interpolation component 140 may re
  • FIG. 4 illustrates example result data 312 that the remote computing system 102 may generate and send to a client device 102 of a user.
  • the result data 312 may comprise an MxN matrix of scores, which the client device may interpolate between as needed in response to receiving varying user input.
  • the result data may be sampled in any pattern in parameter space of any number of dimensions.
  • the matrix or other representation of scores may be sampled at different points other than the MxN points illustrated. For instance, the sampling may occur non-linearly .
  • the number of sampled points may vary (e.g., may be more less than MxN number of points).
  • more than two parameters and/or biasing factors may be used, resulting in higher-dimensionality parameter spaces.
  • the result data 312 corresponds to a parameter of a time range in which a content item was released and a biasing factor corresponding to a popularity of a content item.
  • the training component 120 has trained“N” number of values of this parameter, as indicated by the result data having“N” number of values of the parameter (time ranges 1-N).
  • the biasing component 124 may biased each score output by the correlation component 122 using“M” number of values of the biasing factor (popularity 1-M).
  • the result data 312 comprises an MxN number of scores.
  • the interpolation component may identify the nearest scores and may perform interpolation to computing interpolated scores, which may be used to determine an order in which to recommend content items to the user. For example, envision that a user operating the client device 102 requests to view recommendations for content items that have been released at a time between“time range 1” and“time range 2” and having a popularity that falls between“popularity 1” and“popularity 2”. In response to receiving this input data, the interpolation component 140 may identify the four scores that are nearest this input, such as“biased score 11”,“biased score 12”,“biased score 21”, and“biased score 22”.
  • the interpolation component may then use these scores to perform bilinear interpolation or any other type of suitable interpolation algorithm to generate an interpolated score.
  • This score may then be used to determine an order in which to present the content-item recommendations to the user. For example, a highest-scoring content item may be presented first, a second-highest- scoring content item may be presented second, and so forth.
  • FIG. 5 illustrates an example user interface (UI) 500 that a client device 102 may present in response to receiving a request from a user for content-item recommendations.
  • the UI 500 may present the custom list 114 of recommended content items 114(1), 114(2), 114(3), and 114(4), with this list 114 being determined based on a consumption history of the user for which the UI 500 was generated and respective values of each of the UI controls 112(1) and 112(2).
  • the UI control 112(1) may comprise a slider or other type of UI control to enable the user to select a desired level of popularity
  • the UI control 112(2) may comprise a slider or other type of UI control to enable the user to select a desired recency of the recommended content items.
  • the UI 500 presented to the user may comprise a default value for each of these UI controls 112(1), such as in the middle of each range, or the like.
  • the scores for the content items and, hence, the list 114 may be generated based on these default values, which may be changed as discussed below with reference to FIG. 6.
  • the UI 500 may include a UI control 502 in which a user may request to filter out games having a user-designated tag, as well as a UI control 504 in which a user may request to filter out games not having a user- designated tag.
  • the user may type in or select via a drop-down menu a tag in either of the controls 502 or 504.
  • the client device 102 may change the list 114 by removing any content items associated with the designated tag.
  • the client device 102 may change the list 114 by removing any content items that are not associated with the designated tag.
  • FIG. 6 illustrates an example UI 600 that the client device 102 may present in response to the user altering a value of a first parameter (in this example, popularity of the content items) and a value of a second parameter (in this example, recency of the content items).
  • a value of a first parameter in this example, popularity of the content items
  • a value of a second parameter in this example, recency of the content items.
  • the interpolation component 114 may computed a new list 602 of content items corresponding to these user inputs.
  • the list 602 may include content items that were released within a relatively large time range and that are not overly popular.
  • FIG. 7 illustrates a flow diagram of an example process 700 that the remote computing system may employ for generating one or more trained machine -learning models and generating result data for use by a client device in providing content-item recommendations to a user.
  • This process, and each process described herein, is illustrated as a collection of blocks in a logical flow graph, which represent a sequence of operations that can be implemented in hardware, software, or a combination thereof.
  • the blocks represent computer- executable instructions that, when executed by one or more processors, perform the recited operations.
  • computer-executable instructions include routines, programs, objects, components, data structures, and the like that perform particular functions or implement particular abstract data types.
  • the order in which the operations are described is not intended to be construed as a limitation, and any number of the described blocks can be combined in any order and/or in parallel to implement the processes.
  • the computing system 106 may be configured to perform some or all of the operations, although in other instances other devices may additionally or alternatively perform some or all of the operations.
  • the computing system 106 may train a machine-learning model. As illustrated, this operation may comprise a series of sub-operations. For example, training the model may include, at a sub-operation 702(1), determining a consumption history of one or more users. As described above, this could include a gameplay history, a consumption history of movies or music, and/or the like.
  • a sub-operation 702(2) may comprise selecting a title from each of the histories as output for the model that is to be trained. In some instances, this selection may be made: (1) from those content items in the history that meet the value of the parameter associated with the model currently being trained, and (2) with reference to how much each content item was consumed by the user.
  • the computing system 106 may apply a larger weight to those content items consumed by the respective user more greatly, thus increasing the likelihood that these content items will be selected as output.
  • the computing system 106 may train one or more internal layers of the model using each consumption history as input and each selected title as output. These sub-operations may be performed serially for each of multiple consumption histories.
  • the process 700 determines whether there is an additional model to be trained. For example, if the process 700 is train“N” number of models, the operation 704 represents determining whether any of the N models are still to be trained. If so, then the process 700 returns to the operation 702. If not, then the process 700 proceeds to an operation 706.
  • the computing system 106 may receive, from a client device of a user, a request for content-item recommendations.
  • the computing system 106 determines a consumption history of the user and, at an operation 710 inputs this history into a trained model.
  • the computing system 106 generates, as output from the trained model, score data indicating a correlation between the consumption history of the user and each respective content item. In some instances, a relatively higher score corresponds to a relatively higher level of correlation.
  • the computing system 106 modifies each score of the score data based on a particular value of a basing factor, as described above.
  • An operation 716 represents determining whether an additional value of a biasing factor is to be used to biased the scores. For example, if the computing system 106 is to bias the score data using“M” number of values of the biasing factor, then the operation 716 represents determining whether any of the“M” values are still to be used to bias the scores. If so, then the process 700 returns to the operation 714 for a new value of the biasing factor.
  • the process 700 proceeds to an operation 718, which represents determining whether an additional model is to be used to generate score data for the user. For example, if the computing system 106 is configured to use“N” number of models for generating score data that is tailored to the user, then the operation 718 represents determining whether any of the“N” models are still to be used to generate the score data for the user. If so, then the process 700 returns inputting the history data of the user into the next model at the operation 710. If not, then the process 700 proceeds to send the result data (e.g., the matrix or other representation of biased score data) to the client device for enabling the presentation of recommendations to the user.
  • the result data e.g., the matrix or other representation of biased score data
  • FIG. 8 illustrates a flow diagram of an example process 800 that a client device of a user may employ for determining an order in which to present one or more content items to the user. While this process 800 is described as being performed on the client device 102, it is to be appreciated that the process 800 may be additionally or alternatively be performed on other devices.
  • the client device 102 receives score data indicating a correlation between consumption history of the user of the client device each of multiple content items.
  • the score data may comprise the result data 312 discussed above.
  • the client device 102 receives an input from the user specifying at least one of a value of a parameter on which the model was trained or a value of the biasing factor on which the score data was biased.
  • the input may specify a desired recency of the content-items to be recommended (e.g., a time range associated with their respective release dates), a desire cost range of the content items, a desired popularity level of the content items, and/or the like.
  • the client device 102 determines an order in which to present the recommended content items based at least in part on the score data and the received input. For example, the client device 102 may perform interpolation on the score data based on the value of the parameter and/or biasing factor specified by the user.
  • the client device 102 presents the content-item recommendations in the determined order and, at an operation 810, determines whether an additional input is received from the user. If so, then the process 800 proceeds to again determine an order in which to (re-)present the content-item recommendations to the user based on the input. If not, then the process 800 ends at an operation 812.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Multimedia (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Medical Informatics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Business, Economics & Management (AREA)
  • Computer Security & Cryptography (AREA)
  • General Business, Economics & Management (AREA)
  • Computational Linguistics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Information Transfer Between Computers (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
EP20836901.7A 2019-07-08 2020-06-26 Empfehlungen von inhaltselementen Pending EP3972709A4 (de)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US16/505,112 US11423103B2 (en) 2019-07-08 2019-07-08 Content-item recommendations
PCT/US2020/039885 WO2021007051A1 (en) 2019-07-08 2020-06-26 Content-item recommendations

Publications (2)

Publication Number Publication Date
EP3972709A1 true EP3972709A1 (de) 2022-03-30
EP3972709A4 EP3972709A4 (de) 2023-04-19

Family

ID=74101908

Family Applications (1)

Application Number Title Priority Date Filing Date
EP20836901.7A Pending EP3972709A4 (de) 2019-07-08 2020-06-26 Empfehlungen von inhaltselementen

Country Status (6)

Country Link
US (2) US11423103B2 (de)
EP (1) EP3972709A4 (de)
JP (1) JP7443486B2 (de)
KR (1) KR20220031569A (de)
CN (1) CN114007710A (de)
WO (1) WO2021007051A1 (de)

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11423103B2 (en) 2019-07-08 2022-08-23 Valve Corporation Content-item recommendations
US11194879B2 (en) 2019-07-08 2021-12-07 Valve Corporation Custom compilation videos
US20210304285A1 (en) * 2020-03-31 2021-09-30 Verizon Patent And Licensing Inc. Systems and methods for utilizing machine learning models to generate content package recommendations for current and prospective customers
US20210313060A1 (en) * 2020-04-07 2021-10-07 Clover Health Recommendation prioritization and task throttling
US11070881B1 (en) * 2020-07-07 2021-07-20 Verizon Patent And Licensing Inc. Systems and methods for evaluating models that generate recommendations
US11693897B2 (en) * 2020-10-20 2023-07-04 Spotify Ab Using a hierarchical machine learning algorithm for providing personalized media content
US12013874B2 (en) * 2020-12-14 2024-06-18 International Business Machines Corporation Bias detection
CN113032589B (zh) * 2021-03-29 2024-01-23 北京奇艺世纪科技有限公司 多媒体文件推荐方法、装置、电子设备及可读存储介质
KR20230059595A (ko) * 2021-10-26 2023-05-03 삼성전자주식회사 사용자의 선호도를 모델링하는 방법 및 장치
JP2024008550A (ja) * 2022-07-08 2024-01-19 キヤノン株式会社 情報処理装置及びその制御方法とプログラム
US20240139629A1 (en) * 2022-10-31 2024-05-02 Samsung Electronics Co., Ltd. Effective multi-scale multi-granular targeting for game users
US20240220859A1 (en) * 2022-12-31 2024-07-04 Maplebear Inc. Counterfactual Policy Evaluation of Model Performance
CN116077942B (zh) * 2023-04-06 2023-06-27 深圳尚米网络技术有限公司 一种交互式内容推荐的实现方法

Family Cites Families (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8641424B2 (en) 2003-10-23 2014-02-04 Monvini Limited Method of publication and distribution of instructional materials
US7465231B2 (en) * 2004-05-20 2008-12-16 Gametap Llc Systems and methods for delivering content over a network
US8214264B2 (en) * 2005-05-02 2012-07-03 Cbs Interactive, Inc. System and method for an electronic product advisor
US7580853B2 (en) * 2006-04-17 2009-08-25 Electronic Entertainment Design And Research Methods of providing a marketing guidance report for a proposed electronic game
KR101895062B1 (ko) 2006-10-06 2018-09-05 로비 가이드스, 인크. 인터랙티브 미디어 안내 어플리케이션들에서 미디어를 획득, 카테고리화 및 전달하기 위한 시스템 및 방법
US8407226B1 (en) * 2007-02-16 2013-03-26 Google Inc. Collaborative filtering
EP1975813A1 (de) 2007-03-31 2008-10-01 Sony Deutschland Gmbh Verfahren zur Inhaltsempfehlung
JP2010002944A (ja) 2008-06-18 2010-01-07 Hitachi Ltd コンテンツ推薦装置とその方法
US8298087B1 (en) 2009-01-02 2012-10-30 Nintendo Of America Inc. Recommendation engine for electronic game shopping channel
US8683521B1 (en) 2009-03-31 2014-03-25 Google Inc. Feature-based video suggestions
WO2011011465A1 (en) 2009-07-21 2011-01-27 Wms Gaming, Inc. Using bots in a wagering game system
US20110054980A1 (en) * 2009-09-02 2011-03-03 Dave Simons Game revenue optimization system
US20110302103A1 (en) 2010-06-08 2011-12-08 International Business Machines Corporation Popularity prediction of user-generated content
US9104670B2 (en) * 2010-07-21 2015-08-11 Apple Inc. Customized search or acquisition of digital media assets
US9317598B2 (en) 2010-09-08 2016-04-19 Nokia Technologies Oy Method and apparatus for generating a compilation of media items
JP5095794B2 (ja) 2010-10-05 2012-12-12 ヤフー株式会社 レコメンデーション連携装置及びその方法
JP5578040B2 (ja) 2010-11-15 2014-08-27 ソニー株式会社 情報処理装置および方法、情報処理システム、並びに、プログラム
US20130080224A1 (en) * 2011-09-23 2013-03-28 Scott Thomas O'Brien Method and apparatus for an online story sharing service
US9292160B2 (en) * 2012-11-30 2016-03-22 Verizon and Redbox Digital Entertainment Services, LLC Systems and methods for presenting media program accessibility information
US9762629B1 (en) 2012-12-04 2017-09-12 Google Inc. Customizing content in a social stream
US9352226B2 (en) 2012-12-21 2016-05-31 Sony Interactive Entertainment America Llc Automatic generation of suggested mini-games for cloud-gaming based on recorded gameplay
CN103902808B (zh) 2012-12-27 2017-04-26 索尼互动娱乐美国有限责任公司 用于生成和共享云供应游戏的视频剪辑的***和方法
US20150254998A1 (en) 2014-03-05 2015-09-10 Drexel University Training systems
US9412229B2 (en) * 2014-03-21 2016-08-09 Igt System for providing a game at a gaming machine
US9560399B2 (en) 2014-06-13 2017-01-31 Hulu, LLC Personalized generation of watch list of shows in a video delivery system
US9965788B2 (en) 2014-09-26 2018-05-08 Wal-Mart Stores, Inc. System and method for prioritized product index searching
US10207185B2 (en) 2015-03-07 2019-02-19 Sony Interactive Entertainment America Llc Using connection quality history to optimize user experience
US20180130075A1 (en) 2015-03-16 2018-05-10 Amazon Technologies, Inc. Analysis of media consumption for new media production
KR101627261B1 (ko) 2015-09-08 2016-06-13 주식회사 홍인터내셔날 다트 게임 레슨 모드를 제공하는 방법, 장치 및 컴퓨터 판독 가능 매체에 저장된 컴퓨터 프로그램
CN109479110A (zh) 2016-03-08 2019-03-15 优确诺股份有限公司 动态创建个人化锻炼视频的***和方法
EP3429710A2 (de) 2016-03-15 2019-01-23 Skillz Inc. Across-match-analytik in peer-to-peer-spieleturnieren
US10432749B2 (en) 2016-08-24 2019-10-01 Facebook, Inc. Application bookmarks and recommendations
WO2018057530A1 (en) 2016-09-21 2018-03-29 GumGum, Inc. Machine learning models for identifying objects depicted in image or video data
US10762136B2 (en) * 2017-09-15 2020-09-01 Telefonaktiebolaget Lm Ericsson (Publ) Tag-based, user directed media recommendations
US11194879B2 (en) 2019-07-08 2021-12-07 Valve Corporation Custom compilation videos
US11423103B2 (en) 2019-07-08 2022-08-23 Valve Corporation Content-item recommendations

Also Published As

Publication number Publication date
US11423103B2 (en) 2022-08-23
JP2022539802A (ja) 2022-09-13
JP7443486B2 (ja) 2024-03-05
WO2021007051A1 (en) 2021-01-14
EP3972709A4 (de) 2023-04-19
KR20220031569A (ko) 2022-03-11
US20210011958A1 (en) 2021-01-14
US12026208B2 (en) 2024-07-02
CN114007710A (zh) 2022-02-01
US20230044538A1 (en) 2023-02-09

Similar Documents

Publication Publication Date Title
US12026208B2 (en) Content-item recommendations
US11194879B2 (en) Custom compilation videos
US11379514B2 (en) User-specific media playlists
CN110781321A (zh) 一种多媒体内容推荐方法及装置
US20150302009A1 (en) Adaptive Media Library for Application Ecosystems
CN115867919A (zh) 用于推荐***的图结构感知增量学习
GB2576938A (en) System and method for improved content discovery
US11771999B2 (en) Personalized game notifications
US11175807B1 (en) Intelligent contextual video thumbnail display
CN114041179B (zh) 自定义合集视频
US12015816B2 (en) User preferences based synthesis of group recommendations
CN118400582A (zh) 一种教育视频播放方法及***
CN116361537A (zh) 一种推荐方法、装置及计算机可读存储介质

Legal Events

Date Code Title Description
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20211221

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

DAV Request for validation of the european patent (deleted)
DAX Request for extension of the european patent (deleted)
A4 Supplementary search report drawn up and despatched

Effective date: 20230322

RIC1 Information provided on ipc code assigned before grant

Ipc: G06Q 99/00 20060101ALI20230316BHEP

Ipc: G06Q 30/02 20120101ALI20230316BHEP

Ipc: A63F 13/35 20140101ALI20230316BHEP

Ipc: A63F 13/00 20140101ALI20230316BHEP

Ipc: A63F 9/24 20060101ALI20230316BHEP

Ipc: A63F 13/79 20140101AFI20230316BHEP

P01 Opt-out of the competence of the unified patent court (upc) registered

Effective date: 20230428