U.S. Pat. No. 10,576,380
Artificial Intelligence (AI) Model Training Using Cloud Gaming Network
AssigneeSony Interactive Entertainment LLC
Issue DateNovember 5, 2018
Illustrative Figure
Abstract
A method for processing an artificial intelligence (AI) model for a gaming application. The method includes training the AI model from a plurality of game plays of a scenario of the gaming application using training state data collected from the plurality of game plays of the scenario and associated success criteria of each of the plurality of game plays. The method includes receiving first input state data during a first game play of the scenario. The method includes applying the first input state data to the AI model to generate an output indicating a degree of success for the scenario for the first game play. The method includes performing an analysis of the output based on a predefined objective. The method includes performing an action to achieve the predefined objective based on the output that is analyzed.
Description
DETAILED DESCRIPTION Although the following detailed description contains many specific details for the purposes of illustration, anyone of ordinary skill in the art will appreciate that many variations and alterations to the following details are within the scope of the present disclosure. Accordingly, the aspects of the present disclosure described below are set forth without any loss of generality to, and without imposing limitations upon, the claims that follow this description. Generally speaking, the various embodiments of the present disclosure describe systems and methods implementing deep learning (also referred to as machine learning) techniques to build an AI model that is related to a gaming application, and/or game plays of the gaming application. In particular, the AI model is configured to provide various functionalities in relation to a gaming application and/or to game play of the gaming application, including predicting and/or determining what actions to take in response to a given condition (e.g., game state) of the gaming application. For example, the AI model can be used to train pro-gamers to be the best of all gamers. The AI model is trained and/or built using a network of servers (e.g., through a cloud game system) executing instances of the gaming application supporting one or more game plays. The AI model may be trained based on success criteria, such as following one path over another similar path through the AI model that is more successful in terms of the success criteria. That is, the AI model learns to take the more successful path. In addition, the AI model can be continually refined through the continued collection of training data, and comparing the new training data to existing training data, with selection of the best training data based on the success criteria. With the above general understanding of the various embodiments, example ...
DETAILED DESCRIPTION
Although the following detailed description contains many specific details for the purposes of illustration, anyone of ordinary skill in the art will appreciate that many variations and alterations to the following details are within the scope of the present disclosure. Accordingly, the aspects of the present disclosure described below are set forth without any loss of generality to, and without imposing limitations upon, the claims that follow this description.
Generally speaking, the various embodiments of the present disclosure describe systems and methods implementing deep learning (also referred to as machine learning) techniques to build an AI model that is related to a gaming application, and/or game plays of the gaming application. In particular, the AI model is configured to provide various functionalities in relation to a gaming application and/or to game play of the gaming application, including predicting and/or determining what actions to take in response to a given condition (e.g., game state) of the gaming application. For example, the AI model can be used to train pro-gamers to be the best of all gamers. The AI model is trained and/or built using a network of servers (e.g., through a cloud game system) executing instances of the gaming application supporting one or more game plays. The AI model may be trained based on success criteria, such as following one path over another similar path through the AI model that is more successful in terms of the success criteria. That is, the AI model learns to take the more successful path. In addition, the AI model can be continually refined through the continued collection of training data, and comparing the new training data to existing training data, with selection of the best training data based on the success criteria.
With the above general understanding of the various embodiments, example details of the embodiments will now be described with reference to the various drawings.
FIG. 1Ais a system100A for training an artificial intelligence (AI) model160through a network of back-end servers (e.g., game cloud system) executing instances of a gaming application, in accordance with one embodiment of the present disclosure. The AI model is used for provide various functionalities in relation to the gaming application and/or game plays of the gaming application, in accordance with one embodiment of the present disclosure. In particular, the trained AI model160may be implemented and/or supported by a back-end server through a network150(e.g., internet), wherein the back-end server provides artificial intelligence and/or deep, machine learning (e.g., through the deep, machine, learning engine190) for building and applying the trained AI model160that relates to a gaming application and/or game plays of the gaming application of one or more players in one or more locations throughout the world. The trained AI model160is trained to learn the intricacies of the gaming application, and/or intricacies of playing the gaming application (e.g., through a corresponding game play). In that manner, given as input a condition of a game play of the gaming application (e.g., game state of a particular game play), the AI model can analyze the condition and provide an appropriate response to the condition. The response may be dependent on a predefined objective (e.g., provide assistance, etc.). For example, the trained AI model160is able to determine how the gaming application should respond, or how the player should respond during a corresponding game play.
As shown, a plurality of client devices230(e.g., devices230athrough230n) is connected through a network150(e.g., internet) to a back-end game cloud system (GCS)201to support a plurality of game plays of a particular gaming application. For example, a plurality of remote players is playing the gaming application through respective client devices230. The client device can be any type of computing device having at least a memory and a processor module that is capable of connecting to the back-end server system (e.g., GCS201) over network150. A corresponding client device is configured to interact with an instance of a corresponding gaming application as executed locally or remotely to implement a game play of a corresponding player, such as through input commands that are used to drive game play.
The GCS201includes a plurality of servers205(e.g., servers205athough205n) executing a plurality of instances of the gaming application110(e.g., instances110athrough110n). For example, server205ais configured to execute instance110aof the gaming application in support of a corresponding game play of a corresponding player through a corresponding client device. Each of the servers205may be configured as having at least a memory and a processor module that is capable of executing the gaming application, such as through a corresponding instance of the gaming application, in support of a corresponding game play. For example, each server205may be a server console, gaming console, computer, etc. Each server205is configured to stream data105(e.g., rendered images and/or frames of a corresponding game play) back to a corresponding client device through network150. In that manner, a computationally complex gaming application may be executing at the back-end server in response to controller inputs received and forwarded by a corresponding client device. Each server is able to render images and/or frames that are then encoded (e.g., compressed) and streamed to the corresponding client device for display.
In one embodiment, GCS201includes a distributed game engine system and/or architecture. In particular, a distributed game engine executing game logic is configured as a corresponding instance of the gaming application. In general, the distributed game engine takes each of the functions of a game engine and distributes those functions for execution by a multitude of processing entities. Individual functions can be further distributed across one or more processing entities. The processing entities may be configured in different configurations, including as metal or physical hardware, and/or as virtual components or virtual machines, and/or as virtual containers, wherein a container is different from a virtual machine as it virtualizes an instance of the gaming application running on a virtualized operating system. The processing entities may utilize and/or rely on servers and their underlying hardware on one or more servers (compute nodes) of the GCS201, wherein the servers may be located on one or more racks. The coordination, assignment, and management of the execution of those functions to the various processing entities are performed by a distribution synchronization layer. In that manner, execution of those functions is controlled by the distribution synchronization layer to enable generation of media (e.g., video frames, audio, etc.) for the gaming application in response to controller input by a player. The distribution synchronization layer is able to efficiently execute (e.g., through load balancing) those functions across the distributed processing entities, such that critical game engine components/functions are distributed and reassembled for more efficient processing.
These various functions performed by a game engine include basic processor based functions for executing the gaming application and services associated with the gaming application. For example, processor based functions include 2D or 3D rendering, physics, physics simulation, scripting, audio, animation, graphics processing, lighting, shading, rasterization, ray tracing, shadowing, culling, transformation, artificial intelligence, etc. In addition, services for the gaming application include streaming, encoding, memory management, multi-thread management, quality of service (QoS), bandwidth testing, social networking, management of social friends, communication with social networks of friends, communication channels, texting, instant messaging, chat support, etc.
In addition, the distribution synchronization layer can easily be scaled (on an individual distributed game engine) to handle increased processing demand or complexity of processing by adding more processing entities. That is, the processing power can be elastically increased or decreased based on demand. Also, the distributed game engine system can be scaled across many users, each of which is supported by a corresponding distributed game engine, such as in a multi-player gaming session for the gaming application. As such, the multi-player gaming session is not constrained by hardware limitations that limit the maximum number of players in the session because of performance issues (e.g., latency). Instead, the scalable distributed gaming engine system is able to scale up or down the number of distributed gaming engines depending on the number of players participating in the multi-player gaming session without any hardware constraints. As such, many thousands of players may participate in a single multi-player gaming session.
In system100A, an AI model160can be built (e.g., trained) using the network of servers that are configured to execute instances of the gaming application in support of the multiple game plays. In particular, training data345is collected from the execution of the instances of the gaming application at the plurality of servers205. In particular, the collection for use as training data is performed without the knowledge of any of the remote players that are associated with the various game plays of the gaming application. The training data345includes controller input236used to direct corresponding game plays, game states133of the instances of the gaming application occurring during the game plays, response metrics, success criteria135, success metrics, additional analysis performed on the game plays, etc. In some cases, the training data345can be generated by the instances of the gaming application (e.g., game state metadata, etc.). The training data345is provided as input into deep learning and/or machine learning engine190. The deep learning engine190includes a modeler120that is configured to build and/or train the AI model160using the training data345, as will be further described below in relation toFIGS. 1B, and 3A, 3B-1 and 3B-2.
The AI model160may be implemented during subsequent game play (e.g., after training of the AI model) of the gaming application (e.g., by a player, by an automatic player, etc.). For example, the AI model may be implemented and/or executed at a back-end server in support of the game play, wherein the game play may be executing on a local device (to the player) or at a back-end server. For example, the AI model that is trained is implemented by the deep learning engine190to provide various functionalities to game plays of the gaming application.
As shown, analyzer140is configured to utilize the AI model160that is trained to provide various functionalities in relation to a game play of the gaming application. In particular, an input data stream405is provided as input to the deep learning engine190that is configured to implement the trained AI model160. The trained AI model160provides an output in response to the input, wherein the output is dependent on the predefined functionality and/or predefined objective of the trained AI model160. For example, the trained AI model160may be used by the analyzer140to determine what actions need to be taken during the game play—either by the player, or by the corresponding executing instance of the gaming application. The analyzer140includes an action generator170that is configured to perform an action responsive to the input state data405and in consideration of the predefined objective of the trained AI model160. In that manner, the analyzer through the use of the AI model160can provide various functionalities, including providing services to the player playing the gaming application (e.g., providing recommendations, finding weaknesses of the player, training the player, providing an opponent to the player, finding flaws in the gaming application, etc.).
More particularly, the analyzer140is configured to perform various functionalities in relation to a gaming application and/or game plays of the gaming application. The analyzer140is configured to analyze the output from the trained AI model160for a given input (e.g., controller input236, game state data133, success criteria135), and provide a response (e.g., an action). For example, the analyzer140may provide a profile of a player playing the gaming application; provide a recommendation to a player during game play of a gaming application by the player, wherein the recommendation may be structured in consideration of the user profile; take over the game play; build teams (e.g., teams that are competing against each other in the gaming application) that are fairly balanced using player profiles; auto play the gaming application, such as for purposes of automatically training the AI model; provide a bot opponent; explore the gaming application; determine a weakness of a corresponding player, and to perform services that help the player to overcome the weakness, etc.
FIG. 1Billustrates an example neural network190(e.g., deep learning engine) used to build a trained AI model that learns, in part, the intricacies of a gaming application and how to play the corresponding gaming application. In that manner, given as input a condition of a game play of the gaming application (e.g., game state of a particular game play) the AI model can analyze the condition and provide an appropriate response to the condition. For example, the AI model can be used to provide various functionalities in relation to a gaming application and/or to game play of the gaming application, including predicting and/or determining what actions to take in response to a given condition (e.g., game state) of the gaming application.
More particularly, the AI model is trained and/or built using a network of servers (e.g., through a cloud game system) executing instances of the gaming application supporting one or more game plays. The training is implemented by the deep learning engine190, in accordance with one embodiment of the present disclosure. The neural network190may be implemented within an AI processor or engine210at a back-end server, in one embodiment. In particular, the modeler120of the deep learning engine190in system100A ofFIG. 1Ais configured to learn everything about the gaming application for use in subsequent game plays of the gaming application by any player (e.g., real or virtual).
Specifically, the deep learning or machine learning engine190in cooperation with modeler120is configured to analyze training data345that is collected at a plurality of back-end servers configured for executing a plurality of instances of a gaming application. The deep learning engine190utilizes artificial intelligence, including deep learning algorithms, reinforcement learning, or other artificial intelligence-based algorithms to build a trained AI model that relates to the gaming application and/or game plays of the gaming application. In that manner, a vast collection of training data may be efficiently collected, in part, to help define the gaming application, and/or any game play of the gaming application, to include the contexts within which the gaming application may present (e.g., different scenarios, and parts within those scenarios, as defined within the trained AI model. Further, the AI model can be used (e.g., during training or after training, such as by the analyzer140of the AI processor engine210) to provide various functionalities in relation to the gaming application and/or game plays of the gaming application in any subsequent game play of the gaming application. As such, the deep learning engine190is able to learn everything about the gaming application and/or game plays of the gaming application, so that the trained AI model can be used to provide the best services for subsequent game plays of the gaming application, such as to a player. For example, the trained AI model may be used to train a pro-gamer to be the best gamer in the world, or to provide various tutorial sessions (e.g., video instruction, game play challenges, etc.) that are designed to address the weaknesses of the player. The deep learning engine190may be configured to continually refine the trained AI model given any updated training data. The refinements are based on determining which sets of training data can be used for training based on how those sets perform within the gaming application based on corresponding success criteria.
More particularly, during the learning and/or modeling phase, the training data345is used by the deep learning engine190to predict and/or determine how successful a particular game play of a gaming application will turn out given a set of input data. The resulting AI model of the gaming application can be used to determine actions to be performed, for a given game play of a gaming application given the set of input data. That is, the trained AI model can be used (e.g., by analyzer140) to provide various functionalities related to a gaming application and/or game plays of the gaming application. For example, the input data may be game state data (e.g., including controller input, etc.), and the trained AI model160may be used to generate a response to the input data. The response may be provided with or without any request or knowledge by the user.
For example, the trained AI model160may be used by the analyzer140to provide recommendations to a player playing the gaming application. The AI model160may be used by the analyzer140to create a user profile of the player that is specific to the gaming application, or to a scenario of the gaming application. The AI model160may be used by the analyzer140to control a bot opponent within the gaming application, such as when training the player against the best opponent in the world as created through training data, when training the player against a virtual me opponent so that the user is able to get incrementally better through self-play (e.g., playing against a virtual version of himself or herself) (and that continually gets refined through that self-play). The AI model160may be used by the analyzer140to discover weaknesses of the player and to provide tutorial sessions to address those weaknesses. The AI model160may be used by the analyzer140to discover problems within the gaming application (e.g., holes in the code that lead to glitches, etc.). Still other functionalities are supported, though not necessarily described.
The neural network190represents an example of an automated analysis tool for analyzing data sets to determine the intricacies of playing a gaming application, to include the responses and/or actions that can be determined and/or performed during a game play of a gaming application. Different types of neural networks190are possible. In an example, the neural network190supports deep learning that may be implemented by deep learning engine190. Accordingly, a deep neural network, a convolutional deep neural network, and/or a recurrent neural network using supervised or unsupervised training can be implemented. In another example, the neural network190includes a deep learning network that supports reinforcement learning, or rewards based learning (e.g., through the use of success criteria, success metrics, etc.). For instance, the neural network190is set up as a Markov decision process (MDP) that supports a reinforcement learning algorithm.
Generally, the neural network190represents a network of interconnected nodes, such as an artificial neural network. Each node learns some information from data. Knowledge can be exchanged between the nodes through the interconnections. Input to the neural network190activates a set of nodes. In turn, this set of nodes activates other nodes, thereby propagating knowledge about the input. This activation process is repeated across other nodes until an output is provided.
As illustrated, the neural network190includes a hierarchy of nodes. At the lowest hierarchy level, an input layer191exists. The input layer191includes a set of input nodes. For example, each of these input nodes is mapped to instances of a game play of the gaming application, wherein the instances include one or more features that define that instance (e.g., controller input, game state, results data, etc.). The intermediary predictions of the model are determined through a classifier that creates labels (e.g., outputs, features, nodes, classifications, etc.).
At the highest hierarchical level, an output layer193exists. The output layer193includes a set of output nodes. An output node represents a decision (e.g., action, prediction, prediction of success of a game play for a given set of input data, etc.) that relates to one or more components of a trained AI model160, for example. As previously described, the output nodes may identify the predicted or expected actions, or learned actions for a given set of inputs, wherein the inputs may define various scenarios or parts of scenarios of a gaming application. These results can be compared to predetermined and true results, or learned actions and results, as obtained from the game plays used for collecting the training data in order to refine and/or modify the parameters used by the deep learning engine190to iteratively determine the appropriate predicted or expected responses and/or actions for a given set of inputs. That is, the nodes in the neural network190learn the parameters of the trained AI model160that can be used to make such decisions when refining the parameters.
In particular, a hidden layer192exists between the input layer191and the output layer193. The hidden layer192includes “N” number of hidden layers, where “N” is an integer greater than or equal to one. In turn, each of the hidden layers also includes a set of hidden nodes. The input nodes are interconnected to the hidden nodes. Likewise, the hidden nodes are interconnected to the output nodes, such that the input nodes are not directly interconnected to the output nodes. If multiple hidden layers exist, the input nodes are interconnected to the hidden nodes of the lowest hidden layer. In turn, these hidden nodes are interconnected to the hidden nodes of the next hidden layer, and so on and so forth. The hidden nodes of the next highest hidden layer are interconnected to the output nodes. An interconnection connects two nodes. The interconnection has a numerical weight that can be learned, rendering the neural network190adaptive to inputs and capable of learning.
Generally, the hidden layer192allows knowledge about the input nodes to be shared among all the tasks corresponding to the output nodes. To do so, a transformation ƒ is applied to the input nodes through the hidden layer192, in one implementation. In an example, the transformation ƒ is non-linear. Different non-linear transformations ƒ are available including, for instance, a rectifier function ƒ(x)=max(0,x).
The neural network190also uses a cost function c to find an optimal solution. The cost function measures the deviation between the prediction that is output by the neural network190defined as ƒ(x), for a given input x and the ground truth or target value y (e.g., the expected result). The optimal solution represents a situation where no solution has a cost lower than the cost of the optimal solution. An example of a cost function is the mean squared error between the prediction and the ground truth, for data where such ground truth labels are available. During the learning process, the neural network190can use back-propagation algorithms to employ different optimization methods to learn model parameters (e.g., the weights for the interconnections between nodes in the hidden layers192) that minimize the cost function. An example of such an optimization method is stochastic gradient descent.
In an example, the training dataset for the neural network190can be from a same data domain. For instance, the neural network190is trained for learning the predicted or expected responses and/or actions to be performed for a given set of inputs or input data. In this illustration, the data domain includes game play data collected through multiple game plays of multiple users to define baseline input data. In another example, the training dataset is from different data domains to include input data other than baseline.
As such, the neural network190may predict or determine, in part, predicted or expected responses and/or actions to be performed for a given set of inputs (a condition of a gaming application, such as game state). Based on these predictive results, the neural network190may also define the trained AI model160that is used to provide determine those results and/or actions (e.g., the various functionalities relating to the gaming application and/or game plays of the gaming application) to be performed given a set of inputs.
FIG. 2Aillustrates a system200A supporting the collection of training data used to build and/or train an AI model relating to a gaming application and/or game plays of the gaming application, in accordance with one embodiment of the present disclosure. In particular, system diagram200A enables access and playing of video games stored in a game cloud system (GCS)201. Generally speaking, game cloud system GCS201may be a cloud computing system operating over a network220to support a plurality of players playing a gaming application through corresponding game plays. Data related to those game plays may be provided as training data used for building and/or training an AI model relating to the gaming application and/or game plays of the gaming application. In particular, system200A includes GCS201, one or more social media providers240, and a client device230, all of which are connected via a network150(e.g., internet). One or more user devices may be connected to network150to access services provided by GCS201and social media providers240.
In one embodiment, game cloud system201includes a game server205, a video recorder271, a tag processor273, and account manager274that includes a user profile manager, a game selection engine275, a game session manager285, user access logic280, a network interface290, and a social media manager295. GCS201may further include a plurality of gaming storage systems, such as a game state store, random seed store, user saved data store, snapshot store, which may be stored generally in datastore260. Other gaming storage systems may include a game code store261, a recorded game store262, a tag data store263, video game data store264, and a game network user store265. In one embodiment, GCS201is a system that can provide gaming applications, services, gaming related digital content, and interconnectivity among systems, applications, users, and social networks. GCS201may communicate with user device230and social media providers240through social media manager295via network interface290. Social media manager295may be configured to relate one or more friends. In one embodiment, each social media provider240includes at least one social graph245that shows user social network connections.
Player/user5is able to access services provided by GCS201via the game session manager285. For example, account manager274enables authentication and access by player5to GCS201. Account manager274stores information about member users/players. For instance, a user profile for each member user may be managed by account manager274. In that manner, member information can be used by the account manager274for authentication purposes. For example, account manager2274may be used to update and manage user information related to a member user. Additionally, game titles owned by a member user may be managed by account manager274. In that manner, video games stored in data store264are made available to any member user who owns those video games.
In one embodiment, a user, e.g., player5, can access the services provided by GCS201and social media providers240by way of client device230through connections over network150. Client device230can include any type of device having a processor and memory, wired or wireless, portable or not portable. In one embodiment, client device230can be in the form of a smartphone, a tablet computer, or hybrids that provide touch screen capability in a portable form factor. One exemplary device can include a portable phone device that runs an operating system and is provided with access to various applications (apps) that may be obtained over network150, and executed on the local portable device (e.g., smartphone, tablet, laptop, desktop, etc.).
Client device230includes a display232that acts as an interface for player5to send input commands236and display data and/or information235received from GCS201and social media providers240. Display232can be configured as a touch-screen, or a display typically provided by a flat-panel display, a cathode ray tube (CRT), or other device capable of rendering a display. Alternatively, the client device230can have its display232separate from the device, similar to a desktop computer or a laptop computer.
In one embodiment, client device230is configured to communicate with GCS201to enable player5to play a video game. For example, player5may select (e.g., by game title, etc.) a video game that is available in the video game data store264via the game selection engine275. In that manner, the selected video game is enabled and loaded for execution by game server205on the GCS201. In one embodiment, game play is primarily executed in the GCS201, such that client device230will receive a stream of game video frames235from GCS201, and user input commands236for driving the game play is transmitted back to the GCS201. The received video frames235from the streaming game play are shown in display232of client device230.
In one embodiment, after player5chooses an available game title to play, a game session for the chosen game title may be initiated by the player5through game session manager285. Game session manager285first accesses game state store in data store140to retrieve the saved game state of the last session played by the player5(for the selected game), if any, so that the player5can restart game play from a previous game play stop point. Once the resume or start point is identified, the game session manager285may inform game execution engine in game processor210to execute the game code of the chosen game title from game code store261. After a game session is initiated, game session manager285may pass the game video frames235(i.e., streaming video data), via network interface290to a client device, e.g., client device230.
During game play, game session manager285may communicate with game processor210, recording engine271, and tag processor273to generate or save a recording (e.g., video) of the game play or game play session. In one embodiment, the video recording of the game play can include tag content entered or provided during game play, and other game related metadata. Tag content may also be saved via snapshots. The video recording of game play, along with any game metrics corresponding to that game play, may be saved in recorded game store262. Any tag content may be saved in tag data store263.
During game play, game session manager285may communicate with game processor204to deliver and obtain user input commands236that are used to influence the outcome of a corresponding game play of a video game. Input commands236entered by player5may be transmitted from client device230to game session manager285of GCS201. Input commands (e.g., controller inputs)236, including input commands used to drive game play, may include user interactive input, such as including tag content (e.g., texts, images, video recording clips, etc.). Game input commands as well as any user play metrics (how long the user plays the game, etc.) may be stored in game network user store. Select information related to game play for a video game may be used to enable multiple features that may be available to the user.
FIG. 2Billustrates a system200B providing the collection of training data used to build and/or train an AI model160relating to a gaming application and/or game plays of the gaming application, wherein instances and/or instantiations of the gaming application in support of the game plays are executed over a cloud game network, in accordance with one embodiment of the present disclosure. In addition, the system200A is configured to support implementation of the AI model160that is trained to provide various functionalities relating to the gaming application and/or game plays of the gaming application (e.g., provide recommendations, train the player, discover weaknesses of the player, provide bot opponents, etc.).
As shown, system200B provides gaming control to a plurality of players215(e.g., players5L,5M . . .5Z) playing a gaming application as executed over a cloud game network, in accordance with one embodiment of the present disclosure. In some embodiments, the cloud game network may be a game cloud system210that includes a plurality of virtual machines (VMs) running on a hypervisor of a host machine, with one or more virtual machines configured to execute a game processor module utilizing the hardware resources available to the hypervisor of the host. In other embodiments, GCS201includes a distributed game engine system and/or architecture that is executing game logic, as configured as a corresponding instance of the gaming application. In general, the distributed game engine takes each of the functions of a game engine and distributes those functions for execution by a multitude of processing entities across one or more servers of the GCS201. Individual functions can be further distributed across one or more processing entities. Referring now to the drawings, like referenced numerals designate identical or corresponding parts.
As shown, the game cloud system210includes a game server205that provides access to a plurality of interactive video games or gaming applications. Game server205may be any type of server computing device available in the cloud, and may be configured as one or more virtual machines executing on one or more hosts. For example, game server205may manage a virtual machine supporting a game processor that instantiates an instance of a gaming application for a user. As such, a plurality of game processors of game server205associated with a plurality of virtual machines is configured to execute multiple instances of the gaming application associated with game plays of the plurality of users215. In that manner, back-end server support provides streaming of media (e.g., video, audio, etc.) of game plays of a plurality of gaming applications to a plurality of corresponding users. As will be further described below, training data collected from the plurality of game processors executing instances of the gaming application is used to build and/or train an AI model relating to the gaming application and/or game plays of the gaming application.
A plurality of players215accesses the game cloud system210via network150, wherein players (e.g., players5L,5M . . .5Z) access network150via corresponding client devices230′. Client device230′ may be configured similarly as client device230ofFIG. 1A, or may be configured as a thin client providing that interfaces with a back end server providing computational functionality. In particular, a client device230′ of a corresponding player5L is configured for requesting access to gaming applications over a network150, such as the internet, and for rendering instances of gaming application (e.g., video game) executed by the game server205and delivered to a display device associated with the corresponding player5L. For example, player5L may be interacting through client device230′ with an instance of a gaming application executing on game processor of game server205. More particularly, an instance of the gaming application is executed by the game title execution engine211(e.g., game engine) that is running game logic177corresponding to the gaming application. Game logic (e.g., executable code)177implementing the gaming application is stored and accessible through data store260, or game code store261, or video game store264, etc., previously described, and is used to execute the gaming application. Game title processing engine211is able to support a plurality of gaming applications using a plurality of game logics177, as shown.
As previously described, client device230′ may receive input from various types of input devices11, such as game controllers, tablet computers, keyboards, gestures captured by video cameras, mice touch pads, etc. Client device230′ can be any type of computing device having at least a memory and a processor module that is capable of connecting to the game server205over network150. Also, client device230′ of a corresponding player is configured for generating rendered images executed by the game title execution engine211executing remotely, and for displaying the rendered images on a display, including a head mounted display (HMD)102. For example, a corresponding client device230′ is configured to interact with an instance of a corresponding gaming application as executed remotely to implement a game play of a corresponding player, such as through input commands that are used to drive game play.
In another embodiment, multi-player processing engine119, previously described, provides for controlling a multi-player gaming session for a gaming application. In particular, when the multi-player processing engine119is managing the multi-player gaming session, the multi-player session controller116is configured to establish and maintain communication sessions with each of the users and/or players in the multi-player session. In that manner, players in the session can communicate with each other as controlled by the multi-player session controller116.
Further, multi-player processing engine119communicates with multi-player logic118in order to enable interaction between players within corresponding gaming environments of each player. In particular, state sharing module117is configured to manage states for each of the players in the multi-player gaming session. For example, state data may include game state data that defines the state of the game play (of a gaming application) for a corresponding player (e.g., player5L) at a particular point, as previously described. Further, state data may include user/player saved data that includes information that personalizes the video game for the corresponding player, as previously described. For example, state data includes information associated with the user's character, so that the video game is rendered with a character that may be unique to that user (e.g., shape, look, clothing, weaponry, etc.). In that manner, the multi-player processing engine119using the state sharing data117and multi-player logic118is able to overlay/insert objects and characters into each of the gaming environments of the users participating in the multi-player gaming session. This allows for interaction between users in the multi-player gaming session via each of their respective gaming environments (e.g., as displayed on a screen).
In addition, back-end server support via the AI processor210, which may be integrated within GCS201or located remote from GCS201, may provide for the building and/or training of the AI model160relating to a gaming application and/or game plays of the gaming application, and also provide for the implementation and/or application of the AI model160. In particular, the back-end AI processor210includes the deep learning engine190, previously described, that is configured for learning and/or modeling, in part, the responses and/or actions (e.g., controller inputs, etc.) to be performed for any given set of inputs (e.g., that define a condition of a game play of a gaming application, including game state, etc.) in order to build (e.g., through the modeler120) and apply (e.g., through analyzer140) the trained AI model in relation to a gaming application and/or subsequent game play of the gaming application. For example, the modeler120within the deep learning engine190may operate to set the parameters defined within the deep learning engine190that define the various nodes in the input layer191, hidden layers192, and output layer193, for purposes applying the trained AI model160within the deep learning engine190. The modeler120may set the parameters in the AI model160based on one or more success criteria used during training, as previously described. In that manner, the AI model160is trained to learn the intricacies of the gaming application and/or the intricacies of playing the gaming application so that the AI model160can be used to provide various functionalities in relation to the gaming application and/or a game play of the gaming application (e.g., predicting and/or determining what actions, to include controller inputs, to take in response to a given condition, such as game state). As such, the analyzer140is able to analyze the output from the AI model160responding to input defining a condition of a game play, and optionally the input (e.g., input state data), in order to provide an appropriate response to the condition, wherein the response may be dependent on a predefined objective (e.g., provide assistance, providing coaching, etc.).
With the detailed description of the various modules of the system100A and system200B, flow diagram300A discloses a method for training an AI model through a network of back-end servers executing instances of the gaming application, the AI model being used for provide various functionalities in relation to the gaming application and/or game plays of the gaming application, in accordance with one embodiment of the present disclosure. Flow diagram300A may be implemented within a back-end server (e.g., within the game cloud system201in combination with a deep learning engine190), previously described.
In particular, at302the method includes executing at a plurality of servers a plurality of instances of a gaming application, the plurality of instances supporting a plurality of game plays of the gaming application. For example, the plurality of servers may be operating within a game cloud system (e.g., GCS201) previously described, wherein the gaming application is executing remotely from a corresponding player, and media (e.g., video frames, audio, etc.) is streamed to the client device of the player over a network. In that case, the plurality of game plays of the gaming application is controlled by a plurality of players via corresponding client devices, wherein the client devices are remote from the servers. In other implementations, the gaming application is executing locally to a user/player and metadata from the executing gaming application is delivered over a network to the back-end servers for purposes of analysis (e.g., for purposes of training the AI model), or for supporting multi-player gaming sessions. In still other embodiments, the plurality of game plays may be controlled automatically (e.g., through AI) (e.g., for purposes of self-training an AI model).
At304, the method includes collecting at the plurality of servers training state data associated with corresponding game plays of a scenario of the gaming application. The training state data may include metadata associated with the game plays, to include controller inputs, game state, progress through the game play, results (e.g., success or failure) of the scenario, user profile information, etc. That is, the training state data includes any data that may be relevant to understanding the gaming application and/or game plays of the gaming application. Because the instances of the gaming application are executing at the back-end servers, access to the training state data is readily available without active participation of the associated players of those game plays. That is, the players may not even know that the training state data is being collected.
The training state data may include game state data that defines the state and/or condition of a gaming application at a particular point (e.g., during a game play). For example, game state data may include game characters, game objects, game object attributes, game attributes, game object state, graphic overlays, etc. In that manner, game state data allows for the generation of the gaming environment that existed at the corresponding point in the video game. Game state data may also include the state of every device used for rendering the game play, such as states of CPU, GPU, memory, register values, program counter value, programmable DMA state, buffered data for the DMA, audio chip state, CD-ROM state, etc. Game state data may also identify which parts of the executable code need to be loaded to execute the video game from that point. Not all the game state data need be captured and stored, just the data that is sufficient for the executable code to start the game at the point corresponding to the snapshot.
Also, the training state data may include user saved data that personalizes the gaming application for the corresponding user. This includes information associated with the user's character, so that the video game is rendered with a character that may be unique to that user (e.g., shape, look, clothing, weaponry, etc.). In that manner, the user saved data enables generation of a character for the game play of a corresponding user, wherein the character has a state that corresponds to the point in the video game associated with the snapshot. For example, user saved data may include the game difficulty selected by the corresponding player when playing the gaming application, game level, character attributes, character location, number of lives left, the total possible number of lives available, armor, trophy, time counter values, etc. User saved data may also include user profile data that identifies the corresponding player.
In addition, training state data may also include random seed data that is generated through AI. The random seed data may not be part of the original game code, but may be added in an overlay to make the gaming environment seem more realistic and/or engaging to the user. That is, random seed data provides additional features for the gaming environment that exists at the corresponding point in the game play of the player. For example, AI characters may be randomly generated and provided in the overlay. The AI characters are placed into the gaming environment to enhance the user's experience, and may or may not affect the game play. As an illustration, these AI characters may randomly walk the streets in a city scene. In addition, other objects maybe generated and presented in an overlay. For instance, clouds in the background and birds flying through space may be generated and presented in an overlay.
The gaming application may include one or more scenarios. A scenario may be a critical point in the gaming application (e.g., necessary to progress through the gaming application), such as combating a boss at the end of a level, or jumping over an object or obstacle that is blocking the only path to a destination, or passing an obstacle (e.g., climbing a mountain, crossing a lake or river with aggressive alligators, etc.). A scenario may be less critical, such as when completing an intermediary task in the gaming application. In these cases, the scenario may involve completing the task to obtain a reward (e.g., money, important sword, etc.). The collection of data for purposes of training an AI model may be limited to data that is related to game play during one or more scenarios of interest. In that manner, the trained AI model understands the intricacies of playing the gaming application during the scenario using data that is relevant to that scenario, and without contamination from data that may not be relevant to playing that scenario.
The scenario may be predefined, such as by the developer of the gaming application. For example, the scenario may be designed to have a high degree of difficulty, such that it is expected that many players would fail to progress through scenario. In other cases, the scenario may be discovered through analysis of the collected training state data. That is, it may become clear that a particular portion of the gaming application is difficult for players to progress through during their corresponding game plays. In that case, that portion of the gaming application may be identified as a scenario, such that training state data is collected during game plays of that identified scenario.
At306, the method includes defining success criteria for the training state data. For example, the success criteria may be used by the deep learning engine for purposes of training the AI model. Specifically, the success criteria may be applied to define the interrelationships between the nodes of layers when the AI model (e.g., refine the weights defining the interrelationships between two nodes of different layers, etc.). For example, the success criteria may be used to differentiate training state data that are similar and that gives some insight in how to play or choose how to play the gaming application during the scenario. In a simplistic example, the scenario may involve accomplishing a task (e.g., beat a boss, navigate through an obstacle, etc.), wherein two sets of training data each describe or outline processes in how to play through the scenario (e.g., controller inputs, strategy, etc.). The success criteria may be utilized to determine which set of training state data is more successful in accomplishing the task. The more successful set of training state data may be more heavily weighted (e.g., when defining interrelationships between nodes of the AI model) than the other set of training state data that is less successful, for purposes of training an AI model relating to the scenario of the gaming application and/or game play of the scenario of the gaming application.
At308, the method includes training an AI model for the scenario by providing the training state data and the success criteria to a deep learning engine, wherein the trained AI model provides a plurality of outputs for a plurality of inputs. Because the training state data relates to the scenario, the AI model is trained to learn everything about the scenario in the gaming application and/or game play of the scenario based on one or more success criteria. In that manner, when the AI model is given a set of inputs (e.g., game state, controller input, etc.) related to a subsequent game play of the scenario, the AI model can provide an output that may be beneficial to that game play. In one embodiment, the output may indicate a degree of success for that game play. That is, given the current condition of the game play (e.g., game state), the AI model can predict where that game play is going and predict how successful that game play may be in progressing through the corresponding scenario. Additional analysis of the output (e.g., by the analyzer140), as well as analysis of the set of inputs (e.g., current and past sets of inputs), may provide a response to the set of inputs. In particular, an action may be performed as the output, and in response to the set of inputs. For example, if the set of inputs indicates that the game play is leading towards being unsuccessful in progressing through the scenario (e.g., via the output of the AI model), the output from the AI model may be used (e.g., by the analyzer140) to provide a recommendation or advice on how to progress through the encountered condition during game play of the scenario of the gaming application. If the game play is associated with a remote player, the recommendation may be provided without any prompting from the remote player, such that the set of inputs is provided to the trained AI model automatically during the game play. The recommendation may help the player to successfully progress through the scenario, efficiently progress through the scenario, earn the most assets playing the scenario, accomplish tasks in the scenario that give the player the best chance to progress through a later stage or scenario in the gaming application, etc. In other embodiments, the recommendation is provided at the request of the remote player, such that in response to the request, the set of inputs is provided to the AI model.
FIGS. 3B-1 and 3B-2illustrate data flow diagrams illustrating processes for training an AI model through a network of back-end servers executing instances of a gaming application, in accordance with one embodiment of the present disclosure. As previously described, the AI model is knowledgeable about the gaming application and/or game plays of the gaming application, such that given an input (e.g., that relates to a condition of the gaming application—game state, etc.) the AI model can be used to provide various functionalities (e.g., provide assistance, etc.) in relation to the gaming application and/or game plays of the gaming application. The process ofFIG. 3B-2provides a different functionality as performed by modeler120′ when compared to the functionality as performed by modeler120ofFIG. 3B-1, as will be described below.
As shown inFIGS. 3B-1 and 3B-2, a plurality of game plays310(e.g.,310athrough310n) of a gaming application is shown. Instances of the gaming application may be executing in support of the game plays, wherein the instances are executed on back-end servers, such as a game cloud system, as previously described. The gaming application may include one or more scenarios of interest. As shown, the gaming application includes a first scenario (S-1), a second scenario (S-2), and a third scenario (S-3). For example, a scenario may be rated as being difficult by the developer, or discovered through game play of the gaming application. In other examples, a scenario may be an intriguing part of the gaming application that is popular with players, such that those players would be eager to learn everything about that scenario.
In one embodiment, the plurality of game plays310is controlled by a plurality of players P-1through P-n, through respective client devices. In another embodiment, the plurality of game plays310may be automatically controlled, such as for purposes of self-training the AI model using the plurality of back-end servers. As shown, the game plays provide various game play data320athrough320n. The game play data may include metadata, including game state data, as previously described. For example, game state data describes the state of the game at a particular point, and may include controller input data. In addition, the game play data320athrough320nmay include recordings of the game plays310athrough310nfor purposes of extracting the metadata and/or training state data.
As shown inFIGS. 3B-1 and 3B-2, capture engine340captures the game play data320athrough320n, as well as other data that may be provided, such as success criteria330. As previously described, the success criteria may be used to differentiate training state data that are similar for purposes of training the AI model. In particular, the success criteria is used to train the AI model to include defining the interrelationships between the nodes of layers within the AI model, wherein the nodes may define features related to the gaming application and/or game play of the gaming application (e.g., controller inputs, game states, etc.). In an example, the success criteria may be utilized to determine which set of training state data is more successful, such as to define the weights of one or more paths (between nodes of one or more layers) through the AI model. In that manner, the AI model can be used to give insight (e.g., strategy) in how to play or choose how to play the gaming application during the scenario. The captured data is provided to a deep learning engine190as training state data345.
As shown inFIGS. 3B-1 and 3B-2, the deep learning engine190includes a modeler120that is configured for training and/or building the AI model using the training state data345based one or more success criteria. The modeler120may implement artificial intelligence through various neural networks (e.g., convolutional, recurrent, etc.). The modeler120may implement AI through various algorithms, including as examples for purposes of illustration, deep learning, reinforcement learning, supervised learning, unsupervised learning, rewards based learning (e.g., through the use of success criteria, success metrics, etc.), incremental learning, etc. The deep learning engine190was previously introduced inFIG. 1B. In particular, the modeler120identifies a set of feature dependent rules that make predictions given a set of inputs (e.g., features that may define a context or condition—game state—of a gaming application) when building the AI model. The predictions may include how successful a given set of inputs may be when playing a scenario. In that manner, the AI model can be used to determine actions to be taken given the set of inputs.
As shown inFIGS. 3B-1 and 3B-2, modeler120of deep learning engine190includes a feature identification engine350that is configured for identifying a plurality of features of the training state data. For each game play of a corresponding scenario, the training state data includes features. For example, at a particular point in the game play, an instance of training state data may be collected, wherein the training instance includes one or more features (e.g., a set of features for the training instance), wherein features may include variables, parameters, controller inputs, game state metadata, etc.
As such, the feature identification engine350is configured to parse through the training state data for purposes of identifying and/or extracting features from the data. The feature identification engine350may also be configured to learn features. At each training cycle through the deep learning engine190, a training instance (e.g., set of features) is provided as input, wherein the training instance may be associated with a particular point in the game play of the scenario. In this manner, the deep learning engine190is configured to incrementally learn about the gaming application, the scenario of the gaming application, and/or game play of the scenario of the gaming application.
As shown inFIG. 3B-1, the modeler120is configured to learn rules defining relationships between features and outputs (e.g., predictions, actions, etc.), wherein features may be defined within one or more nodes that are located at one or more hierarchical levels of the AI model160being trained. The modeler120builds the AI model160by linking the features between the layers such that a given input set of data leads to a particular output of the AI model. As such, the modeler120may be configured to generate the features and/or nodes of the AI model160, as defined by rules that link the features at the various layers. For example, a rule may link (e.g., using relationship parameters including weights) one or more features or nodes through the AI model between an input and an output. That is, one or more linked features make a rule. The AI model160is trained and includes a set of the rules, each corresponding to a labeled or classified output. A more detailed description of the functionality of the modeler120is provided inFIG. 3B-2below.
In particular, as shown inFIG. 3B-2, in the modeler120the identified features identified and/or extracted from the input data by the feature identification engine130may be delivered to a classifier360, which is configured to learn rules defining relationships between features and outputs (e.g., predictions, actions, etc.). The features may be defined within one or more nodes that are located at one or more hierarchical levels of an AI model being trained. Each feature may be linked with one or more features at other layers, wherein one or more relationship parameters define interconnections between a first feature and other features (e.g., a second feature, a third feature, etc.) at other layers in the AI model.
For example, as shown inFIG. 3B-2the classifier360is configured to determine which label or output to which a set of features (the set making a rule) belongs. That is, a rule links a given set of features which may be defined as nodes within the AI model (i.e., the nodes describing a training instance or point in the game play of the scenario of the gaming application) to a specific output, that is labeled by the classifier360. For example, a rule may link one or more features or nodes (the links or interrelationships between features defined through one or more relationship parameters) through the AI model between an input and an output. The classifier may be configured to generate the features and/or nodes of the AI model160, wherein the features and/or nodes are used to define the rules, as described above. The output may be associated with a label that is generated, assigned, and/or determined by the classifier360.
More particularly, the learned rules may be learned paths and/or learned patterns (e.g., through the nodes or features of the AI model that lead to an output node) for a given set of inputs and/or input data relating to game play of a scenario of a gaming application. For example, one or more linked features and/or nodes make a rule. The trained AI model is a set of the rules and labels (i.e., outputs). In a supervised learning environment, the output is predetermined for a given set of features, and the deep learning engine learns the rule that links the set of features (e.g., through the labels) to the output. In an unsupervised learning environment, the given set of features are not automatically tied to an output, wherein the rule may be learned by looking for similarities, or clusters of data points relating to other similar sets of features. Clusters may be preferred over other clusters depending on success criteria defined for training. In either case, an existing rule for the input set of features may be matched, or a new rule may be generated for the input set of features (that is most likely similar to or an evolution of an existing rule or rules). The resulting output according to the learned rule of the AI model may predict how successful the corresponding input set of features may be when used to play the scenario of the gaming application. Further, the output from the AI model (and optionally the input set of features) may be used (e.g., via the analyzer) to determine a course of action to be taken for that particular point in the game play of the scenario (as determined by the condition or game state of the gaming application indicated by the input set of data). For example, the action may include a set of controller inputs to be suggested as a recommendation, control commands for the gaming application to respond to the inputs, etc. One or more rules may be generated for a given set of features, or similar sets of features. Depending on the corresponding success criteria or reward criteria, one rule may be preferred for a given set of features or similar set of features. For a given set of features, the output that is most successful may be chosen for the given set of features. That is, for a given set of features (e.g., input), the most successful rule (and output) is chosen to be used within the trained AI model160, which is evidence of incremental learning.
In a simple illustration, success criteria may be defined by earned points, wherein the path (e.g., linked labels) generating the most points is more successful. In another example, success criteria may be related to the level of a life bar of the player's character or an opponent, as is further described inFIG. 3C-1. Other success criteria may be defined based on the player. For example, by identifying an expert through one or more success criteria, their corresponding game plays may be targeted for use as training state data, and generally used for learning the best actions to take for a given state or condition of a game play of a scenario of a gaming application. In that manner, success criteria may be defined to determine skill level of the player, to include how quick is the players' response time, how accurate is the player in targeting one or more targets (e.g., generally a skilled player has a fast trigger and moves from one target to another quickly, decisively, and accurately), how quick is the period between controller inputs, etc.
As shown inFIGS. 3B-1 and 3B-2, the modeler120builds and/or outputs the trained AI model160, which links the learned paths and/or learned patterns (e.g., linking labels of the AI model) to a given set of inputs and/or input data relating to game play of a scenario of a gaming application. The AI model160can be later used to provide one or more functionalities related to the gaming application and/or game play of the gaming application. That is, given a set of inputs that may indicate a condition of a subsequent game play by a player, the resulting output of the trained AI model160can be used (e.g., via the analyzer) to predict and/or determine the best course of action to be taken for that particular point in the game play of the scenario as defined by the corresponding set of input data. For example, a player may be playing the scenario of the gaming application after the AI model160has been trained. The player is also encountering difficulty in progressing through the scenario, which may be reflected in the output of the AI model. New and subsequent input state data (e.g., game state) may be related to any data related to that particular point in a game play of that player (where difficulty is experienced). That input state data for the scenario is received and provided to the AI model via the deep learning engine190, wherein the AI model may predict as an output how successful the game play will be when playing the scenario given the given condition of the gaming application. The output from the AI model can be analyzed and used to perform various functionalities related to the gaming application and/or the game play of the gaming application. For example, the output may be analyzed to determine the best course of action to be taken for that particular point in the game play of the scenario. An action may be performed based on the output. For example, the trained AI model160may provide a recommendation to the player to advance his or her game play.
Other functionalities may also be determined and generated based on the output of the AI model160for a given input set of data, wherein the functionalities are manifested in the form of an output or action. That is, input state data is received during a first game play of the scenario, and provided as input to a deep learning engine. The deep learning engine applies the AI model to the input state data. An action is performed based on an analysis of the output of the AI model, wherein the action is performed in response to the input state data.FIGS. 4A-4Billustrate the application of the AI model for a given game play of a gaming application.FIGS. 5A-5Fprovide various illustrations of different actions or responses that can be performed depending on the objective predefined. In particular, the AI model, as implemented through the deep learning engine190, matches a given input state data to one or more rules (each rule providing linked or interconnected nodes and/or features) defined within the trained AI model. Each rule is associated with an output. A success criteria may be applied to generate the rule. In addition, an analyzer140takes the output and performs additional analysis to determine the appropriate action in relation to the corresponding input data. For example, when the rule is satisfied with respect to the success criteria for a given set of input state data, a corresponding action may be identified and/or performed.
FIG. 3C-1is a screen shot300C of a game play of a gaming application and illustrates one type of success criteria that may be used to train the AI model that understands how to play the gaming application, in accordance with one embodiment of the present disclosure. In particular, player P-1controls the game play. As shown, screen shot300C shows a frontal view of the battle between Kratos391and the enemy396. In the “God of War” gaming application, Kratos is a Spartan warrior of Greek mythology, who is tasked with killing Ares, the God of War. In the game play, a player may control Kratos391. Training state data320afrom the game play associated with the screen shot300C is fed to the deep and/or machine learning engine190for purposes of training the AI model, previously described. In addition, success criteria330is supplied to the deep and/or machine learning engine190, wherein the success criteria is used by the deep learning engine190to train the AI model. For example, the success criteria may be used to differentiate paths through the AI model (where the paths traverse nodes in the AI model). In embodiments, one or more success criteria may be defined for use in training the AI model. For example, success criteria may include earning the most points, earning the most assets or most powerful assets or a critical asset, etc.
In one example, the success criteria330may be defined by whether a life bar397of the enemy396becomes low during the game play, wherein the life bar indicates health of the corresponding character. Also, the life bar392indicates the health of Kratos391during the game play. As shown, the life bar397of the enemy is extremely low, and may indicate that the game play is successful or has a high degree of success. In another example, the success criteria may be more sophisticated and is defined by how quickly the life bar397of enemy396is depleted. If the life bar397is quickly depleted, this indicates that the game play is being controlled by a highly skilled player. For example, the highly skilled player understands how to battle enemy397, and knows the sequences of controller inputs used to defeat the enemy396, and has the skill and/or ability to implement those sequences quickly and accurately (e.g., without deviation). During training of the AI model, using success criteria that focuses on how quickly the life bar is depleted may differentiate successful inputs over less successful inputs, and help identify controller input sequences to be used during a particular condition or game state of the game play that would be successful if implemented by another player.
FIG. 3C-2is an illustration of the collection of training state data from a plurality of game plays, wherein a plurality of instances is executing on a plurality of servers in support of the game plays, in accordance with one embodiment of the present disclosure. A plurality of game plays310(e.g.,310athrough310n) of a gaming application is shown. The plurality of game plays310is controlled by a plurality of players P-1through P-n, through respective client devices. As shown, the game plays provide various game play data320athrough320n, of which extracted data may be used as training state data345. The game play data may include metadata, including game state data, as previously described. Each of the game plays is playing scenario1(S-1) of the gaming application, wherein each game play is unique and has a unique outcome. That is, each game play may be associated with a measured degree of success of progressing through scenario S-1. That degree of success may be measured in part using success criteria330, wherein the success criteria330may be used by the modeler120of the deep and/or machine learning engine190to train the AI model160, as previously described. For example, success criteria may be used to differentiate training state data that are similar for purposes of training the AI model.
With the detailed description of the various modules of the system100A and system200B, flow diagram400A discloses a method for applying an AI model that knows everything about the gaming application and/or game plays of the gaming application based on one or more success criteria (e.g., understands how to play a gaming application), the AI model being used for provide various functionalities in relation to the gaming application and/or game plays of the gaming application, in accordance with one embodiment of the present disclosure. Flow diagram400A may be implemented within a back-end server (e.g., within the game cloud system201in combination with a deep learning engine190), previously described.
In particular, at410, the method includes training an AI model from a plurality of game plays of a scenario of the gaming application. For example, the AI model is trained for playing the scenario of a gaming application using training state data collected from a plurality of game plays of the scenario and associated success criteria of each of the game plays. That is, the AI model is trained to learn the intricacies of the gaming application and/or the intricacies of playing the gaming application or a scenario of the gaming application.
In one embodiment, the AI model is trained using training state data collected across a plurality of game plays of the gaming application, wherein the plurality of game plays is controlled by a plurality of players via a plurality of client devices. For example, a plurality of servers may be executing a plurality of instances of the gaming application, wherein the instances support the plurality of game plays. Training state data is collected at the plurality of servers, wherein the training state data is associated with corresponding game plays of the scenario. For example, the training state data may include metadata associated with the game plays, to include controller inputs, game state defining the state and/or condition of the gaming application at a particular point during a game play, progress through the game play, results (e.g., success or failure) of the scenario, user profile information, etc., as previously described. That is, the training state data includes any data that may be relevant to understanding the gaming application and/or game plays of the gaming application.
In addition, one or more success criteria is defined for the training state data. For example, success criteria is used to train the AI model that understands how to play the gaming application. For example, the success criteria may be used to define favorable links between nodes of the AI model, or to define favorable paths through the nodes of the AI model for a given set of input data. The success criteria and the training state data are provided to a deep learning engine to train the AI model.
At420, the method includes receiving first input state data during a first game play of the scenario. In one embodiment, the first game play occurs after training the AI model. That is, the first game play is not used for training the AI model. In another embodiment, the first game play may occur during training, such that the first game play is used for training the AI model (e.g., self-training mode). In still another embodiment, the first game play may not initially be used for training the AI model, and is used primarily for determining an appropriate action to be taken given the first input state data. Thereafter, the first game play (e.g., the first input state data) may be used to refine the AI model through additional training.
At430, the method includes applying the first input state data to the AI model to generate an output. In one embodiment, the output may indicate a degree of success for the scenario for the first game play. That is, the output may predict how successful the first game play will be in progressing through the scenario.
At440, the method includes performing an analysis of the output based on a predefined objective. In addition, the set of inputs (e.g., current and past sets of inputs) may also be analyzed. Depending on the predefined objective, the analysis may produce an action to be performed for a particular point in the corresponding game play of the scenario (as determined by the condition or game state of the gaming application indicated by the input set of data). For example, if the predefined objective is to provide assistance, the analysis may produce a recommendation or advice on how to progress through the encountered condition during game play of the scenario of the gaming application. If the predefined objective is to provide coaching, the analysis may determine a weakness of the player, and provide tutorial sessions for the player to address the weakness. Other predefined objectives are supported, such as to provide gaming support, provide parity in game plays, to automatically train the AI model, to find flaws in the gaming application, to automatically test the gaming application, etc.
As such, at450, the method includes performing the action to achieve the predefined objective based on the output that is analyzed.FIGS. 5A-5Fprovide various illustrations and examples of different actions or responses that can be performed depending on the objective predefined.
FIG. 4Bis block diagram of AI processor engine210ofFIG. 2Bthat is configured for applying an AI model that is configured to understand how to play a gaming application, the AI model being used for provide various functionalities in relation to the gaming application and/or game plays of the gaming application, in accordance with one embodiment of the present disclosure. As shown, the AI processor engine210includes the deep learning engine or machine learning engine190that is configured to train and/or apply the AI model160, wherein the AI model160is configured to provide an output for a given input. The AI processor engine210also includes an analyzer for using the output from the AI model160to determine a response for the input.
In particular, input state data405is provided as input to the deep learning engine190that is configured to implement the AI model160. The input state data may be associated with game play of the gaming application, to include controller inputs, game state data, user data, success criteria, etc., as previously described. The input state data405may be provided after the AI model160is trained, in some embodiments. In other embodiments, the input state data405may be provided during the training of the AI model160(e.g., self-training the AI model). As previously described, the AI model160generates an output173for the given input state data405. For example, the output may indicate how successful the game play may be in progressing through a scenario of the gaming application. The output may be analyzed by the analyzer140to provide various functionalities in relation to the game play of the gaming application. The analyzer140determines the appropriate action or response depending on a predefined objective, as previously described (e.g., coaching, providing assistance, etc.). That is, the AI model160may be used by the analyzer140to determine what is needed to direct the game play to a successful end during the scenario. As such, the analyzer140may determine and perform an action depending on the predefined objective for a given input state data. In that manner, the analyzer through the use of the AI model160can provide various functionalities related to the gaming application and/or the game play of the gaming application.
As shown, as part of the analysis the analyzer140may perform one or more functions. For example, the player profiler144aof the analyzer140is configured to perform profiling of the player playing the gaming application (e.g., determine skill level of the player). The weakness identifier141aof the analyzer140is configured to identify a weakness in the gaming ability of the player. The input control sequence parser147aof the analyzer140is configured to determine the sequence of controller inputs used by the player to control the game play. The sequence map and/or route tracker148aof the analyzer140is configured to track the progress of the game play, to include tracking the progress through a gaming environment. For example, the tracker148ais configured to track routes taken in the gaming environment, or to build a map of the gaming environment, or to build a node map of the game play indicating progress through the gaming application. Still other modules are supported that provide an analysis of the output of the AI model for a given input state data405.
The analyzer140is further configured to determine and perform an action by the action generator170in response to the analysis of the output determined in response to the given input state data405. The action is determined and performed depending on the predefined objective. For example, the action may provide services to the player playing the gaming application (e.g., provide a profile of a player playing the gaming application, providing recommendations to a player during game play of a gaming application by the player wherein the recommendation may be structured in consideration of the user profile, finding weaknesses of the player, provide services to address those weaknesses, training the player, providing a bot opponent to the player, take over game play for the player, etc.). The action may also provide services to the game developer or other entity, such as finding flaws in the gaming application, building equal teams using player profiles that are competing against each other in the gaming application such as in a multi-player online game, automatically training the AI model through auto-play, explore the gaming environment of the gaming application to discover flaws, etc.FIGS. 5A-5Fprovide various illustrations of different actions or responses that can be performed depending on the objective predefined.
In particular,FIG. 5Aillustrates a data flow diagram illustrating a process for providing recommendations to a player playing a gaming application, the recommendation provided through an AI model160trained through a network of back-end servers executing instances of a gaming application, in accordance with one embodiment of the present disclosure. A game play501xof a gaming application is shown. The game play may be controlled by a player P-x through a corresponding client device, wherein an instance of the gaming application is executing at a back-end server of a streaming gaming service (e.g., cloud game system), as previously described. In other embodiments, the gaming application may be executing locally on the client device, and metadata is delivered to a back-end server for AI model support. The game play501xis occurring during scenario S-1.
Input state data505xfrom the game play501xis provided to the analyzer140that is configured to analyze the output of the trained AI model160, wherein the AI model160is implemented through the deep learning engine190(in an application phase and not a learning phase). The input state data505xis received after the AI model160is trained, and as such may not be part of the training state data used for training the AI model, previously described. The AI model160is configured to provide an output173, and the analyzer140is configured to perform or provide an action to be performed based on an analysis of the output173.
In particular, the action generator170of the analyzer140includes a recommendation engine145athat is configured to provide a recommendation to a player during game play of a gaming application by the player. For example, the predefined objective may be to provide gaming support to player P-x, as implemented through the analyzer140. The recommendation may be provided in response to a condition (e.g., game state) in the game play, such as when the player is having difficulty navigating through a particular scenario of the gaming application. In one embodiment, the analyzer140can determine that the player is struggling by determining that the first input state data, or the features in the first input state data does not satisfy a corresponding success criteria. The success criteria may provide an indication on how successful the game play will turn out, and in particular how successful that portion of the game play will turn out. For example, the success criteria may indicate how many points are accumulated during the scenario. When the analyzer140determines from the output of the AI model160that the first input state data does not meet the success criteria, then a recommendation515may be provided, as an action, to the client device of the player P-x.
In one embodiment, the recommendation may take into account the user/player profile510xof the player P-x (e.g., consider the skill level of the player), wherein the player profiler144amay analyze the output data from the trained AI model160and/or the input state data to determine the user profile510x(e.g., how the player reacts to the gaming application during the scenario, skill level, and other player characteristic metrics) for the scenario. That is, the recommendation reflects the skill of the player P-x. In other embodiments, the user profile510xmay be determined through historical game play of the subject gaming application, as well as other gaming applications. For example, if the user profile510xindicates that the player is an expert gamer, then the recommendation515for player P-x may provide small hints without too much detail. On the other hand, if the user profile510xindicates that the player P-x is a beginning gamer, then the recommendation515may provide detailed instructions for the player to follow.
In one embodiment, the recommendation515may include a query asking if the player P-x would like to engage an auto play feature. If yes, the task auto play engine145bis configured to take over the game play of player P-x. In that manner, the task auto play engine145bautomatically (and with success) performs the scenario or a task within the scenario in the gaming application.
FIG. 5Billustrates a data flow diagram illustrating a process for balancing teams of players playing a gaming application using player profiles determined through an AI model trained through a network of back-end servers executing instances of a gaming application, in accordance with one embodiment of the present disclosure. A plurality of game plays501(e.g.,501athrough501n) of a gaming application is shown. The game plays may be controlled by a plurality of players P-1through P-n through corresponding client devices. As previously described, instances of the gaming application are executing on back-end servers of a streaming gaming service (e.g., cloud game system), as previously described. In other embodiments, the gaming application may be executing locally on the client device, and metadata is delivered to a back-end server for AI model support. The game plays501a-501nmay be occurring during one or more scenarios S-1, S-2, and S-3, for example.
Input state data505athrough505nfrom the game plays501athrough501nis provided to the analyzer140that is configured to analyze the output of the trained AI model160, wherein the AI model160is implemented through the deep learning engine190(in an application phase and not a learning phase). In one embodiment, the plurality of input state data505a-505nis received during a second plurality of game plays of a scenario of a gaming application, wherein the game plays may be conducted after training the AI model. The second plurality of game plays is controlled by a plurality of players, and wherein the plurality of input state data includes a plurality of player characteristic metrics of the plurality of players. The plurality of input state data is applied to the AI model to generate a plurality of outputs indicating a plurality of degrees of success for the second plurality of game plays of the scenario. In addition, the plurality of outputs is analyzed to determine a plurality of player profiles based on the plurality of player characteristic metrics. The AI model160is configured to provide an output, and the analyzer140is configured to perform or provide an action to be performed based on that analysis of the output. In particular, the analyzer140is configured to build balanced teams of players from the plurality of players based on corresponding player profiles to achieve a predefined objective.
During the analysis, the analyzer140may perform user profiling, such as by the player profiler144a. In particular, the output from the trained AI model160is provided to player profiler144athat is configured to provide profiles of the players P-1through P-n playing the gaming application. The player profiles510athrough510nmay be limited to game play of a particular scenario of the gaming application (e.g., profile determines or predicts how a corresponding player reacts to the gaming application during the scenario, the skill level of the corresponding player, etc.). For example, player profiles may be defined based on a plurality of player characteristic metrics, wherein each player profile includes one or more corresponding player characteristic metrics. The input state data may be received during game plays of the gaming application by the players. The input state data including a plurality of player characteristic metrics of the plurality of players. The input state data is also applied to the AI model160to determine an output, as previously described. The player profiler144aanalyzes the output173and/or the input state data505a-505nto determine corresponding player profiles of corresponding players. For example, a corresponding player profile is based on corresponding player characteristic metrics provided as an input to the trained AI model. For example, player characteristic metrics may include: accuracy of the player; or speed of generating a sequence of input control data by a corresponding player; or reaction time of the corresponding player when responding to an event in the scenario; or consistency of the corresponding player; or transition time of the corresponding player between a first target and a second target, etc. In other embodiments, a corresponding player profile may be determined through historical game play of the subject gaming application, as well as other gaming applications, a previously described.
Also, the action generator170of the analyzer140includes a team balancer module144bthat is configured to build teams (e.g., teams that are competing against each other in the gaming application) that are fairly balanced using player profiles according to a predefined objective. For example, the predefined objective may be to provide parity in game plays (e.g., in a massively multi-player online gaming application—MMO), as implemented through the analyzer140. For example, based on the player profiles510athrough510n, the team balancer144bis configured to build teams (e.g., teams that are competing against each other in the gaming application) that are fairly balanced. In that manner, the game plays of the gaming application between the two teams will be fair, engaging, and worthwhile to the players on each team. For example, balanced teams will help to avoid blowout games. Consider player profiles that are further compartmentalized into two groups (G1and G2) based on skill levels in one or more categories. It is desirable to spread players, and/or skills of players of each group evenly throughput all teams (e.g., team-1through team-n). Each team in a gaming application centered on combat includes one or more positions, including a leader, a sniper, a trained killer (e.g., hand to hand combat). In one illustration, you can group all the skill positions and then evenly spread them out between the teams. In the illustration, the leader is excluded, and for simplicity all leaders are selected from a single group G-2, such that all teams have equally strong leaders. It is assumed that all other skilled positions are valued equally. That is, each team should be made up of a number of skilled positions with equal numbers of players taken from each group G1and G2. For example, each team includes on player from group G1, and one player from group G2(excluding the leader). As shown, team-1and team-2each includes one player from group G1(the sniper) and one player from G2(e.g., the killer). Also, team-3includes one player from group G1(the killer) and one player from G2(e.g., the sniper). As such, each team is balanced according user profiles (e.g., skill level).
FIG. 5Cillustrates a data flow diagram illustrating a process for training an AI model through auto play directed by the AI model160, in accordance with one embodiment of the present disclosure. As previously described, the AI model160is trained through a network of back-end servers executing instances of a gaming application. A plurality of game plays310athrough310nof a gaming application is shown. As previously described, game play data320athrough320nis produced from the game plays310athrough310n. In one embodiment, the plurality of game plays is controlled by the AI model160during training. In that manner, the input state data is generated by the analyzer140using data from the AI model160, wherein the input state data is provided as a previous action by the analyzer140during training. The game play data may include metadata, including game state data that describes the state of the game at a particular point, and may include controller input data, and may include recordings of the game plays310athrough310nfor purposes of extracting the metadata and/or training state data. Capture engine340captures the game play data320athrough320n, as well as other data that may be provided, such as success criteria330, as previously described. Success criteria may be used to differentiate training state data that are similar for purposes of training. That is success criteria may be used by the deep learning engine190to train the AI model160. Training state data345is provided to the deep learning engine190. The function of the deep learning engine190was described in relation to FIGS.3B1and3B-2, and is equally applicable to the deep learning engine190shown inFIG. 5C. Not all components of the deep learning engine190are shown inFIG. 5Cfor simplicity. Generally, the deep learning engine190includes a modeler120that is configured for training and/or building the AI model160using the training state data345and success criteria330. The modeler120may implement artificial intelligence through various neural networks (e.g., convolutional, recurrent, etc.). In particular, the modeler120identifies a set of feature dependent rules that make predictions and/or determine actions to be taken given a set of inputs (e.g., features that may define a context or condition—game state—of a gaming application). For example, the output of the AI model160may predict the success of a given set of input data (e.g., defining a condition of a game play of a gaming application) in progressing through a scenario of the gaming application. The set of rules connecting features and/or nodes make up the AI model160.
As shown, the deep learning engine acts in both the learning and application phases. In particular, the game plays310are automatically executed, such as by the AI model160. In that manner, the AI model160is self-trained. Specifically, the analyzer140includes an input control sequence parser571, a player profiler144a, and a map/route tracker572, each of which is previously introduced. For example, the input control sequence parser147ais configured to determine the sequence of controller inputs used by the player to control the game play. The sequence map and/or route tracker148aof the analyzer140is configured to track the progress of the game play, to include tracking the progress through a gaming environment. The player profiler144aof the analyzer140is configured to perform profiling of the player playing the gaming application (e.g., determine skill level of the player).
In addition, the action generator170of the analyzer140includes an auto player143athat is configured to auto play the gaming application, such as for purposes of automatically training the AI model, as directed by the AI model trainer143b, according to the predefined objective. For example, the predefined objective may be to automatically train the AI model160. Specifically, the analyzer140includes the auto player143athat is configured to auto play the gaming application, as directed by the AI model trainer143b. For example, for a given set of training state data, such as for a training instance, a learned output of the AI model160may be analyzed by the analyzer140to determine the next set of controller inputs for controlling the corresponding game play. In particular, the bot selector and input control selection/prediction engine537of the AI model trainer143bcombined are configured to determine the next set of controller inputs depending on the objective. For example, a bot may be selected that is optimized for getting the through the game with the best success and efficiency, or exploring the different permutations of the gaming application, etc. Depending on the objective, the input control selection/prediction engine537is configured to predict what the next set of controller inputs would be given the objective and the selected bot. The controller inputs are delivered back to the auto-player143a, which then sends the controller inputs to the corresponding instance of the gaming application supporting a corresponding game play. In that manner, by utilizing a bank of back-end servers for executing instances of the gaming application for corresponding game plays310a-310n, the training state data345may be automatically generated quickly and efficiently.
In one embodiment, for the first input state data, a plurality of sets of input controls or controller inputs may be determined to be played subsequent to the first input state data. A first set of input controls is selected that is predicted to satisfy the success criteria, and predicted to have a highest degree of success among the plurality of sets of input controls. In that manner, the first set of input controls can be delivered as the action to a corresponding instance of the gaming application for execution.
FIG. 5Dillustrates a data flow diagram illustrating a process for automatically sweeping through a gaming application using an auto-sweeper/player146adirected by an AI model160that is trained through a network of back-end servers executing instances of the gaming application, in accordance with one embodiment of the present disclosure. A plurality of game plays501(e.g.,501athrough501n) of a gaming application is shown. The game plays may be controlled automatically, such as through a AI model160. In that case, the AI model160may be configured to provide the next input state data (e.g., controller input that is used to generate game state and other input state data). Instances of the gaming application may be executing on back-end servers of a streaming gaming service (e.g., cloud game system. The game plays501a-501nmay be occurring during one or more scenarios S-1, S-2, and S-3, for example. Input state data505athrough505nfrom the game plays501athrough501nis provided to the analyzer140that is configured to analyze the output of the AI model160. The deep learning engine190implements the AI model160(in an application phase and not a learning phase). The AI model160is configured to provide an output, and the analyzer140is configured to perform or provide an action to be performed based on that analysis of the output.
In particular, the game plays310are automatically executed, such as by the AI model160. For example, the predefined objective may be to automatically test the gaming application after the AI model has been trained. For instance, the AI model may have been previously automatically self-trained and the AI model may now be used to test the gaming application. In another embodiment, the auto-sweep feature is performed during training. Specifically, the analyzer140includes an input control sequence parser571and a map/route tracker572, each of which is previously introduced. For example, the input control sequence parser147ais configured to determine the sequence of controller inputs previously tested. The sequence map and/or route tracker148aof the analyzer140is configured to track the progress of the game play during testing, to include tracking the progress through a gaming environment.
In one embodiment, first input state data is generated by the AI model as a previously action taken during training. During analysis of the output of the AI model, different permutations are then determined for responding to the first input state data. For example, each permutation includes a unique set of controller inputs to be taken. Thereafter, one or more actions are taken, such as executing the different permutations. In that manner, the system may be configured to discover any anomaly in playing the gaming application.
In particular, the action generator170of the analyzer140includes the auto sweeper engine/player146athat is configured to explore the gaming application, such as by using a permutation engine146bto determine the various permutations that a gaming application can perform. For example, the input control permutation engine146bis configured to determine the various permutations that a gaming application can perform, such as in response to a given condition (e.g., a particular game state). That is, the input control permutation engine146bis able to determine what should be the next set of controller input for a given input state data405a-405n(e.g., game state). In this case, the permutation engine146bdiscovers the different permutations for responding to the input state data, each permutation including a different set of actions to be taken. The auto sweeper engine146ais then configured to go through the various permutations by controlling the game plays401a-401n(e.g., by submitting appropriate controller inputs to the executing instances of the gaming application).
Further, the analyzer140is configured to perform quality analysis on the gaming application, such as for purposes of discovering weak points in the gaming application (e.g., excessively long and boring sequences, difficult sections, etc.), or flaws (e.g., glitches, loops, etc.). For example, the ma/route analyzer441is configured to analyze the output (e.g., game states) of the different permutations of the gaming application to discover the weak points in the gaming application. In one implementation, the game code identifier443is configured to discover a problem in the coding of the gaming application, wherein the code location447is provided as an output.
FIG. 5Eillustrates a data flow diagram illustrating a process for providing an opponent for a player, wherein the opponent is directed by an AI model that is trained through a network of back-end servers executing instances of the gaming application, in accordance with one embodiment of the present disclosure. A game play501xof a gaming application is shown. The game play may be controlled by a player P-x through a corresponding client device, wherein an instance of the gaming application is executing at a back-end server of a streaming gaming service (e.g., cloud game system), as previously described. In other embodiments, the gaming application may be executing locally on the client device, and metadata is delivered to a back-end server for AI model support. The game play501xis occurring during scenario S-1, and is controlled by player input controls503x.
Input state data505xfrom the game play501xis provided to the analyzer140that is configured to analyze the output of the trained AI model160, wherein the trained AI model160is implemented through the deep learning engine190(in an application phase and not a learning phase). The input state data505xis received after the AI model160is trained, and as such may not be part of the training state data used for training the AI model, previously described. The AI model160is configured to provide an output, and the analyzer140is configured to perform or provide an action to be performed based on that analysis of the output.
Specifically, the analyzer140includes an input control sequence parser571and a player profiler144a, each of which is previously introduced. For example, the input control sequence parser147ais configured to determine the sequence of controller inputs used by the player to control the game play. The player profiler144aof the analyzer140is configured to perform profiling of the player playing the gaming application (e.g., determine skill level of the player).
In addition, the analyzer140includes a bot builder575, which is configured to build one or more bots (automatic player robots or opponents to be used for game play or to control characters in the game play). For example, the bot builder575may be configured to build an ultimate bot142dthat is learned by the AI model160through application of success criteria. As previously described, for a given set of inputs (e.g., input training data), a more successful pattern (e.g., rule including linked features and/or labels) is learned and chosen over a less successful pattern. In that manner, the best or ultimate bot142dis trained that is the most successful at playing the gaming application. In another example, the bot builder575is configured to build a virtual player or virtual me bot142cthat simulates the first player. In one embodiment, the training state data is obtained from game plays by the first player at one or more client devices. That is, data from game plays of other players are not used for training the virtual me bot142c. In that manner, the AI model is learned off of metadata created by game plays of the first player, and as such, the AI model would be a direct reflection of that first player.
In still another example, the bot builder575is configured to build one or more bots of varying skill levels. For example, various skill bots may include an expert bot576, an intermediate bot577, and a beginner bot578. In one implementation, the training state data may be parsed to reflect the corresponding skill level, such that the AI model is trained using data of a corresponding skill level. For example, the success criteria may be defined such that only game plays of expert players are used to train the AI model in order so that the bot builder575can build the expert bot576. In another example, success criteria may be defined such that only game plays of players of intermediate skill are used to train the AI model in order so that the bot builder575can build the intermediate bot577. In still another example, success criteria may be defined such that only game plays of players of a beginner skill are used to train the AI model in order so that the bot builder575can build the beginner bot578. In still another embodiment, a bot of a particular skill level may be implemented by using the virtual ultimate bot142dand applying one or more conditions to the performance of the ultimate bot, to include introducing randomness and/or latency. For example, the performance of the ultimate bot may be compromised by introducing latency between execution of controller inputs in a sequence, or by introducing random controller inputs into a given sequence of controller inputs that are known to be highly successful in accomplishing a task.
In addition, the action generator170of the analyzer140includes an auto player143athat is configured to auto play the gaming application, such as for purposes of automatically training the AI model, as directed by the AI model trainer143b, according to the predefined objective. For example, the predefined objective may be to automatically train the AI model160. Specifically, the analyzer140includes the auto player143athat is configured to auto play the gaming application, as directed by the AI model trainer143b.
The analyzer140includes the auto player143athat is configured to auto play the gaming application for purposes of providing a bot opponent (e.g., automatically reactive robot opponent). The bot opponent selector142ais configured to select the appropriate bot, such as those bots previously introduced (e.g., ultimate bot142d, virtual me bot142c, and variously skilled bots to include the expert bot576, intermediate bot577, or beginner bot578). A player controlling the game play may control a character that is going up against a bot opponent. In particular, the auto player143ais configured to implement the selected automatic robot (bot). For a given set of input state data505x, an output may be analyzed by the analyzer140by the auto player143ato determine the next set of instructions for controlling a bot found in the gaming application.
In addition, a bot throttler142bis configured to apply a difficulty setting as reflected in the corresponding bot (e.g., to the operations of the bot). The bot throttler may start with the ultimate bot142d, or any other learned bot. The difficulty setting may be applied actively by the player, or may be applied according to a user profile. For example, if the player was an expert as indicated through the profile, then the difficulty would be set higher (opponent bot would be difficult to beat). On the other hand, if the player were a beginner, then the difficulty setting would be lower (opponent bot would be easy to beat). As shown, the bot throttler142bmay include a randomizer551configured to introduce random instructions. A corresponding bot with a high degree of difficulty would have a low level of random instructions inserted into a stream of instructions that is normally used for controlling the ultimate bot142d. As such, the resulting bot would be similar to the ultimate bot142d. On the other hand, a resulting bot having a low degree of difficulty would have a high level of random instructions inserted into the stream of instructions used for normally controlling the ultimate bot142d. Because there are random instructions, the resulting bot would act erratically, and not perform as well as the ultimate bot142d. In addition, a latency engine553may be implemented to apply a difficulty setting. For a resulting bot with a high degree of difficulty, there would be limited or no latency introduced into the stream of instructions that would normally be applied for an ultimate bot142d. On the other hand, for a resulting bot with a low degree of difficulty (e.g., for a beginner), there would be a greater amount of latency introduced into the stream of instructions that would normally be applied for an ultimate bot142d. In that manner, the resulting bot having a low difficulty setting would then be acting very slowly, such as during an attack, or during a defensive maneuver and as such would likely be easily defeated.
FIG. 5Fillustrates a data flow diagram illustrating a process for identifying weaknesses of a player, and providing various services to train the player to overcome those weaknesses, in accordance with one embodiment of the present disclosure. A game play501xof a gaming application is shown. The game play may be controlled by a player P-x through a corresponding client device, wherein an instance of the gaming application is executing at a back-end server of a streaming gaming service (e.g., cloud game system), as previously described. In other embodiments, the gaming application may be executing locally on the client device, and metadata is delivered to a back-end server for AI model support. The game play501xis occurring during scenario S-1, and is controlled by player input controls503x.
Input state data505xfrom the game play501xis provided to the analyzer140that is configured to analyze the output of the trained AI model160, wherein the trained AI model160is implemented through the deep learning engine190(in an application phase and not a learning phase). The input state data505xis received after the AI model160is trained, and as such may not be part of the training state data used for training the AI model, previously described. The AI model160is configured to provide an output, and the analyzer140is configured to perform or provide an action to be performed based on that analysis of the output.
In addition, the action generator170of the analyzer140includes an auto player143athat is configured to auto play the gaming application, such as for purposes of automatically training the AI model, as directed by the AI model trainer143b, according to the predefined objective. For example, the predefined objective may be to automatically train the AI model160.
In particular, the analyzer140includes a weakness identifier141athat is configured to determine a weakness of a corresponding player that is controlling a corresponding game play according to a predefined objective to provide coaching. The weakness is determined through analysis of the game play of the player.
For example, the analysis may include comparisons to a success criteria. In particular, a weakness may be identified by determining that first input state data has a lower than average degree of success, as determined by analysis of the output from the AI model in comparison to other outputs resulting from comparable input state data. As an illustration, for a given input state data (e.g., first input state data), a first set of interconnected nodes that produce an output in response to the first input state data may have a lower value when satisfying a corresponding success criteria than an average value determined for similar game plays having comparable input state data. That is, the comparison is made with game plays having the same or similar input state data. As such, the weakness identifier141amay determine a weakness of the player.
Further, a weakness trainer141bis configured to perform services that help the player to overcome the weakness. For example, the weakness trainer141bmay provide one or more tutorials561(e.g., videos, gaming sessions, etc.) that are targeted at improving the skills of the player in relation to the player's weakness. The tutorials may include video tutorials that provide instructions on how to strengthen skills of the player related to the weakness, or gaming sessions that are directed at strengthen skills related to the weakness. In addition, a training session addressing the identified weakness may be presented to the player via a corresponding client device. Also, the weakness trainer141bmay be configured to provide a bot that is specialized to direct the game play in a manner that exposes the weakness of the player, in one embodiment. The bot may be a virtual ultimate opponent565for the player to play against. The virtual ultimate opponent565may be the ultimate bot142d, as previously described.
FIG. 6illustrates components of an example device600that can be used to perform aspects of the various embodiments of the present disclosure. For example,FIG. 6illustrates an exemplary hardware system suitable for training an AI model that is capable of performing various functionalities in relation to a gaming application and/or game plays of the gaming application, in accordance with one embodiment of the present disclosure. This block diagram illustrates a device600that can incorporate or can be a personal computer, a server computer, gaming console, mobile device, or other digital device, each of which is suitable for practicing an embodiment of the invention. Device600includes a central processing unit (CPU)602for running software applications and optionally an operating system. CPU602may be comprised of one or more homogeneous or heterogeneous processing cores.
In accordance with various embodiments, CPU602is one or more general-purpose microprocessors having one or more processing cores. Further embodiments can be implemented using one or more CPUs with microprocessor architectures specifically adapted for highly parallel and computationally intensive applications, such as media and interactive entertainment applications, of applications configured for deep learning, content classification, and user classifications. For example, CPU602may be configured to include the AI engine (e.g., deep learning)190that is configured to support and/or perform learning operations with regards to providing various functionalities (e.g., predicting) in relation to a gaming application and/or game plays of the gaming application. The deep leaning engine may include a modeler120that is configured for building and/or training the AI model that is configured to provide the various functionalities related to the gaming application and/or game plays of the gaming application. Further, the CPU602includes an analyzer140that is configured for implementing the trained AI model. The trained AI model provides an output in response to the input, wherein the output is dependent on the predefined functionality of the trained AI model. The trained AI model may be used to determine what actions can be taken during the game play. The analyzer140determines which is the appropriate action to be taken. That is, the analyzer140is configured to perform various functionalities in relation to a gaming application and/or game plays of the gaming application. The analyzer140is configured to analyze the output from the trained AI model160for a given input (e.g., controller input, game state data, success criteria), and provide a response.
Memory604stores applications and data for use by the CPU602. Storage606provides non-volatile storage and other computer readable media for applications and data and may include fixed disk drives, removable disk drives, flash memory devices, and CD-ROM, DVD-ROM, Blu-ray, HD-DVD, or other optical storage devices, as well as signal transmission and storage media. User input devices608communicate user inputs from one or more users to device600, examples of which may include keyboards, mice, joysticks, touch pads, touch screens, still or video recorders/cameras, and/or microphones. Network interface614allows device600to communicate with other computer systems via an electronic communications network, and may include wired or wireless communication over local area networks and wide area networks such as the internet. An audio processor612is adapted to generate analog or digital audio output from instructions and/or data provided by the CPU602, memory604, and/or storage606. The components of device600, including CPU602, memory604, data storage606, user input devices608, network interface610, and audio processor612are connected via one or more data buses622.
A graphics subsystem614is further connected with data bus622and the components of the device600. The graphics subsystem614includes a graphics processing unit (GPU)616and graphics memory618. Graphics memory618includes a display memory (e.g., a frame buffer) used for storing pixel data for each pixel of an output image. Graphics memory618can be integrated in the same device as GPU616, connected as a separate device with GPU616, and/or implemented within memory604. Pixel data can be provided to graphics memory618directly from the CPU602. Alternatively, CPU602provides the GPU616with data and/or instructions defining the desired output images, from which the GPU616generates the pixel data of one or more output images. The data and/or instructions defining the desired output images can be stored in memory604and/or graphics memory618. In an embodiment, the GPU616includes 3D rendering capabilities for generating pixel data for output images from instructions and data defining the geometry, lighting, shading, texturing, motion, and/or camera parameters for a scene. The GPU616can further include one or more programmable execution units capable of executing shader programs. In one embodiment, GPU616may be implemented within AI engine190to provide additional processing power, such as for the AI or deep learning functionality.
The graphics subsystem614periodically outputs pixel data for an image from graphics memory618to be displayed on display device610, or to be projected by projection system640. Display device610can be any device capable of displaying visual information in response to a signal from the device600, including CRT, LCD, plasma, and OLED displays. Device600can provide the display device610with an analog or digital signal, for example.
Accordingly, the present disclosure described systems and methods implementing deep learning (also referred to as machine learning) techniques to build an AI model using training data collected from a network of servers executing instances of a gaming application supporting one or more game plays, and for using the trained AI model to provide various functionalities relating to the gaming application and/or game plays of the gaming application, in various embodiments.
It should be understood that the various embodiments defined herein may be combined or assembled into specific implementations using the various features disclosed herein. Thus, the examples provided are just some possible examples, without limitation to the various implementations that are possible by combining the various elements to define many more implementations. In some examples, some implementations may include fewer elements, without departing from the spirit of the disclosed or equivalent implementations.
Embodiments of the present disclosure may be practiced with various computer system configurations including hand-held devices, microprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers and the like. Embodiments of the present disclosure can also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a wire-based or wireless network.
With the above embodiments in mind, it should be understood that embodiments of the present disclosure can employ various computer-implemented operations involving data stored in computer systems. These operations are those requiring physical manipulation of physical quantities. Any of the operations described herein that form part of embodiments of the present disclosure are useful machine operations. Embodiments of the disclosure also relate to a device or an apparatus for performing these operations. The apparatus can be specially constructed for the required purpose, or the apparatus can be a general-purpose computer selectively activated or configured by a computer program stored in the computer. In particular, various general-purpose machines can be used with computer programs written in accordance with the teachings herein, or it may be more convenient to construct a more specialized apparatus to perform the required operations.
The disclosure can also be embodied as computer readable code on a computer readable medium. The computer readable medium is any data storage device that can store data, which can be thereafter be read by a computer system. Examples of the computer readable medium include hard drives, network attached storage (NAS), read-only memory, random-access memory, CD-ROMs, CD-Rs, CD-RWs, magnetic tapes and other optical and non-optical data storage devices. The computer readable medium can include computer readable tangible medium distributed over a network-coupled computer system so that the computer readable code is stored and executed in a distributed fashion.
Although the method operations were described in a specific order, it should be understood that other housekeeping operations may be performed in between operations, or operations may be adjusted so that they occur at slightly different times, or may be distributed in a system which allows the occurrence of the processing operations at various intervals associated with the processing, as long as the processing of the overlay operations are performed in the desired way.
Although the foregoing disclosure has been described in some detail for purposes of clarity of understanding, it will be apparent that certain changes and modifications can be practiced within the scope of the appended claims. Accordingly, the present embodiments are to be considered as illustrative and not restrictive, and embodiments of the present disclosure is not to be limited to the details given herein, but may be modified within the scope and equivalents of the appended claims.
Claims
- A method for processing an artificial intelligence (AI) model for a gaming application, comprising: training the AI model from a plurality of game plays of a scenario of the gaming application using training state data collected from the plurality of game plays of the scenario and associated success criteria of each of the plurality of game plays;receiving first input state data during a first game play of the scenario;applying the first input state data to the AI model to generate an output indicating a degree of success for the scenario for the first game play;performing an analysis of the output based on a predefined objective;and performing an action to achieve the predefined objective based on the output that is analyzed.
- The method of claim 1 , wherein the training an AI model includes: executing at a plurality of servers a plurality of instances of the gaming application supporting the plurality of game plays;collecting at the plurality of servers the training state data that is associated with corresponding game plays of the scenario;defining the success criteria;and providing the training state data and the success criteria to a deep learning engine.
- The method of claim 2 , wherein the plurality of game plays is controlled by a plurality of players via a plurality of client devices.
- The method of claim 2 , wherein the performing an action includes: wherein the predefined objective is to automatically train the AI model, wherein the plurality of game plays is controlled by the AI model during training, wherein the first input state data is generated by the AI model as a previous action taken during training, and determining as the analysis a plurality of sets of controller inputs for the first game play to be played subsequent to the first input state data;selecting a first set of controller inputs predicted to satisfy the success criteria and predicted to have a highest degree of success among the plurality of sets of controller inputs;and delivering as the action the first set of controller inputs to a corresponding instance of the gaming application for execution.
- The method of claim 2 , wherein the performing an action includes: wherein the predefined objective is to automatically test the gaming application, wherein the first input state data is generated by the AI model as a previous action during training, determining as the analysis different permutations for responding to the first input state data, each permutation including a unique set of controller inputs to be taken;and executing as the action the different permutations to discover any anomaly in playing the gaming application.
- The method of claim 2 , wherein the plurality of game plays of the gaming application is controlled by a first player through one or more client devices, wherein the trained AI model is a virtual player that simulates the first player.
- The method of claim 1 , further comprising: wherein the predefined objective is to provide gaming support, determining the first input state data does not satisfy the success criteria based on the output that is analyzed;determining a first user profile for a first player controlling the first game play based on the output that is analyzed and the first input state data;and providing as the action a recommendation on how to play the scenario to a client device of the first player, the recommendation reflecting a skill of the first player based on the first user profile.
- The method of claim 1 , wherein performing an analysis of the output includes: wherein the predefined objective is to provide coaching, identifying a weakness of a first player controlling the first game play by determining that the first input state data has a lower than average degree of success;and executing as the action a training session corresponding to the weakness for delivery to a client device of the first player.
- The method of claim 1 , further comprising: wherein the predefined objective is to provide parity in game plays, receiving a plurality of input state data during a second plurality of game plays of the scenario conducted after training the AI model, the second plurality of game plays being controlled by a plurality of players, the plurality of input state data including a plurality of player characteristic metrics of the plurality of players;applying the plurality of input state data to the AI model to generate a plurality of outputs indicating a plurality of degrees of success for the second plurality of game plays of the scenario;analyzing the plurality of outputs to determine a plurality of player profiles based on the plurality of player characteristic metrics;and building balanced teams of players from the plurality of players based on corresponding player profiles to achieve the predefined objective.
- The method of claim 9 , wherein a player characteristic metric includes: accuracy of a corresponding player;or speed of generating a sequence of controller inputs by the corresponding player;or reaction time of the corresponding player when responding to an event in the scenario;or consistency of the corresponding player;or transition time of the corresponding player between a first target and a second target.
- A non-transitory computer-readable medium storing a computer program for artificial intelligence (AI) training, said computer-readable medium comprising: program instructions for training the AI model from a plurality of game plays of a scenario of the gaming application using training state data collected from the plurality of game plays of the scenario and associated success criteria of each of the plurality of game plays;program instructions for receiving first input state data during a first game play of the scenario;program instructions for applying the first input state data to the AI model to generate an output indicating a degree of success for the scenario for the first game play;program instructions for performing an analysis of the output based on a predefined objective;and program instructions for performing an action to achieve the predefined objective based on the output that is analyzed.
- The non-transitory computer-readable medium of claim 11 , wherein the program instructions for training an AI model includes: program instructions for executing at a plurality of servers a plurality of instances of the gaming application supporting the plurality of game plays;program instructions for collecting at the plurality of servers the training state data that is associated with corresponding game plays of the scenario;program instructions for defining the success criteria;and program instructions for providing the training state data and the success criteria to a deep learning engine.
- The non-transitory computer-readable medium of claim 12 , wherein the plurality of game plays is controlled by a plurality of players via a plurality of client devices.
- The non-transitory computer-readable medium of claim 11 , further comprising: wherein the predefined objective is to provide gaming support, program instructions for determining the first input state data does not satisfy the success criteria based on the output that is analyzed;program instructions for determining a first user profile for a first player controlling the first game play based on the output that is analyzed and the first input state data;and program instructions for providing as the action a recommendation on how to play the scenario to a client device of the first player, the recommendation reflecting a skill of the first player based on the first user profile.
- The non-transitory computer-readable medium of claim 11 , wherein program instructions for performing an analysis of the output includes: wherein the predefined objective is to provide coaching, program instructions for identifying a weakness of a first player controlling the first game play by determining that the first input state data has a lower than average degree of success;and program instructions for executing as the action a training session corresponding to the weakness for delivery to a client device of the first player.
- A computer system comprising: a processor;and memory coupled to the processor and having stored therein instructions that, if executed by the computer system, cause the computer system to execute a method for artificial intelligence (AI) training comprising: training the AI model from a plurality of game plays of a scenario of the gaming application using training state data collected from the plurality of game plays of the scenario and associated success criteria of each of the plurality of game plays;receiving first input state data during a first game play of the scenario;applying the first input state data to the AI model to generate an output indicating a degree of success for the scenario for the first game play;performing an analysis of the output based on a predefined objective;and performing an action to achieve the predefined objective based on the output that is analyzed.
- The computer system of claim 16 , wherein in the method the training an AI model includes: executing at a plurality of servers a plurality of instances of the gaming application supporting the plurality of game plays;collecting at the plurality of servers the training state data that is associated with corresponding game plays of the scenario;defining the success criteria;and providing the training state data and the success criteria to a deep learning engine.
- The computer system of claim 17 , wherein in the method the plurality of game plays is controlled by a plurality of players via a plurality of client devices.
- The computer system of claim 16 , the method further comprising: wherein the predefined objective is to provide gaming support, determining the first input state data does not satisfy the success criteria based on the output that is analyzed;determining a first user profile for a first player controlling the first game play based on the output that is analyzed and the first input state data;and providing as the action a recommendation on how to play the scenario to a client device of the first player, the recommendation reflecting a skill of the first player based on the first user profile.
- The computer system of claim 16 , wherein in the method performing an analysis of the output includes: wherein the predefined objective is to provide coaching, identifying a weakness of a first player controlling the first game play by determining that the first input state data has a lower than average degree of success;and executing as the action a training session corresponding to the weakness for delivery to a client device of the first player.
Disclaimer: Data collected from the USPTO and may be malformed, incomplete, and/or otherwise inaccurate.