U.S. Pat. No. 7,090,576
PERSONALIZED BEHAVIOR OF COMPUTER CONTROLLED AVATARS IN A VIRTUAL REALITY ENVIRONMENT
AssigneeMicrosoft Corporation
Issue DateJune 30, 2003
U.S. Patent No. 7,090,576: Personalized behavior of computer controlled avatars in a virtual reality environment
Summary:
The ‘576 patent describes a racing video game which offers enhanced computer AI to give the computer-controlled cars a more “human” feel to them. The game records a player’s movements during training and then applies samplings from this behavior to the computer-controlled cars during game play. Thus, instead of acting solely as a computer (with quicker, more perfect reaction times than a human), the computer-controlled cars act with the same reaction times and in the same manner as the human players. This invention allows a player to enjoy playing the computer as much as he does when he plays another human.
Abstract:
Racing-based computer games typically include a mode in which one or more human players can compete against one or more computer-controlled opponents. For example, a human player may drive a virtual race car against a computer-controlled virtual race car purported to be driven by Mario Andretti or some other race car driver. Such computer controlled opponents may be enhanced by including a sampling of actual game behavior of a human subject into the opponent’s artificial intelligence control system. Such a sampling can allow the game system to personalize the behavior of the computer control opponent to emulate the human subject.
Illustrative Claim:
1. A method comprising: computing an avatar behavior definition based on environmental context of a virtual reality environment and a randomly selected training behavior from a training set of personalized sample behaviors; and generating at least one control signal to guide behavior of an entity in the virtual reality environment in accordance with the avatar behavior definition.
Illustrative Figure
Abstract
Racing-based computer games typically include a mode in which one or more human players can compete against one or more computer-controlled opponents. For example, a human player may drive a virtual race car against a computer-controlled virtual race car purported to be driven by Mario Andretti or some other race car driver. Such computer controlled opponents may be enhanced by including a sampling of actual game behavior of a human subject into the opponent's artificial intelligence control system. Such a sampling can allow the game system to personalize the behavior of the computer control opponent to emulate the human subject.
Description
Racing-based computer games typically include a mode in which one or more human players can compete against one or more computer-controlled opponents. For example, a human player may drive a virtual race car against a computer-controlled virtual race car purported to be driven by Mario Andretti or some other race car driver. As described herein, such computer controlled opponents may be enhanced by including a sampling of actual game behavior of a human subject. Such a sampling can allow the game system to personalize the behavior of the computer control opponent to emulate the human subject. In this manner, the computer controlled opponent may be personalized to behave like famous competitors in the field or like friends and family of the human player. For example, two friends are playing a computer racing game before one of the friends needs leave. The departing friend can enter a training mode to create personalized training data storing it in a persistent storage medium accessible by the gaming system. The remaining friend can then choose to play an avatar that is personalized by that training data to emulate his or her departed friend. Likewise, training data may be generated and/or obtained for more famous participants (e.g., Mario Andretti) and then stored in a persistent storage medium accessible by the gaming system. FIG. 1illustrates a race course and two training racing lines for use in parameterizing an exemplary computer controlled avatar. A virtual race course100includes a racing surface102and an infield104. It should be understood that an alternative course may not be closed or may be applied to a non-racing environment, such as a first person shooter game, a military simulation, or a reality simulation. As such, a computer controlled driver may be replaced by a computer controlled (or AI controlled) player represented by any moving ...
Racing-based computer games typically include a mode in which one or more human players can compete against one or more computer-controlled opponents. For example, a human player may drive a virtual race car against a computer-controlled virtual race car purported to be driven by Mario Andretti or some other race car driver. As described herein, such computer controlled opponents may be enhanced by including a sampling of actual game behavior of a human subject. Such a sampling can allow the game system to personalize the behavior of the computer control opponent to emulate the human subject.
In this manner, the computer controlled opponent may be personalized to behave like famous competitors in the field or like friends and family of the human player. For example, two friends are playing a computer racing game before one of the friends needs leave. The departing friend can enter a training mode to create personalized training data storing it in a persistent storage medium accessible by the gaming system. The remaining friend can then choose to play an avatar that is personalized by that training data to emulate his or her departed friend. Likewise, training data may be generated and/or obtained for more famous participants (e.g., Mario Andretti) and then stored in a persistent storage medium accessible by the gaming system.
FIG. 1illustrates a race course and two training racing lines for use in parameterizing an exemplary computer controlled avatar. A virtual race course100includes a racing surface102and an infield104. It should be understood that an alternative course may not be closed or may be applied to a non-racing environment, such as a first person shooter game, a military simulation, or a reality simulation. As such, a computer controlled driver may be replaced by a computer controlled (or AI controlled) player represented by any moving object.
Two exemplary training racing lines106and108are shown within the virtual race course100. Notice that, in one lap, the human driver hit the wall at109. Such imperfections contribute to the experience of competing against a human competitor. Racing lines, examples of behavior definitions, may be characterized by various parameters, including without limitation locations on the course, and steering, braking, speed, and acceleration parameter sequences. Such training racing lines may be captured, for example, by monitoring and recording a human driver's performance during multiple laps of a race.
The training racing lines106and108are used to develop personalized training data specific to the human subject that recorded them. Using this personalized training data, the system controls the AI controlled avatar to emulate the human subject's performance. Multiple racing lines provide additional variation that contributes to the unpredictable behavior of the AI controlled avatar.
During a race, at a specified interval (such as a simulation interval or a frame interval, which may be equal in some implementations), an AI motion control system receives stimulus inputs and computes control signals based on these inputs to change the direction, speed, etc. of the computer racing vehicle or other entity. The changes are generally intended to bring the AI driver back into alignment with a racing line (such as an ideal racing line or a personalized racing line). As such, a racing line is a form of a behavior definition that provides guidance to the AI motion control system in controlling the AI driver's path along the track.
A course is divided up into track segments (see, for example, track segment110). Generally, track segments are a type of game segment, which may also represent a game turn in a strategy game, a scene in a role playing game, a level in a first person shooter, etc., or combinations thereof.
Each racing line is characterized relative to waypoints (such as waypoints112,114,116, and118) within a track segment. Track segments are examples of a more generic game segment, such as a distance through a maze, one or more rooms in a virtual reality setting, etc. In one implementation, for example, at waypoint118, a steering angle120between a normal vector (normal to the waypoint line across the track) and the sampled racing line characterizes the racing line to the system. Waypoints and the steering angles are then used by the AI motion control system to compute proper motion control signals for the AI controlled entity (e.g., the racing vehicle).
FIG. 2illustrates an exemplary system200for generating a personalized training data. A set of personalized training samples202are generated and input to the training module204. In one implementation, personalized training samples15are generated by monitoring performance of a human subject during game play. Characteristics of the environment (e.g., track geometry, track conditions, etc) and subject's behavior (e.g., a racing line, control signals, etc.) are parameterized and recorded in personalized training data206. In at least one implementation, the personalized training data may be persistently stored to allow the human player to save training data profiles of various avatar competitors.
FIG. 3illustrates exemplary personalized training data300including personalized sample behaviors. Each row corresponds to a particular game segment, as represented by the track segment pictorials along the leftmost column. The first column of each row represents a given game segment in the course. Each game segment may be represented by game segment descriptors, which may include without limitation track segment geometry, track surface conditions, time of day, etc. Each subsequent column corresponds to a sample of the human subject's behavior (i.e., personalized sample behaviors) in the associated game segment.
In addition, the rate of accumulating personalized training data may effectively be increased by “mirroring” training data for a given track segment, thereby doubling the amount of training data extracted for a given amount of training. For example, for some track segments, by extracting training data for a right hand turn and then mirroring both the track segment (to represent a left hand turn) and the personalized racing line sample, two training samples are captured for a single track segment.
In the illustrated table300, the sample of the human subject's behavior is represented by a racing line for that track segment. As shown in the last row, for example, the human subject provided four sample behavior definitions (e.g., racing lines) for the associated track segment, which appears to be a sharp curve. Differences among the behavior definitions taken by the human subject can be seen in each cell in that row, thereby providing the opportunity to contribute to the unpredictable behavior of the AI avatar on this track segment or a track segments that are similar to this track segment.
In one implementation of a racing game, a racing line may be defined by a sequence of steering angle values at waypoints in a track segment. A steering angle value characterizes the steering angle of the racing line relative to a vector that is normal to the waypoint lines in the track segment. It should be understood that racing lines may be characterized by other parameters or combinations of other parameters, including without limitation sampled locations or vectors within the course, game play actions (e.g., fight sequences), etc. In an implementation of a first person shooter game, a behavior definition may represent a player's movement path.
The track segments included in the training set may or may not be from the current track. In one implementation, a probability density function operation determines similarities between training track segments and the current track segment and computes weighted contributions by sampling racing lines associated with each similar training track segment. The weighting may apply contributions of one or more sample racing lines to the AI avatar's current racing line within the current track segment. In an alternative implementation, personalized avatars may be executed only on tracks on which the human subject has actually trained.
FIG. 4illustrates an exemplary system400for generating control signals for a computer controlled avatar. Personalized training data402may be provided in a variety of ways, including dynamic generation of the training data based on past or concurrent game play by the human subject, downloading training data from a web resource or storage media, computing training data for fictitious players, etc. Note: With the concurrent game play option, a player could compete against an increasingly improving avatar of himself or herself by starting with an initial set of default or personalized training data and adding to the training set as the player plays the game. As the game play proceeds, the racing lines taken by the user during the game may be added to the training data to enhance the richness of the training samples.
Environmental context information404relating to the game segment is provided by the virtual reality environment, such as track location, track segment geometry, etc. The context information parameters, which are similar in format and/or content to the track segment descriptors, are used to match one or more track segments in the training data402, so as to identify personalized sample behaviors that contribute to the behavior definition of the AI controlled avatar.
The training data402and the context information404are received by an avatar behavior definition module406, which computes a behavior definition408. In one implementation, the behavior definition408specifies a personalized racing line to be followed by the AI behavior control system410. However, other behavior definitions may also be specified, such as an AI avatar's behavior in a first person shooter game, etc.
In one implementation, an avatar behavior definition module406randomly selects a sample index (e.g., to select a column of sample data) and evaluates the segment descriptors against the context information404to identify appropriate sample data. The sample behavior for the best matching track segment under the randomly selected index may be forwarded as the behavior definition408to the AI behavior control system410.
In another implementation, a probability density function module412determines the weighted contribution that one or more samples provides to a behavior definition for a given track segment. The higher the similarity between the context information404and the track segment descriptor, the greater the weight applied to the sample data when computing the behavior definition. The weighted sum of these contributions is then forwarded as the behavior definition408to the AI behavior control system410. Other contribution algorithms are also contemplated.
Stimuli414from the virtual reality environment is input to the AI behavior control system410, which generates control signals416for controlling the avatar in the virtual reality environment. The control signals416are input to an implementation engine418to implement the motion or behavior specified by the control signals416. For example, the implementation engine418may cause a virtual race car to change direction and speed.
FIG. 5illustrates exemplary operations500for generating control signals for a computer controlled avatar. A receiving operation502receives personalized training data. Another receiving operation504receives context information for the track segments of the current track.
A generation operation506generates an avatar behavior definition for each track segment of the current track. In one implementation, a sample index is randomly selected and the sample data corresponding to the track segment having the closest match with context information504of the current track is deemed the avatar behavior definition for the current track segment. This selection and evaluation to identify the avatar behavior definition for a current track segment is repeated for each track segment in the current track.
In another implementation, sample data from multiple track segments in the training data may contribute to the avatar behavior definition for a given track segment of the current track. Assume that BDirepresents the behavior definition for a given track segment i in the current track and SDj|kjrepresents the sample data from row j, column kjof the training data. An exemplary algorithm is described below:
1For all track segments i in current track2For all track segments j in the training data3kj=randomly selected sample index over all samplecolumns in the training data4sj=ⅇ-τi-τj22ω2,whereinsjrepresentsaweightingfactorfor the jthtrack segment in the training data, τirepresents the context information for the ithtrack segment in the current track, τjrepresentsthe track descriptor of the jthtrack segment inthe training data, and ω represents a sensitivityfactor5End6Normalizesj(e.g.,sj′=sj∑jsi)7BDi=∑jsji·SDj❘kj8End
The sensitivity factor ω specifies how sensitive the weighting factor is to similarities. For example, if ω is very small, then sjis small and sample data for fewer training track segments contribute heavily to the behavior definition BDi. In contrast, if ω is very large, then sj large and sample data for more training track segments contribute heavily to the behavior definition BDi. The sensitivity factor ω may be preset at game design time or configured by the player or the system.
In one implementation, avatar behavior definitions for all track segments of the current track are computed before race time (e.g., to conserve computational capacity for other functions during race time). However, it should also be understood that, in an alternative implementation, avatar behavior definitions for individual track segments in the current track may be dynamically computed as the track segment is encountered in the virtual reality environment. This approach allows race time events to be considered when determining matches with training data. In addition, a human player's behavior during the current race may be used to dynamically supplement the training data during the race (e.g., to allow the human player to compete against a continually changing avatar of himself or herself).
A receiving operation508receives game stimuli from the virtual reality environment for the current simulation interval. Such stimuli is input to a generation operation510, which generates the control signals used to converge avatar behavior to the avatar behavior definition (e.g., generated in the generation operation506) based on the stimuli. For example, the control signals in a racing game are generally intended to steer the AI control vehicle toward a racing line.
An implementation operation512applies the control signals to the virtual reality environment through an implementation module. For example, exemplary implementation modules may include a racing physics engine. Operations508,510, and512repeat each simulation interval to progress the action in the virtual reality game.
The exemplary hardware and operating environment ofFIG. 6for implementing the invention includes a general purpose computing device in the form of a computer20, including a processing unit21, a system memory22, and a system bus23that operatively couples various system components include the system memory to the processing unit21. There may be only one or there may be more than one processing unit21, such that the processor of computer20comprises a single central-processing unit (CPU), or a plurality of processing units, commonly referred to as a parallel processing environment. The computer may be a conventional computer, a distributed computer, or any other type of computer; the invention is not so limited.
The system bus23may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures. The system memory may also be referred to as simply the memory, and includes read only memory (ROM)24and random access memory (RAM)25. A basic input/output system (BIOS)26, containing the basic routines that help to transfer information between elements within the computer20, such as during start-up, is stored in ROM24. The computer20further includes a hard disk drive27for reading from and writing to a hard disk, not shown, a magnetic disk drive28for reading from or writing to a removable magnetic disk29, and an optical disk drive30for reading from or writing to a removable optical disk31such as a CD ROM or other optical media.
The hard disk drive27, magnetic disk drive28, and optical disk drive30are connected to the system bus23by a hard disk drive interface32, a magnetic disk drive interface33, and an optical disk drive interface34, respectively. The drives and their associated computer-readable media provide nonvolatile storage of computer-readable instructions, data structures, program modules and other data for the computer20. It should be appreciated by those skilled in the art that any type of computer-readable media which can store data that is accessible by a computer, such as magnetic cassettes, flash memory cards, digital video disks, Bernoulli cartridges, random access memories (RAMs), read only memories (ROMs), and the like, may be used in the exemplary operating environment.
A number of program modules may be stored on the hard disk, magnetic disk29, optical disk31, ROM24, or RAM25, including an operating system35, one or more application programs36, other program modules37, and program data38. A user may enter commands and information into the personal computer20through input devices such as a keyboard40and pointing device42. Other input devices (not shown) may include a microphone, joystick, game pad, satellite dish, scanner, or the like. These and other input devices are often connected to the processing unit21through a serial port interface46that is coupled to the system bus, but may be connected by other interfaces, such as a parallel port, game port, or a universal serial bus (USB). A monitor47or other type of display device is also connected to the system bus23via an interface, such as a video adapter48. In addition to the monitor, computers typically include other peripheral output devices (not shown), such as speakers and printers.
The computer20may operate in a networked environment using logical connections to one or more remote computers, such as remote computer49. These logical connections are achieved by a communication device coupled to or a part of the computer20; the invention is not limited to a particular type of communications device. The remote computer49may be another computer, a server, a router, a network PC, a client, a peer device or other common network node, and typically includes many or all of the elements described above relative to the computer20, although only a memory storage device50has been illustrated inFIG. 6. The logical connections depicted inFIG. 6include a local-area network (LAN)51and a wide-area network (WAN)52. Such networking environments are commonplace in office networks, enterprise-wide computer networks, intranets and the Internal, which are all types of networks.
When used in a LAN-networking environment, the computer20is connected to the local network51through a network interface or adapter53, which is one type of communications device. When used in a WAN-networking environment, the computer20typically includes a modem54, a type of communications device, or any other type of communications device for establishing communications over the wide area network52. The modem54, which may be internal or external, is connected to the system bus23via the serial port interface46. In a networked environment, program modules depicted relative to the personal computer20, or portions thereof, may be stored in the remote memory storage device. It is appreciated that the network connections shown are exemplary and other means of and communications devices for establishing a communications link between the computers may be used.
In an exemplary implementation, AI motion control systems, physics engines, training modules, avatar control modules, avatar behavior definition s modules, AI behavior control modules, or other modules may be incorporated as part of the operating system35, application programs36, or other program modules37. The stimuli signals, behavior definitions, training data, context information, and control signals may be stored as program data38.
The embodiments of the invention described herein are implemented as logical steps in one or more computer systems. The logical operations of the present invention are implemented (1) as a sequence of processor-implemented steps executing in one or more computer systems and (2) as interconnected machine modules within one or more computer systems. The implementation is a matter of choice, dependent on the performance requirements of the computer system implementing the invention. Accordingly, the logical operations making up the embodiments of the invention described herein are referred to variously as operations, steps, objects, or modules.
The above specification, examples and data provide a complete description of the structure and use of exemplary embodiments of the invention. Since many embodiments of the invention can be made without departing from the spirit and scope of the invention, the invention resides in the claims hereinafter appended.
Claims
- A method comprising: computing an avatar behavior definition based on environmental context of a virtual reality environment and a randomly selected training behavior from a training set of personalized sample behaviors;and generating at least one control signal to guide behavior of an entity in the virtual reality environment in accordance with the avatar behavior definition.
- The method of claim 1 wherein the avatar behavior definition is associated with a track segment.
- The method of claim 1 further comprising: generating the personalized sample behaviors by recording behavior of a player in one or more training sessions in the virtual reality environment.
- The method of claim 1 wherein the computing operation comprises: combining a plurality of the personalized sample behaviors for a current game segment to yield the avatar behavior definition for the current game segment.
- The method of claim 1 wherein the computing operation comprises: combining weighted contributions of a plurality of the personalized sample behaviors for a current game segment to yield the avatar behavior definition for the current game segment.
- The method of claim 1 wherein the computing operation comprises: randomly selecting a personalized sample behavior from the training set using a randomly selected sample index that designates at least one of the personalized sample behaviors.
- The method of claim 1 further comprising: storing the personalized sample behaviors in a persistent storage medium.
- The method of claim 1 further comprising: recording behavior of a player in competition against the entity during game play in the virtual reality environment;and adding the recorded behavior to the personalized sample behaviors associated with the player during the game play.
- The method of claim 1 wherein the computing operation comprises: identifying game segments in the training set that match the current game segment by evaluating game segment descriptors associated with personalized sample behaviors against environmental context information associated with the current game segment.
- A computer program product encoding a computer program for executing on a computer system a computer process, the computer process comprising: computing an avatar behavior definition based on environmental context of a virtual reality environment and a randomly selected training behavior from a training set of personalized sample behaviors;and generating at least one control signal to guide behavior of an entity in the virtual reality environment in accordance with the avatar behavior definition.
- The computer program product of claim 10 wherein the avatar behavior definition is associated with a game segment.
- The computer program product of claim 10 wherein the computer process further comprises: generating the personalized sample behaviors by recording behavior of a player in one or more training sessions in the virtual reality environment.
- The computer program product of claim 10 wherein the computing operation comprises: combining a plurality of the personalized sample behaviors for a current game segment to yield the avatar behavior definition for the current game segment.
- The computer program product of claim 10 wherein the computing operation comprises: combining weighted contributions of a plurality of the personalized sample behaviors for a current game segment to yield the avatar behavior definition for the current game segment.
- The computer program product of claim 10 wherein the computing operation comprises: randomly selecting a personalized sample behavior from the training set using a randomly selected sample index that designates at least one of the personalized sample behaviors.
- The computer program product of claim 10 wherein the computer process further comprises: storing the personalized sample behaviors in a persistent storage medium.
- The computer program product of claim 10 wherein the computer process further comprises: recording behavior of a player in competition against the entity during game play in the virtual reality environment;and adding the recorded behavior to the personalized sample behaviors associated with the player during the game play.
- The computer program product of claim 10 wherein the computing operation comprises: identifying game segments in the training set that match the current game segment by evaluating game segment descriptors associated with personalized sample behaviors against environmental context information associated with the current game segment.
- A system comprising: an avatar behavior definition module computing an avatar behavior definition based on environmental context of a virtual reality environment and a randomly selected training behavior from a training set of personalized sample behaviors;and a behavior control system generating at least one control signal to guide behavior of an entity in the virtual reality environment in accordance with the avatar behavior definition.
- The system of claim 19 wherein the avatar behavior definition is associated with a game segment.
- The system of claim 19 further comprising: a training module generating the personalized sample behaviors by recording behavior of a player in one or more training sessions in the virtual reality environment.
- The system of claim 19 wherein the computing operation comprises: a probability density function module combining a plurality of the personalized sample behaviors for a current game segment to yield the avatar behavior definition for the current segment.
- The system of claim 19 wherein the computing operation comprises: a probability density function module combining weighted contributions of a plurality of the personalized sample behaviors for a current game segment to yield the avatar behavior definition for the current segment.
- The system of claim 19 wherein the computing operation comprises: a probability density function module randomly selecting a personalized sample behavior from the training set using a randomly selected sample index that designates at least one of the personalized sample behaviors.
- The system of claim 19 further comprising: a training module storing the personalized sample behaviors in a persistent storage medium.
- The system of claim 19 further comprising: a training module recording behavior of a player in competition against the entity during game play in the virtual reality environment and adding the recorded behavior to the personalized sample behaviors associated with the player during game play.
- The system of claim 19 wherein the computing operation comprises: a probability density function module identifying game segments in the training set that match the current game segment by evaluating game segment descriptors associated with personalized sample behaviors against environmental context information associated with the current game segment.
Disclaimer: Data collected from the USPTO and may be malformed, incomplete, and/or otherwise inaccurate.
