U.S. Pat. No. 9,550,121

System and Method for Enabling Characters to be Manifested Within A Plurality of Different Virtual Spaces

AssigneeDisney Enterprises, Inc.

Issue DateNovember 17, 2011

Illustrative Figure

Abstract

A system and method for providing virtual spaces, where a character associated with a user can be manifested within instances of a plurality of the different virtual spaces. Since a single character can be manifested within instances of different virtual spaces, the character can be transferred by the corresponding user between instances of different virtual spaces and controlled by the user to interact with the different virtual spaces. When the user transfers the character between instances of different virtual spaces (and/or different types of virtual spaces), various aspects of the character may persist between the different virtual spaces (and/or the different types of virtual spaces). This may provide an enhanced continuity to the character between the different virtual spaces.

Description

DETAILED DESCRIPTION FIG. 1illustrates a system10configured to provide one or more virtual spaces that may be accessible to users. In some embodiments, system10may include a storage module12, a server14, a client16, and/or other components. Storage module12, server14, and client16may be in operative communication with each other. System10may be configured such that information related to a given virtual space may be transmitted from storage module12to server14, which may then execute an instance of the virtual space. Views of the virtual space may be generated by server14from the instance of the virtual space. Information related to the views may be transmitted from server14to client16to enable client16to format the views for display to a user. In some embodiments, system10may enable a user to control a character within an instance of a virtual space. The character may be manifested within the instance of the virtual space as an object, such as, for example, an avatar, through which the user may interact with the instance of the virtual space. For example, the character may be controlled by the user to move about within the instance, interact with one or more other objects within the instance, communicate with other objects (e.g., characters associated with other users), fight other objects, race other objects, observe and/or interact with topography within the instance of the virtual space, and/or otherwise interact with the virtual space. System10may further be configured such that the character can be transferred by the user between instances of different virtual spaces (including different types of virtual spaces). When the user transfers the character between instances of different virtual spaces (and/or different types of virtual spaces), various aspects of the character may persist between the different virtual spaces (and/or the different types of virtual spaces). System10may implement a markup language for communication between components (e.g., storage module12, ...

DETAILED DESCRIPTION

FIG. 1illustrates a system10configured to provide one or more virtual spaces that may be accessible to users. In some embodiments, system10may include a storage module12, a server14, a client16, and/or other components. Storage module12, server14, and client16may be in operative communication with each other. System10may be configured such that information related to a given virtual space may be transmitted from storage module12to server14, which may then execute an instance of the virtual space. Views of the virtual space may be generated by server14from the instance of the virtual space. Information related to the views may be transmitted from server14to client16to enable client16to format the views for display to a user.

In some embodiments, system10may enable a user to control a character within an instance of a virtual space. The character may be manifested within the instance of the virtual space as an object, such as, for example, an avatar, through which the user may interact with the instance of the virtual space. For example, the character may be controlled by the user to move about within the instance, interact with one or more other objects within the instance, communicate with other objects (e.g., characters associated with other users), fight other objects, race other objects, observe and/or interact with topography within the instance of the virtual space, and/or otherwise interact with the virtual space. System10may further be configured such that the character can be transferred by the user between instances of different virtual spaces (including different types of virtual spaces). When the user transfers the character between instances of different virtual spaces (and/or different types of virtual spaces), various aspects of the character may persist between the different virtual spaces (and/or the different types of virtual spaces).

System10may implement a markup language for communication between components (e.g., storage module12, server14, client16, etc.). Information may be communicated between components via markup elements of the markup language. By virtue of communication between the components of system10in the markup language, various enhancements may be achieved. For example, information may be transmitted from storage module12to server14that configures server14to execute an instance of the virtual space may be provided to server14via the markup language at or near the time of instantiation. Similarly, information transmitted from server14to client16may enable client16to generate views of the virtual space by merely assembling the information indicated in markup elements communicated thereto. The implementation of the markup language may facilitate the execution of instances of various types of virtual spaces by system10. The types of virtual spaces may include, for example, two-dimensional spaces, three-dimensional spaces, gaming spaces, social spaces, first-person spaces, third-person spaces, spaces that draw from different genres, action gaming spaces, role-playing spaces, and/or other types of spaces. The implementation of the markup language may facilitate creation of a new virtual space by the user of client16, and/or the customization/refinement of existing virtual spaces.

As used herein, a virtual space may comprise a simulated space (e.g., a physical space) instanced on a server (e.g., server14) that is accessible by a client (e.g., client16) located remotely from the server, to format a view of the virtual space for display to a user of the client. The simulated space may have a topography, express real-time interaction by the user, and/or include one or more objects positioned within the topography that are capable of locomotion within the topography. In some implementations, the topography may be a 2-dimensional topography. In other instances, the topography may be a 3-dimensional topography. In some implementations, the topography may be a single node. The topography may include dimensions of the virtual space, and/or surface features of a surface or objects that are “native” to the virtual space. In some implementations, the topography may describe a surface (e.g., a ground surface) that runs through at least a substantial portion of the virtual space. In some implementations, the topography may describe a volume with one or more bodies positioned therein (e.g., a simulation of gravity-deprived space with one or more celestial bodies positioned therein). A virtual space may include a virtual world, but this is not necessarily the case. For example, a virtual space may include a game space that does not include one or more of the aspects generally associated with a virtual world (e.g., gravity, a landscape, etc.). By way of illustration, the well-known game Tetris may be formed as a two-dimensional topography in which bodies (e.g., the falling tetrominoes) move in accordance with predetermined parameters (e.g., falling at a predetermined speed, and shifting horizontally and/or rotating based on user interaction).

As used herein, the term “markup language” may include a language used to communicate information between components via markup elements. Generally, a markup element is a discrete unit of information that includes both content and attributes associated with the content. The markup language may include a plurality of different types of elements that denote the type of content and the nature of the attributes to be included in the element. For example, in some embodiments, the markup elements in the markup language may be of the form [O_HERE]|objectId|artIndex|y|z|name|templateId. This may represent a markup element for identifying a new object in a virtual space. The parameters for the mark-up element include: assigning an object Id for future reference for this object, telling the client what art to draw associated with this object, the relative x, y, and z position of the object, the name of the object, and data associated with the object (comes from the template designated). As another non-limiting example, a mark-up element may be of the form [O_GONE]|objId. This mark-up element may represent an object going away from the perspective of a view of the virtual space. As yet another example, a mark-up element may be of the form [O_MOVE]|objectId|x|y|z. This mark-up element may represent an object that has teleported to a new location in the virtual space. As still another example, a mark-up element may be of the form [O_SLIDE]|objectId|x|y|z|time. This mark-up element may represent an object that is gradually moving from one location in the virtual space to a new location over a fixed period of time. It should be appreciated that these examples are not intended to be limiting, but only to illustrate a few different forms of the markup elements.

Storage module12may include information storage18, a server communication module20, and/or other components. Generally, storage module12may store information related to one or more virtual spaces. The information stored by storage module12that is related to a given virtual space may include topographical information related to the topography of the given virtual space, manifestation information related to the manifestation of one or more objects positioned within the topography and/or unseen forces experienced by the one or more objects in the virtual space, interface information related to an interface provided to the user that enables the user to interact with the virtual space, space parameter information related to parameters of the virtual space, and/or other information related to the given virtual space.

The manifestation of the one or more objects may include the locomotion characteristics of the one or more objects, the size of the one or more objects, the identity and/or nature of the one or more objects, interaction characteristics of the one or more objects, and/or other aspect of the manifestation of the one or more objects. The interaction characteristics of the one or more objects described by the manifestation information may include information related to the manner in which individual objects interact with and/or are influenced by other objects, the manner in which individual objects interact with and/or are influenced by the topography (e.g., features of the topography), the manner in which individual objects interact with and/or are influenced by unseen forces within the virtual space, and/or other characteristics of the interaction between individual objects and other forces and/or objects within the virtual space. The interaction characteristics of the one or more objects described by the manifestation information may include scriptable behaviors and, as such, the manifestation stored within storage module12may include one or both of a script and a trigger associated with a given scriptable behavior of a given object (or objects) within the virtual space. The unseen forces present within the virtual space may include one or more of gravity, a wind current, a water current, an unseen force emanating from one of the objects (e.g., as a “power” of the object), and/or other unseen forces (e.g., unseen influences associated with the environment of the virtual space such as temperature and/or air quality).

In some embodiments, the manifestation information may include information related to the sonic characteristics of the one or more objects positioned in the virtual space. The sonic characteristics may include the emission characteristics of individual objects (e.g., controlling the emission of sound from the objects), the acoustic characteristics of individual objects, the influence of sound on individual objects, and/or other characteristics of the one or more objects. In such embodiments, the topographical information may include information related to the sonic characteristics of the topography of the virtual space. The sonic characteristics of the topography of the virtual space may include acoustic characteristics of the topography, and/or other sonic characteristics of the topography.

According to various embodiments, content included within the virtual space (e.g., visual content formed on portions of the topography or objects present in the virtual space, objects themselves, etc.) may be identified within the information stored in storage module12by reference only. For example, rather than storing a structure and/or a texture associated with the structure, storage module12may instead store an access location at which visual content to be implemented as the structure (or a portion of the structure) or texture can be accessed. In some implementations, the access location may include a URL that points to a network location. The network location identified by the access location may be associated with a network asset22. Network asset22may be located remotely from each of storage module12, server14, and client16. For example, the access location may include a network URL address (e.g., an internet URL address, etc.) at which network asset22may be accessed.

It should be appreciated that not only solid structures within the virtual space may be identified in the information stored in storage module12may be stored by reference only. For example, visual effects that represent unseen forces or influences may be stored by reference as described above. Further, information stored by reference may not be limited to visual content. For example, audio content expressed within the virtual space may be stored within storage module12by reference, as an access location at which the audio content can be accessed. Other types of information (e.g., interface information, space parameter information, etc.) may be stored by reference within storage module12.

The interface information stored within storage module12may include information related to an interface provided to the user that enables the user to interact with the virtual space. More particularly, in some implementations, the interface information may include a mapping of an input device provided at client16to commands that can be input by the user to system10. For example, the interface information may include a key map that maps keys in a keyboard (and/or keypad) provided to the user at client16to commands that can be input by the user to system10. As another example, the interface information may include a map that maps the inputs of a mouse (or joystick, or trackball, etc.) to commands that can be input by the user to system10. In some implementations, the interface information may include information related to a configuration of an interface display provided to the user at client16, through which the user may input information to system10. For example, the interface display may receive communication to other users interacting with the virtual space, input dictating actions to be performed by one or more objects within the virtual space, a request for a different point of view for the view, a request for a more (or less) sophisticated view (e.g., a 2-dimensional view, a 3-dimensional view, etc.), a request for one or more additional types of data to be displayed in the interface display, and/or other information.

The interface display may be configured (e.g., by the interface information stored in storage module12) to provide information to the user about conditions in the virtual space that may not be apparent simply from viewing the space. For example, such conditions may include the passage of time, ambient environmental conditions, and/or other conditions. The interface display may be configured (e.g., by the interface information stored in storage module12) to provide information to the user about one or more objects within the space. For instance, information may be provided to the user about objects associated with the topography of the virtual space (e.g., coordinate, elevation, size, identification, age, status, etc.). In some implementations, information may be provided to the user about objects that represent animate characters (e.g., wealth, health, fatigue, age, experience, etc.). For example, such information may be displayed that is related to an object that represents a character associated with client16in the virtual space (e.g., an avatar, a character being controlled by the user, etc.).

The space parameter information may include information related to one or more parameters of the virtual space. Parameters of the virtual space may include, for example, the rate at which time passes, dimensionality of objects within the virtual space (e.g., 2-dimensional vs. 3-dimensional), permissible views of the virtual space (e.g., first person views, bird's eye views, 2-dimensional views, 3-dimensional views, fixed views, dynamic views, selectable views, etc.), and/or other parameters of the virtual space. In some implementations, the space parameter information includes information related to the game parameters of a game provided within the virtual space. For instance, the game parameters may include information related to a maximum number of players, a minimum number of players, the game flow (e.g., turn based, real-time, etc.), scoring, spectators, and/or other game parameters of a game.

The information related to the plurality of virtual spaces may be stored in an organized manner within information storage18. For example, the information may be organized into a plurality of space records24(illustrated as space record24a, space record24b, and space record24c). Individual ones of space records24may correspond to individual ones of the plurality of virtual spaces. A given space record24may include information related to the corresponding virtual space. In some embodiments, the space records24may be stored together in a single hierarchal structure (e.g., a database, a file system of separate files, etc.). In some embodiments, space records24may include a plurality of different “sets” of space records24, wherein each set of space records includes one or more of space records24that is stored separately and discretely from the other space records24.

Information storage18may store information related to one or more characters that can be manipulated by users to interact with the virtual spaces. The information may be organized into one or more character records25, which correspond to individual characters. In some implementations, character record25corresponding to a given character may be included within a user record (not shown) that includes information related to a user that is associated with the character. In such implementations, a single user record may include one or more character records25. For example, where a single user has several characters that can optionally be employed by the user to interact with one or more virtual spaces provided by system10, a single user record may include a character record that corresponds to each of the individual characters associated with the user. In other implementations, character records25may be maintained separately from user records, and associations between characters and users may be recorded by references between character records25and user records. For instance, in cases where a single use has several characters that can be employed by the user to interact with one or more virtual spaces provided by system10, a single user record may include references to each of character records25that correspond to the user's characters.

Some or all of the information included within a given character record25may be persistent between different virtual spaces and/or different types of virtual spaces. As such, character records25may store information that enables individual characters to be manifested within a plurality of the virtual spaces, including different types of virtual spaces. This may enable a user to transfer her character between virtual spaces with some degree of continuity. The information included within a given character record25may include one or more of manifestation information that is specific to the corresponding character, character parameter information, character inventory information, character interface information, and/or other information related to the corresponding character.

In some implementations, a manifestation of a character may include the locomotion characteristics of the character, the size of the character, the strength of the character, the weight of the character, the visual representation of the character, the identity and/or nature of the character, interaction characteristics of the character, movement characteristics of the character, sonic characteristics of the character, a character type or class (e.g., human, animal, alien, warrior, priest, tradesman, etc.), and/or other aspects of the manifestation of the character. The interaction characteristics of a character described by the manifestation within the corresponding character record25may include information related to the manner in which the character interacts with and/or is influenced by other objects and/or characters within the virtual spaces, the topography (e.g., features of the topography) of virtual spaces, unseen forces in the virtual spaces, and/or other characteristics of the manner in which the character interacts with outside forces, objects, and/or characters in the virtual spaces. The interaction characteristics of the character may include scriptable behaviors. Accordingly, the manifestation information stored for a given character in the corresponding character record25may include one or both of a script and/or a trigger associated with a given scriptable behavior of the character within virtual space.

In some implementations, some or all of the manifestation information associated with a given character may be shared in common between the given character and a group of other characters, of which the given character is a part. In such implementations, character record25may include the common manifestation information by reference to a record on information storage18that includes the common manifestation information. For example, if a group of characters known as “elves” exist with a predetermined set of interaction characteristics, these interaction characteristics may be identified in a character record25corresponding to an elf character by referencing some record on information storage18that includes the predetermined set of interaction characteristics that are common to elf characters.

As was the case generally with content found in the virtual spaces, certain content associated with a character (e.g., a texture, an object, an overall appearance of the character, a script describing motion of the character, etc.) may be identified within the corresponding character record25by reference only. Such content maybe referenced by an access location (e.g., a URL, another record on information storage18, etc.) at which the content can be accessed.

Some or all of the manifestation information within a given character record25may be persistent between a plurality of different virtual worlds, and/or virtual world types. As is discussed further below, some of the manifestation information within character record25may not be persistent between all of the virtual spaces provided by system10.

In some implementations, character parameter information may include information related to parameters of a character within one or more of the virtual spaces. By way of non-limiting example, the character parameter information may include information related to one or more of an acquired skill of the character, a skill level of the character, a status of the character, a social connection or friendship between the character and one or more other characters, a score achieved by the character, a permission afforded to the character (e.g., to access a restricted area in a virtual space, to access a restricted virtual space, etc.), and/or other parameters of the character.

One or more of the parameters represented by the character parameter information may be persistent between a plurality of the virtual spaces and/or virtual space types. This may enable parameters gained in one virtual space to be transferred into another virtual space, even where the other virtual space is of a different virtual space type. One or more of the parameters represented by the character parameter information may not be persistent between all of the virtual spaces and/or virtual space types. However, in such implementations, the information related to these parameters may be stored on a per space, and/or per set of spaces (e.g., grouped according to virtual space type, grouped according to a common scheme or theme, etc.), basis. This may enable the character to leave a first virtual space and/or first set of virtual spaces, within which a given parameter is persistent, to enter a second virtual space (or second set of virtual spaces) without maintaining the persistent representation of the parameter, and upon return to the first virtual space (or first set of virtual spaces) the representation of the parameter stored for the first virtual space (or first set of virtual spaces) is restored to the character. In some of these implementations, character record25may store different representations of the same parameter (or similar parameters) for the first virtual space (or first set of virtual spaces) and the second virtual space (or second set of virtual spaces).

As should be appreciated from the foregoing, parameters of a character may be altered in one virtual space, for example, due to achievement within the virtual space. As a non-limiting example, in a first virtual space the character may gain and/or enhance a skill that may be useful in another virtual space and/or type of virtual space. Since the character parameter that represents this parameter (i.e., the skill) may be persistent between a plurality of virtual spaces, when the character moves from the virtual space in which the skill was gained and/or enhanced to another one of the virtual spaces in which the skill is a parameter, even if the character has passed through other virtual spaces in which the skill was not a parameter, the skill will be represented at the gained and/or augmented level. This may provide a continuity to the virtual spaces provided by system10not present in conventional systems that provide monolithic virtual spaces.

In some implementations, character inventory information may include information related to an amount of virtual currency currently possessed by the character, information related to objects in the possession and/or under the control of the character, and/or other information related to an inventory associated with the character. At least a portion of the character inventory information may be persistent between different virtual spaces and/or types of virtual spaces. This may enable the character to take her “possessions” with her between the various virtual spaces. In some cases, a given object in the inventory of a character, the character record25associated with the character may include a relatively complete description of the object (e.g., manifestation information associated with the object, etc.). In some cases, a given object in the inventory of a character may be associated with an object record that is separate from character record25, and includes information that enables the object to be manifested within the virtual spaces. In these cases, character record25and/or the object record may reference each other in such a manner that the object is associated with the character as being in the character's inventory.

As is discussed in greater detail in U.S. patent application Ser. No. 12/249,154, filed Oct. 10, 2008 (U.S. Patent Publication No. US 2010-0095213 A1), which was incorporated by reference above, objects may be manifest within the virtual spaces solely on the basis of information stored in character record25and/or in object records. This may enable flexibility of the platform provided by system10in manifesting objects such that a given object may be manifested within a virtual spaces primarily based on the manifestation information included within the corresponding object record and/or character record25across a variety of virtual spaces and/or virtual space types. In other words, the information stored within an object record and/or character record25may enable an object to be manifested with virtual spaces provided by system10without being preconfigured within the space (and without the space being preconfigured for the object). Accordingly, the character inventory information included within character record25(in some cases in conjunction with an object record) may enable the character corresponding to the character record25to transport the object into at least a substantial set of the virtual spaces provided by system10. This may provide an enhanced level of transportability for possessions of the character between different virtual spaces and/or types of virtual spaces in comparison with traditional systems that require objects to have been previously introduced within a virtual space before they can be possessed, manifested, and/or used in accordance with a character's commands.

In some implementations, character interface information may include information related to an interface provided to the user that enables the user to control a character within the virtual spaces. For example, the character interface information included in a given character record25may include information that configures an input device provided at client16to control the character that corresponds to the given character record in a predetermined manner. The information that configures the input device may include a mapping of the input device provided at client16to commands that can be input to system10to control the character in a predetermined manner. The input device may include, for example, a keyboard, a mouse, a joystick, a trackball, and/or other input devices.

In some implementations, the character interface information may include information that configures a display of information related to the character. For example, the character interface information may include information that configures a display provided to the user at client16. This may include configuring one or more of the type of information that is displayed, the position of the information on a display, the size of the information on the display, and/or other manners of configuring a display of information related to the character.

In some implementations, the character interface information may include information that configures a graphical user interface related to the character. For example, the graphical user interface may display information related to the character and/or enable control of the character. Configuring a graphical user interface may include one or more of determining the size and/or location of displays of information related to the character, determining information related to the commands that can be input through the graphical user interface, determining the manner in which the commands that can be input through the graphical user interface are selected, and/or determining other parameters of the graphical user interface.

Although information storage18is illustrated inFIG. 1as a single entity, this is for illustrative purposes only. In some embodiments, information storage18includes a plurality of informational structures that facilitate management and storage of the information related to the plurality of virtual spaces. Information storage18may include the physical storage elements for storing the information related to the virtual spaces, users of system10, the characters, and/or the information processing and storage assets that enable information storage18to manage, organize, and maintain the stored information. Information storage18may include a relational database, an object oriented database, a hierarchical database, a post-relational database, flat text files (which may be served locally or via a network), XML files (which may be served locally or via a network), and/or other information structures.

In some embodiments, in which information storage18includes a plurality of informational structures that are separate and discrete from each other. In such embodiments, system10may include a central information catalog (e.g., managed by storage module12) that includes information related to the location of the space records, the user records, and/or the character records included in information storage18(e.g., network and/or file system addresses of individual space records). In some embodiments, the central information catalog may form a clearing house of information that enables users to initiate instances a chosen virtual space (e.g., to establish a server executing an instance of the chosen virtual space similar to server14), to grant access of system10to users, to manifest characters within instances of the virtual spaces, and/or to perform other functionality within system10. Accordingly, access to the information stored within the central information catalog may be provided to users and/or servers based on privileges (e.g., earned via monetary payment, administrative privileges, earned via previous game-play, earned via membership in a community, etc.).

Server communication module20may facilitate communication between information storage18and server14. In some embodiments, server communication module20enables this communication by formatting communication between information storage18and server14. This may include, for communication transmitted from information storage18to server14, generating markup elements (e.g., “tags”) that convey the information stored in information storage18, and transmitting the generated markup elements to server14. For communication transmitted from server14to information storage18, server communication module20may receive markup elements transmitted from server14to storage module12and may reformat the information for storage in information storage18.

Server14may be provided remotely from storage module12. Communication between server14and storage module12may be accomplished via one or more communication media. For example, server14and storage module12may communicate via a wireless medium, via a hard-wired medium, via a network (e.g., wireless or wired), and/or via other communication media. In some embodiments, server14may include a communication module26, an instantiation module28, a view module30, a character module31, and/or other modules. Modules26,28,30, and31may be implemented in software; hardware; firmware; some combination of software, hardware, and/or firmware; and/or otherwise implemented. It should be appreciated that although modules26,28,30, and/or31are illustrated inFIG. 1as being co-located within a single unit (server14), in some implementations, server14may include multiple units and modules26,28,30, and/or31may be located remotely from the other modules.

Communication module26may be configured to communicate with storage module12, and/or client16. Communicating with storage module12and/or client16may include transmitting and/or receiving markup elements of the markup language. The markup elements received by communication module26may be implemented by other modules of server14, or may be passed between storage module12and client16via server14(as server14serves as an intermediary therebetween). The markup elements transmitted by communication module26to storage module12or client16may include markup elements being communicated from storage module to client16(or vice versa), or the markup elements may include markup elements generated by the other modules of server14.

Instantiation module28may be configured to execute one or more instances of one or more virtual spaces, such as an instance32of a virtual space present on server14. Instantiation module28may execute instance32of the virtual space according to information received in markup element form from storage module12. Instantiation module28may comprise an application that is configured to execute instances of virtual spaces based on information conveyed thereto in markup element form. The application may be capable of initiating execution of an instance of a virtual space without accessing a local source of information (local to server14) that describes various aspects of the configuration of the virtual space (e.g., manifestation information, space parameter information, etc.), or without making assumptions about such aspects of the configuration of the virtual space when initiating execution of the virtual space. Instead, such information may be obtained by instantiation module28from the markup elements communicated to server14from storage module12. This may provide one or more enhancements over systems in which an application running on a server executes an instance of a virtual space dictates aspects of the virtual space (e.g., in “World of Warcraft”), such as manifestation information and/or space parameter information, or makes assumptions about such aspects. For example, the application included in instantiation module28may be capable of executing instances of a wider variety of “types” of virtual spaces (e.g., virtual worlds, games, 3-D spaces, 2-D spaces, spaces with different views, first person spaces, birds-eye spaces, real-time spaces, turn based spaces, etc.).

Instance32may be characterized as a simulation of the virtual space that is being executed on server14by instantiation module30. The simulation may include determining in real-time the positions, structure, and manifestation of objects, unseen forces, and topography within the virtual space according to the topography, manifestation, and space parameter information that corresponds to the virtual space. As has been discussed above, various portions of the content that make up the virtual space embodied in instance32may be identified in the markup elements received from storage module12by reference. In such implementations, instantiation module28may be configured to access the content at the access location identified (e.g., at network asset22, as described above) in order to account for the nature of the content in instance32.

As instance32is maintained by instantiation module28on server14, and the position, structure, and manifestation of objects, unseen forces, and topography within the virtual space varies, instantiation module28may implement an instance memory module34to store information related to the present state of instance32. Instance memory module34may be provided locally to server14(e.g., integrally with server14, locally connected with server14, etc.), or instance memory module34may be located remotely from server14and an operative communication link may be formed therebetween.

View module30may be configured to implement instance32to determine a view of the virtual space. The view of the virtual space may be from a fixed location or may be dynamic (e.g., may track an object). In some implementations, a character associated with client16(e.g., an avatar) may be included within instance32. In these implementations, the location of the character within instance32may influence the view determined by view module30(e.g., track with the position of the character, be taken from the perspective of the character, etc.). The view determined by view module30may be determined for a variety of different perspectives (e.g., a bird's eye view, an elevation view, a first person view, etc.). The view may be a 2-dimensional view or a 3-dimensional view. These and/or other aspects of the view may be determined for the virtual space based on information stored in a space record24for the virtual space and provided from storage module12via markup elements (e.g., as space parameter information). Determining the view may include determining the identity, shading, size (e.g., due to perspective), motion, and/or position of objects, effects, and/or portions of the topography that would be present in a rendering of the view.

View module30may generate a plurality of markup elements that describe the view based on the determination of the view. The plurality of markup elements may describe identity, shading, size (e.g., due to perspective), and/or position of the objects, effects, and/or portions of the topography that should be present in a rendering of the view. The markup elements may describe the view “completely” such that the view can be formatted for viewing by the user by simply assembling the content identified in the markup elements according to the attributes of the content provided in the markup elements. In such implementations, assembly alone may be sufficient to achieve a display of the view of the virtual space, without further processing of the content (e.g., to determine motion paths, decision-making, scheduling, triggering, etc.).

In some embodiments, the markup elements generated by view module30that describe the view identify content (e.g., visual content, audio content, etc.) to be included in the view by reference only. For example, as was the case with markup elements transmitted from storage module12to server14, the markup elements generated by view module30may identify content by a reference to an access location. The access location may include a URL that points to a network location. The network location identified by the access location may be associated with a network asset (e.g., network asset22). For instance, the access location may include a network URL address (e.g., an internet URL address, etc.) at which network asset22may be accessed.

According to various embodiments, in generating the view, view module30may manage various aspects of content included in views determined by view module30, but stored remotely from server14(e.g., content referenced in markup elements generated by view module30). Such management may include re-formatting content stored remotely from server14to enable client16to convey the content (e.g., via display, etc.) to the user. For example, in some implementations, client16may be executed on a relatively limited platform (e.g., a portable electronic device with limited processing, storage, and/or display capabilities). Server14may be informed of the limited capabilities of the platform (e.g., via communication from client16to server14) and, in response, view module30may access the content stored remotely in network asset22to re-format the content to a form that can be conveyed to the user by the platform executing client16(e.g., simplifying visual content, removing some visual content, re-formatting from 3-dimensional to 2-dimensional, etc.). In such implementations, the re-formatted content may be stored at network asset22by over-writing the previous version of the content, stored at network asset22separately from the previous version of the content, stored at a network asset36that is separate from network asset22, and/or otherwise stored. In implementations in which the re-formatted content is stored separately from the previous version of the content (e.g., stored separately at network asset22, stored at network asset24, cached locally by server14, etc.), the markup elements generated by view module30for client16reflect the access location of the re-formatted content.

As was mentioned above, in some embodiments, view module30may adjust one or more aspects of a view of instance32based on communication from client16indicating that the capabilities of client16may be limited in some manner (e.g., limitations in screen size, limitations of screen resolution, limitations of audio capabilities, limitations in information communication speeds, limitations in processing capabilities, etc.). In such embodiments, view module30may generate markup elements for transmission that reduce (or increase) the complexity of the view based on the capabilities (and/or lack thereof) communicated by client16to server14. For example, view module30may remove audio content from the markup elements, view module30may generate the markup elements to provide a two dimensional (rather than a three dimensional) view of instance32, view module30may reduce, minimize, or remove information dictating motion of one or more objects in the view, view module30may change the point of view of the view (e.g., from a perspective view to a bird's eye view), and/or otherwise generate the markup elements to accommodate client16. In some implementations, these types of accommodations for client16may be made by server14in response to commands input by a user on client16as well as or instead of based on communication of client capabilities by client16. For example, the user may input commands to reduce the load to client16caused by displaying the view to improve the quality of the performance of client16in displaying the view, to free up processing and/or communication capabilities on client16for other functions, and/or for other reasons.

From the description above it should be apparent that as view module30“customizes” the markup elements that describe the view for client16, a plurality of different versions of the same view may be described in markup elements that are sent to different clients with different capabilities, settings, and/or requirements input by a user. This customization by view module30may enhance the ability of system10to be implemented with a wider variety of clients and/or provide other enhancements.

Character module31may receive information related to a character introduced into instance32by the user (e.g., via client16). The information may be implemented by instantiation module28and/or view module30in generating instance32and/or a view thereof. The received information may include information received from client16and/or information storage18. For example, in order to “enter” instance32, client16may transmit a request to server14for access to instance32. The request may include an identification that enables character module31to determine a character that should be manifested within instance32as a representation of the user. The identification may be information that identifies one or more of a user associated with the character, a user record corresponding to the user, the character, and/or a character record corresponding to the character. Determination of the character by character module31may include identifying a character record25from which information related to the character can be accessed. Character module31may then obtain information from the appropriate character record25that will enable instantiation module28to manifest the character within instance32and/or will enable view module30to determine the appropriate view of instance32for transmission to the client.

As has been set forth above, the information within character record25accessed by character module31may include one or more of manifestation information that is specific to the character, character parameter information, character inventory information, character interface information, and/or other information related to the character. At least some of the information within character record25may be persistent between instance32and instances of other virtual spaces. However, this may not be the case for some of the information within character record25. Some of the information within character record may be specific to the virtual space of instance32(and/or some group of virtual spaces of which the virtual space of instance32is a part). For example, aspects of the visual representation of the character may be specific to the virtual space and/or group of virtual spaces, objects and/or currency in the inventory of the character may be specific to the virtual space and/or group of virtual spaces, a score, a skill, a skill level, a permission, a social connection, and/or other parameters of the character may be specific to the virtual space and/or group of virtual spaces, and/or other information included in character record25may be specific to the virtual spaces and/or group of virtual spaces.

In some implementations, character module31may obtain information from character record25by transmitting a request to storage module12that identifies the character record25including information related to the character that is going to be manifested within instance32. This request may include an identification of the virtual space of instance32. In response to the request, storage module12may transmit character record25and/or some portion thereof (e.g., including information relevant to the virtual space identified in the request) back to character module31.

According to various embodiments of the invention, if a character is being introduced into a virtual space (e.g., manifested into an instance of the virtual space) for the first time, character record25may not include all of the information necessary to manifest the character. For example, the virtual space may be configured such that some of the information within character record25is specific to the virtual space. If the character has not been introduced into the virtual space, then this information may not yet have been entered to character record25. In such cases, character module31may enable the user to configure the character for the virtual space.

By way of non-limiting example, a virtual space may feature aspects of visual representations of the characters present therein (e.g., may have specialized clothing, may require characters to be represented as a one of one or more specific types of creatures (e.g., a dragon, an elf, a wizard, soldier, etc.), etc.). Upon entry into an instance of the virtual space, character module31may assign default values for these aspects of the visual representation of the character. These default values may be written to character record25that corresponds to the character. Character module31may present an interface to the user (e.g., via client16) that enables the user to change and/or configure these aspects of the visual representation of the character of the user in a customizable manner (rather than the default values). These customized aspects of the visual representation may be written to character record25. Writing the configured information (whether default or customized) to character record25enables the information to be retrieved if the character leaves the instance of the virtual world and then returns to the virtual world at some future time such that the character will not have to be reconfigured for future entries into the virtual space.

It should be appreciated that the example of configuring, and writing to character record25, aspects of the visual representation of the character that are specific to the virtual space (or some group of virtual spaces of which this group of virtual spaces is a part) is intended solely for illustrative purposes. Other information (e.g., other manifestation information, character inventory information, character parameter information, character interface information, etc.) related to the character may be determined by character module31(based on defaults and/or customization) upon entry of the character into the virtual space for the first time in a manner that is similar to or substantially the same as the configuration and storage of the aspects of the visual representation of the character discussed above.

In some implementations, information related to the character may change and/or evolve during the presence of the character within instance32. For example, the appearance of the character may change over time to represent a maturation of the character, the character may acquire and/or enhance one or more skills, the character may obtain currency and/or objects, a score associated with the character may changed, manifestation information associated with the character may change (e.g., the character may grow, become injured, get stronger, etc.), the character may gain a permission, the character may make a new social connection or enhance an existing social connection, and/or other information associated with the character may change. These changes may happen spontaneously within instance32(e.g., instance32may cause the character to mature over time), may be caused by game-play within instance32, may be caused by customization and/or configuration of the user (via client16), and/or may have other causes. Character module31may manage the storage of these changes in character record25. These changes may be stored within character record25so that if the character leaves instance32and enters another virtual space, information related to the character that is persistent between instance32and the other virtual space will reflect the changes in the other virtual space. If the information related to the character that is changed within instance32is not persistent between instance32and the other virtual space, then these changes will be stored within character record25until the character re-enters instance32and/or an instance of some other virtual space for which the changed information is persistent.

In some embodiments, client16provides an interface to the user that includes a view display38, a user interface display40, an input interface42, and/or other interfaces that enable interaction of the user with the virtual space. Client16may include a server communication module44, a view display module46, an interface module48, and/or other modules. Client16may be executed on a computing platform that includes a processor that executes modules44and46, a display device that conveys displays38and40to the user, and provides input interface42to the user to enable the user to input information to system10(e.g., a keyboard, a keypad, a switch, a knob, a lever, a touchpad, a touchscreen, a button, a joystick, a mouse, a trackball, etc.). The platform may include a desktop computing system, a gaming system, or more portable systems (e.g., a mobile phone, a personal digital assistant, a hand-held computer, a laptop computer, etc.). In some embodiments, client16may be formed in a distributed manner (e.g., as a web service). In some embodiments, client16may be formed in a server. In these embodiments, a given virtual space instanced on server14may include one or more objects that present another virtual space (of which server14becomes the client in determining the views of the first given virtual space).

Server communication module44may be configured to receive information related to the execution of instance32on server14from server14. For example, server communication module44may receive markup elements generated by storage module12(e.g., via server14), view module30, and/or other components or modules of system10. The information included in the markup elements may include, for example, view information that describes a view of instance32of the virtual space, interface information that describes various aspects of the interface provided by client16to the user, and/or other information. Server communication module44may communicate with server14via one or more protocols such as, for example, WAP, TCP, UDP, and/or other protocols. The protocol implemented by server communication module44may be negotiated between server communication module44and server14.

View display module48may be configured to format the view described by the markup elements received from server14for display on view display38. Formatting the view described by the markup elements may include assembling the view information included in the markup elements. This may include providing the content indicated in the markup elements according to the attributes indicated in the markup elements, without further processing (e.g., to determine motion paths, decision-making, scheduling, triggering, etc.). As was discussed above, in some implementations, the content indicated in the markup elements may be indicated by reference only. In such implementations, view display module46may access the content at the access locations provided in the markup elements (e.g., the access locations that reference network assets22and/or36, or objects cached locally to server14). In some of these implementations, view display module46may cause one or more of the content accessed to be cached locally to client16, in order to enhance the speed with which future views may be assembled. The view that is formatted by assembling the view information provided in the markup elements may then be conveyed to the user via view display38.

As has been mentioned above, in some implementations, the capabilities of client16may be relatively limited. In some such implementations, client16may communicate these limitations to server14, and the markup elements received by client16may have been generated by server14to accommodate the communicated limitations. However, in some such implementations, client16may not communicate some or all of the limitations that prohibit conveying to the user all of the content included in the markup elements received from server14. Similarly, server14may not accommodate all of the limitations communicated by client16as server14generates the markup elements for transmission to client16. In these instances, view display module48may be configured to exclude or alter content contained in the markup elements in formatting the view. For example, view display module48may disregard audio content if client16does not include capabilities for providing audio content to the user. As another example, if client16does not have the processing and/or display resources to convey movement of objects in the view, view display module48may restrict and/or disregard motion dictated by motion information included in the markup elements.

Interface module48may be configured to configure various aspects of the interface provided to the user by client16. For example, interface module48may configure user interface display40and/or input interface42according to the interface information provided in the markup elements. User interface display40may enable display of the user interface to the user. In some implementations, user interface display40may be provided to the user on the same display device (e.g., the same screen) as view display38. As was discussed above, the user interface configured on user interface display40by interface module38may enable the user to input communication to other users interacting with the virtual space, input actions to be performed by one or more objects within the virtual space, provide information to the user about conditions in the virtual space that may not be apparent simply from viewing the space, provide information to the user about one or more objects within the space, input commands and/or request for information related to a character in instance32that is associated with the user, and/or provide for other interactive features for the user. In some implementations, the markup elements that dictate aspects of the user interface may include markup elements generated at storage module12(e.g., at startup of instance32) and/or markup elements generated by server14(e.g., by view module30) based on the information conveyed from storage module12to server14via markup elements. This information may include information from within space record24, character record25, and/or other records (e.g., user records) stored in storage module12.

In some implementations, interface module48may configure input interface42according to information received from server14via markup elements. For example, interface module48may map the manipulation of input interface42by the user into commands to be input to system10based on a predetermined mapping that is conveyed to client16from server14via markup elements. The predetermined mapping may include, for example, a key map and/or other types of interface mappings (e.g., a mapping of inputs to a mouse, a joystick, a trackball, and/or other input devices). If input interface42is manipulated by the user, interface module48may implement the mapping to determine an appropriate command (or commands) that correspond to the manipulation of input interface42by the user. Similarly, information input by the user to user interface display40(e.g., via a command line prompt) may be formatted into an appropriate command for system10by interface module48. In some implementations, the availability of certain commands, and/or the mapping of such commands may be provided based on privileges associated with a user manipulating client16(e.g., as determined from a login). For example, a user with administrative privileges, premium privileges (e.g., earned via monetary payment), advanced privileges (e.g., earned via previous game-play), and/or other privileges may be enabled to access an enhanced set of commands. These commands formatted by interface module48may be communicated to server14by server communication module44.

Upon receipt of commands from client16that include commands input by the user (e.g., via communication module26), server14may enqueue for execution (and/or execute) the received commands. The received commands may include commands related to the execution of instance32of the virtual space. For example, the commands may include display commands (e.g., pan, zoom, etc.), object manipulation commands (e.g., to move one or more objects in a predetermined manner), character action commands (e.g., for the character associated with client16to perform a predetermined action), communication commands (e.g., to communicate with other users interacting with the virtual space), information requests, and/or other commands. Instantiation module38may execute the commands in the virtual space by manipulating instance32of the virtual space. The manipulation of instance32in response to the received commands may be reflected in the view generated by view module30of instance32, which may then be provided back to client16for viewing. Thus, commands input by the user at client16enable the user to interact with the virtual space without requiring execution or processing of the commands on client16itself.

It should be appreciated that system10as illustrated inFIG. 1is not intended to be limiting in the numbers of the various components and/or the number of virtual spaces being instanced. For example,FIG. 2illustrates a system50, similar to system10, including a storage module52, a plurality of servers54,56, and58, and a plurality of clients60,62, and64. Storage module52may perform substantially the same function as storage module12(shown inFIG. 1and described above). Servers54,56, and58may perform substantially the same function as server14(shown inFIG. 1and described above). Clients60,62, and64may perform substantially the same function as client16(shown inFIG. 1and described above).

Storage module52may store information related to a plurality of virtual spaces, and may communicate the stored information to servers54,56, and/or58via markup elements of the markup language, as was discussed above. Servers54,56, and/or58may implement the information received from storage module52to execute instances66,68,70, and/or70of virtual spaces. As can be seen inFIG. 2, a given server, for example, server58, may be implemented to execute instances of a plurality of virtual spaces (e.g., instances70and72). Clients60,62, and64may receive information from servers54,56, and/or58that enables clients60,62, and/or64to provide an interface for users thereof to one or more virtual spaces being instanced on servers54,56, and/or58. The information received from servers54,56, and/or58may be provided as markup elements of the markup language, as discussed above.

Due at least in part to the implementation of the markup language to communicate information between the components of system50, it should be appreciated from the foregoing description that any of servers54,56, and/or58may instance any of the virtual spaces stored on storage module52. The ability of servers54,56, and/or58to instance a given virtual space may be independent, for example, from the topography of the given virtual space, the manner in which objects and/or forces are manifest in the given virtual space, and/or the space parameters of the given virtual space. This flexibility may provide an enhancement over conventional systems for instancing virtual spaces, which may only be capable of instancing certain “types” of virtual spaces. Similarly, clients60,62, and/or64may interface with any of the instances66,68,70, and/or72. Such interface may be provided without regard for specifics of the virtual space (e.g., topography, manifestations, parameters, etc.) that may limit the number of “types” of virtual spaces that can be provided for with a single client in conventional systems. In conventional systems, these limitations may arise as a product of the limitations of platforms executing client16, limitations of client16itself, and/or other limitations.

In some embodiments, system10may enable the user to create a virtual space. In such embodiments, the user may select a set of characteristics of the virtual space on client16(e.g., via user interface display48and/or input interface42). The characteristics selected by the user may include characteristics of one or more of a topography of the virtual space, the manifestation in the virtual space of one or more objects and/or unseen forces, an interface provided to users to enable the users to interact with the new virtual space, space parameters associated with the new virtual space, and/or other characteristics of the new virtual space.

Although the invention has been described in detail for the purpose of illustration based on what is currently considered to be the most practical and preferred embodiments, it is to be understood that such detail is solely for that purpose and that the invention is not limited to the disclosed embodiments, but, on the contrary, is intended to cover modifications and equivalent arrangements that are within the spirit and scope of the appended claims. For example, it is to be understood that the present invention contemplates that, to the extent possible, one or more features of any embodiment can be combined with one or more features of any other embodiment.

Claims

  1. An information storage system configured to store information related to characters within virtual spaces, the information storage system comprising: a storage component comprising: character storage that stores a plurality of character records including a first character record corresponding to a first user and a second character record corresponding to a second user;wherein the first character record comprises information that enables a first server executing an instance of a first virtual space to manifest a first character associated with the first user in the first virtual space, the first server being configured to effectuate presentation of the instance of the first virtual space on a first client device associated with the first user;wherein the first character record further comprises information that enables a second server executing an instance of a second virtual space to manifest a second character associated with the first user in the second virtual space, the second server being configured to effectuate presentation of the instance of the second virtual space on a second client device associated with the first user;wherein the first virtual space has one or more aspects that are different from the second virtual space;wherein one or more differences between the first character and the second character result from the one or more aspects of the first virtual space that are different from the second virtual space;and wherein at least a portion of the information in the first character record corresponding to the first character is common with a portion of the information in the first character record corresponding to the second character;one or more physical processors configured to receive requests from the first server executing the instance of the first virtual space and the second sever executing the instance of the second virtual space for information within the character storage, and to transmit the requested information to the first and second servers responsive to such requests;and wherein the storage component is separate and distinct from the first server, first client device, the second server, and the second client device.
  1. The system of claim 1 , wherein the character storage is configured such that the at least a portion of the information in the first character record corresponding to the first character that is common with a portion of the information in the first character record corresponding to the second character comprises information that reflects character progress in one or both of the first virtual space or the second virtual space.
  2. The system of claim 1 , wherein the one or more differences between the first character and the second character result from a difference in one or more of dimensionality, point-of-view, game mechanic, or game genre between the first virtual space and the second virtual space.
  3. The system of claim 1 , wherein the first virtual space is a simulated physical space that has a topography, expresses real-time interaction by the users, and includes one or more objects positioned within the topography that are capable of experiencing locomotion within the topography.
  4. The system of claim 1 , wherein the character storage is configured such that the at least a portion of the information in the first character record corresponding to the first character that is common with a portion of the information in the first character record corresponding to the second character comprises information that reflects a social connection with another user initiated by the first user in the first virtual space.
  5. The system of claim 1 , wherein the character storage is configured such that the at least a portion of the information in the first character record corresponding to the first character that is common with a portion of the information in the first character record corresponding to the second character comprises information that reflects the acquisition of a virtual object by the first character in the first virtual space.
  6. The system of claim 1 , wherein the character storage is configured such that the second character record comprises information that enables the first server executing an instance of the first virtual space to manifest a third character associated with the second user in the first virtual space, wherein the second character record further comprises information that enables the second server executing an instance of the second virtual space to manifest a fourth character associated with the second user in the second virtual space, wherein one or more differences between the third character and the fourth character result from the one or more aspects of the first virtual space that are different from the second virtual space, and wherein at least a portion of the information in the second character record corresponding to the third character is common with a portion of the information in the second character record corresponding to the fourth character.
  7. The method of claim 1 , wherein the second character record comprises information that enables the first server executing an instance of the first virtual space to manifest a third character associated with the second user in the first virtual space, wherein the second character record further comprises information that enables the second server executing an instance of the second virtual space to manifest a fourth character associated with the second user in the second virtual space, wherein one or more differences between the third character and the fourth character result from the one or more aspects of the first virtual space that are different from the second virtual space, and wherein at least a portion of the information in the second character record corresponding to the third character is common with a portion of the information in the second character record corresponding to the fourth character.
  8. A computer-implemented method of storing information related to characters within virtual spaces, the method being implemented in a storage system comprising non-transient electronic storage component and one or more physical processors, the method comprising: storing, to the storage component, a plurality of character records including a first character record corresponding to a first user and a second character record corresponding to a second user;wherein the first character record comprises information that enables a first server executing an instance of a first virtual space to manifest a first character associated with the first user in the first virtual space, the first server being configured to effectuate presentation of the instance of the first virtual space on a first client device associated with the first user;wherein the first character record further comprises information that enables a second server executing an instance of a second virtual space to manifest a second character associated with the first user in the second virtual space, the second server being configured to effectuate presentation of the instance of the second virtual space on a second client device associated with the first user;wherein the first virtual space has one or more aspects that are different from the second virtual space;wherein one or more differences between the first character and the second character result from the one or more aspects of the first virtual space that are different from the second virtual space;and wherein at least a portion of the information in the first character record corresponding to the first character is common with a portion of the information in the first character record corresponding to the second character;receiving a first request from the first server executing an instance of the first virtual space for the information from the first character record corresponding to the first character;transmitting to the first server, responsive to the first request, the information from the first character record corresponding to the first character;receiving a second request from the second server executing an instance of the second virtual space for the information from the first character record corresponding to the second character;transmitting to the second server, responsive to the second request, the information from the second character record corresponding to the second character;and wherein the electronic storage component is separate and distinct from the first server, the first client device, the second server, and the second client device.
  9. The method of claim 9 , wherein the at least a portion of the information in the first character record corresponding to the first character that is common with a portion of the information in the first character record corresponding to the second character comprises information that reflects character progress in one or both of the first virtual space or the second virtual space.
  10. The method of claim 9 , wherein the one or more differences between the first character and the second character result from a difference in one or more of dimensionality, point-of-view, game mechanic, or game genre between the first virtual space and the second virtual space.
  11. The method of claim 9 , wherein the first virtual space is a simulated physical space that has a topography, expresses real-time interaction by the users, and includes one or more objects positioned within the topography that are capable of experiencing locomotion within the topography.
  12. The method of claim 9 , wherein the at least a portion of the information in the first character record corresponding to the first character that is common with a portion of the information in the first character record corresponding to the second character comprises information that reflects a social connection with another user initiated by the first user in the first virtual space.
  13. The method of claim 9 , wherein the at least a portion of the information in the first character record corresponding to the first character that is common with a portion of the information in the first character record corresponding to the second character comprises information that reflects the acquisition of a virtual object by the first character in the first virtual space.
  14. A set of one or more servers configured to execute instances of a plurality of different virtual spaces, the set of one or more servers comprising: one or more physical processors configured by machine-readable instructions to: execute an instance of a first virtual space, which is one of the plurality of different virtual spaces, wherein a virtual space is a simulated physical space that has a topography, expresses real-time interaction by a plurality of users, and includes one or more objects positioned within the topography that are capable of experiencing locomotion within the topography, and execute an instance of a second virtual space having one or more aspects that are different from the first virtual space;implement the instance of the first virtual space to determine a view of the first virtual space, and to generate view information that describes the view of the first virtual space for transmission to a first client that generates a display of the view of the first virtual space for a first user by assembling the view information, and implement the instance of the second virtual space to determine a view of the second virtual space, and to generate view information that describes the view of the second virtual space for transmission to a second client that generates a display of the view of the second virtual space for the first user;receive character information related to a first character and a second character that are associated with the first user from a storage component, wherein the character information related to the first character enables the one or more physical processors to manifest the first character within the first virtual space such that the first character is controllable by the first user via the first client, wherein the character information related to the second character enables the one or more physical processors to manifest the second character within the second virtual space such that the second character is controllable by the first user via the second client, wherein at least a portion of the character information related to the first character is common with a portion of the character information related to the second character;and wherein the storage component is separate and distinct from the set of one or more servers, the first client device, and the second client device.
  15. The set of one or more servers of claim 15 , wherein the one or more differences between the first character and the second character result from the one or more aspects of the first virtual space that are different from the second virtual space.
  16. The set of one or more servers of claim 16 , wherein the one or more differences between the first character and the second character result from a difference in one or more of dimensionality, point-of-view, game mechanic, or game genre between the first virtual space and the second virtual space.
  17. The set of one or more servers of claim 15 , wherein the at least a portion of the information related to the first character that is common with a portion of the information related to the second character comprises information that reflects character progress in one or both of the first virtual space or the second virtual space.
  18. The set of one or more servers of claim 15 , wherein the at least a portion of the information related to the first character that is common with a portion of the information related to the second character comprises information that reflects a social connection with another user initiated by the first user in the first virtual space.
  19. The set of one or more servers of claim 15 , wherein the at least a portion of the information related to the first character that is common with a portion of the information related to the second character comprises information that reflects the acquisition of a virtual object by the first character in the first virtual space.

Disclaimer: Data collected from the USPTO and may be malformed, incomplete, and/or otherwise inaccurate.