U.S. Pat. No. 8,409,003

GAME CONTROLLER AND GAME SYSTEM

AssigneeNintendo Co., Ltd.

Issue DateAugust 14, 2008

Illustrative Figure

Abstract

A first control unit includes a first operation data generation section for generating first operation data in accordance with a motion of a first control unit body included in the first control unit. A second control unit includes a second operation data generation section for generating second operation data in accordance with a direction input operation performed by a player. Further, one of the first control unit and the second control unit includes a transmission section for transmitting the first operation data and the second operation data to a computer at a predetermined timing.

Description

DESCRIPTION OF THE PREFERRED EMBODIMENTS With reference toFIG. 1, a game system1according to one embodiment of the present invention will be described.FIG. 1is an external view illustrating the game system1. In the following description, the game system1according to the present invention includes a stationary game apparatus. As shown inFIG. 1, the game system1includes an installation type game apparatus (herein after, referred to simply as a “game apparatus”)3, which is connected to a display (herein after, referred to as a “monitor”)2of a home-use television receiver or the like having a speaker2avia a connection cord, and a controller7for giving operation information to the game apparatus3. The game apparatus3is connected to a receiving unit6via a connection terminal. The receiving unit6receives transmission data which is wirelessly transmitted from the controller7. The controller7and the game apparatus3are connected to each other by wireless communication. On the game apparatus3, an optical disc4as an example of an exchangeable information storage medium is detachably mounted. The game apparatus3includes a power ON/OFF switch, a game process reset switch, and an OPEN switch for opening a top lid of the game apparatus3on a top main surface of the game apparatus3. When a player presses the OPEN switch, the lid is opened, so that the optical disc4can be mounted or dismounted. Further, on the game apparatus3, an external memory card5is detachably mounted when necessary. The external memory card5has a backup memory or the like mounted thereon for fixedly storing saved data or the like. The game apparatus3executes a game program or the like stored on the optical disc4and displays the result on the monitor2as a game image. The game apparatus3can also reproduce a state of a game played in the past using saved data stored in the external memory card5and display the game image on the monitor2. A player playing with the ...

DESCRIPTION OF THE PREFERRED EMBODIMENTS

With reference toFIG. 1, a game system1according to one embodiment of the present invention will be described.FIG. 1is an external view illustrating the game system1. In the following description, the game system1according to the present invention includes a stationary game apparatus.

As shown inFIG. 1, the game system1includes an installation type game apparatus (herein after, referred to simply as a “game apparatus”)3, which is connected to a display (herein after, referred to as a “monitor”)2of a home-use television receiver or the like having a speaker2avia a connection cord, and a controller7for giving operation information to the game apparatus3. The game apparatus3is connected to a receiving unit6via a connection terminal. The receiving unit6receives transmission data which is wirelessly transmitted from the controller7. The controller7and the game apparatus3are connected to each other by wireless communication. On the game apparatus3, an optical disc4as an example of an exchangeable information storage medium is detachably mounted. The game apparatus3includes a power ON/OFF switch, a game process reset switch, and an OPEN switch for opening a top lid of the game apparatus3on a top main surface of the game apparatus3. When a player presses the OPEN switch, the lid is opened, so that the optical disc4can be mounted or dismounted.

Further, on the game apparatus3, an external memory card5is detachably mounted when necessary. The external memory card5has a backup memory or the like mounted thereon for fixedly storing saved data or the like. The game apparatus3executes a game program or the like stored on the optical disc4and displays the result on the monitor2as a game image. The game apparatus3can also reproduce a state of a game played in the past using saved data stored in the external memory card5and display the game image on the monitor2. A player playing with the game apparatus3can enjoy the game by operating the controller7while watching the game image displayed on the monitor2.

The controller7wirelessly transmits the transmission data from a communication section75included therein (described later) to the game apparatus3connected to the receiving unit6, using the technology of, for example, Bluetooth (registered trademark). The controller7has two control units, a core unit70and a subunit76, connected to each other by a flexible connecting cable79. The controller7is an operation means for mainly operating a player object appearing in a game space displayed on the monitor2. The core unit70and the subunit76each includes an operation section such as a plurality of operation buttons, a key, a stick and the like. As described later in detail, the core unit70includes an imaging information calculation section74for taking an image viewed from the core unit70. As an example of an imaging target of the imaging information calculation section74, two LED modules8L and8R are provided in the vicinity of a display screen of the monitor2. The LED modules8L and8R each outputs infrared light forward from the monitor2. Although in the present embodiment the core unit70and the subunit76are connected to each other by the flexible cable, the subunit76may have a wireless unit, thereby eliminating the connecting cable79. For example, the subunit76has a Bluetooth (registered trademark) unit as the wireless unit, whereby the subunit76can transmit operation data to the core unit70.

Next, with reference toFIG. 2, a structure of the game apparatus3will be described.FIG. 2is a functional block diagram of the game apparatus3.

As shown inFIG. 2, the game apparatus3includes, for example, a RISC CPU (central processing unit)30for executing various types of programs. The CPU30executes a boot program stored in a boot ROM (not shown) to, for example, initialize memories including a main memory33, and then executes a game program stored on the optical disc4to perform game process or the like in accordance with the game program. The CPU30is connected to a GPU (Graphics Processing Unit)32, the main memory33, a DSP (Digital Signal Processor)34, and an ARAM (audio RAM)35via a memory controller31. The memory controller31is connected to a controller I/F (interface)36, a video I/F37, an external memory I/F38, an audio I/F39, and a disc I/F41via a predetermined bus. The controller I/F36, the video I/F37, the external memory I/F38, the audio I/F39and the disc I/F41are respectively connected to the receiving unit6, the monitor2, the external memory card5, the speaker2a, and a disc drive40.

The GPU32performs image processing based on an instruction from the CPU30. The GPU32includes, for example, a semiconductor chip for performing calculation process necessary for displaying 3D graphics. The GPU32performs the image process using a memory dedicated for image process (not shown) and a part of the storage area of the main memory33. The GPU32generates game image data and a movie to be displayed on the monitor2using such memories, and outputs the generated data or movie to the monitor2via the memory controller31and the video I/F37as necessary.

The main memory33is a storage area used by the CPU30, and stores a game program or the like necessary for processing performed by the CPU30as necessary. For example, the main memory33stores a game program read from the optical disc4by the CPU30, various types of data or the like. The game program, the various types of data or the like stored in the main memory33are executed by the CPU30.

The DSP34processes sound data or the like generated by the CPU30during the execution of the game program. The DSP34is connected to the ARAM35for storing the sound data or the like. The ARAM35is used when the DSP34performs a predetermined process (for example, storage of the game program or sound data already read). The DSP34reads the sound data stored in the ARAM35, and outputs the sound data to the speaker2aincluded in the monitor2via the memory controller31and the audio I/F39.

The memory controller31comprehensively controls data transmission, and is connected to the various I/Fs described above. The controller I/F36includes, for example, four controller I/Fs36a,36b,36cand36d, and communicably connects the game apparatus3to an external device which is engageable via connectors of the controller I/Fs36a,36b,36cand36d. For example, the receiving unit6is engaged with such a connector and is connected to the game apparatus3via the controller I/F36. As described above, the receiving unit6receives the transmission data from the controller7and outputs the transmission data to the CPU30via the controller I/F36. The video I/F37is connected to the monitor2. The external memory I/F38is connected to the external memory card5and is accessible to a backup memory or the like provided in the external memory card5. The audio I/F39is connected to the speaker2abuilt in the monitor2such that the sound data read by the DSP34from the ARAM35or sound data directly outputted from the disc drive40can be outputted from the speaker2a. The disc I/F41is connected to the disc drive40. The disc drive40reads data stored at a predetermined reading position of the optical disc4and outputs the data to a bus of the game apparatus3or the audio I/F39.

Next, with reference toFIGS. 3 and 4, the controller7will be described.FIG. 3is a perspective view illustrating an outer appearance of the controller7.FIG. 4is a perspective view illustrating a state of the connecting cable79of the controller7shown inFIG. 3being connected to or disconnected from the core unit70.

As shown inFIG. 3, the controller7includes the core unit70and the subunit76connected to each other by the connecting cable79. The core unit70has a housing71including a plurality of operation sections72. The subunit76has a housing77including a plurality of operation sections78. The core unit70and the subunit76are connected to each other by the connecting cable79.

As shown inFIG. 4, the connecting cable79has a connector791detachably connected to the connector73of the core unit70at one end thereof, and the connecting cable79is fixedly connected to the subunit76at the other end thereof. The connector791of the connecting cable79is engaged with the connector73provided at the rear surface of the core unit70so as to connect the core unit70and the subunit76to each other by the connecting cable79.

With reference toFIGS. 5 and 6, the core unit70will be described.FIG. 5is a perspective view of the core unit70as seen from the top rear side thereof.FIG. 6is a perspective view of the core unit70as seen from the bottom rear side thereof.

As shown inFIGS. 5 and 6, the core unit70includes the housing71formed by plastic molding or the like. The housing71has a generally parallelepiped shape extending in a longitudinal direction from front to rear. The overall size of the housing71is small enough to be held by one hand of an adult or even a child.

At the center of a front part of a top surface of the housing71, a cross key72ais provided. The cross key72ais a cross-shaped four-direction push switch. The cross key72aincludes operation portions corresponding to the four directions (front, rear, right and left) represented by arrows, which are respectively located on cross-shaped projecting portions arranged at intervals of 90 degrees. The player selects one of the front, rear, right and left directions by pressing one of the operation portions of the cross key72a. Through an operation on the cross key72a, the player can, for example, instruct a direction in which a player character or the like appearing in a virtual game world is to move or a direction in which the cursor is to move.

Although the cross key72ais an operation section for outputting an operation signal in accordance with the aforementioned direction input operation performed by the player, such an operation section may be provided in another form. For example, the cross key72amay be replaced with a composite switch including a push switch including a ring-shaped four-direction operation section and a center switch provided at the center thereof. Alternatively, the cross key72amay be replaced with an operation section which includes an inclinable stick projecting from the top surface of the housing71and outputs an operation signal in accordance with the inclining direction of the stick. Still alternatively, the cross key72amay be replaced with an operation section which includes a disc-shaped member horizontally slidable and outputs an operation signal in accordance with the sliding direction of the disc-shaped member. Still alternatively, the cross key72amay be replaced with a touchpad. Still alternatively, the cross key72amay be replaced with an operation section which includes switches representing at least four directions (front, rear, right and left) and outputs an operation signal in accordance with the switch pressed by the player.

Behind the cross key72aon the top surface of the housing71, a plurality of operation buttons72b,72c,72d,72e,72fand72gare provided. The operation buttons72b,72c,72d,72e,72fand72gare each an operation section for outputting a respective operation signal assigned to the operation buttons72b,72c,72d,72e,72for72gwhen the player presses ahead thereof. For example, the operation buttons72b,72c, and72d, are assigned with functions of an X button, a Y button, and a B button. Further, the operation buttons72e,72fand72gare assigned with functions of a select switch, a menu switch and a start switch, for example. The operation buttons72b,72c,72d,72e,72fand72gare assigned with various functions in accordance with the game program executed by the game apparatus3, but this will not be described in detail because the functions are not directly relevant to the present invention. In an exemplary arrangement shown inFIG. 5, the operation buttons72b,72cand72dare arranged in a line at the center in the front-rear direction on the top surface of the housing71. The operation buttons72e,72fand72gare arranged in a line in the left-right direction between the operation buttons72band72don the top surface of the housing71. The operation button72fhas a top surface thereof buried in the top surface of the housing71, so as not to be inadvertently pressed by the player.

In front of the cross key72aon the top surface of the housing71, an operation button72his provided. The operation button72his a power switch for remote-controlling the power of the game apparatus3to be on or off. The operation button72halso has a top surface thereof buried in the top surface of the housing71, so as not to be inadvertently pressed by the player.

Behind the operation button72con the top surface of the housing71, a plurality of LEDs702are provided. The controller7is assigned a controller type (number) so as to be distinguishable from the other controllers7. For example, the LEDs702are used for informing the player of the controller type which is currently set to controller7that he or she is using. Specifically, when the core unit70transmits the transmission data to the receiving unit6, one of the plurality of LEDs702corresponding to the controller type is lit up.

On a bottom surface of the housing71, a recessed portion is formed. As described later in detail, the recessed portion is formed at a position at which an index finger or middle finger of the player is located when the player holds the core unit70. On a rear slope surface of the recessed portion, an operation button72iis provided. The operation button72iis an operation section acting as, for example, an A button. The operation button72iis used, for example, as a trigger switch in a shooting game, or for attracting attention of a player object to a predetermined object.

On a front surface of the housing71, an image pickup element743included in the imaging information calculation section74is provided. The imaging information calculation section74is a system for analyzing image data taken by the core unit70and detecting the position of the center of gravity, the size and the like of an area having a high brightness in the image data. The imaging information calculation section74has, for example, a maximum sampling period of about 200 frames/sec., and therefore can trace and analyze even a relatively fast motion of the core unit70. The imaging information calculation section74will be described later in detail. On a rear surface of the housing71, the connector73is provided. The connector73is, for example, a 32-pin edge connector, and is used for engaging and connecting the core unit70with the connector791of the connecting cable79.

With reference toFIGS. 7A and 7B, an internal structure of the core unit70will be described.FIG. 7Ais a perspective view illustrating a state where an upper casing (a part of the housing71) of the core unit70is removed.FIG. 7Bis a perspective view illustrating a state where a lower casing (a part of the housing71) of the core unit70is removed.FIG. 7Bis a perspective view illustrating a reverse side of a substrate700shown inFIG. 7A.

As shown inFIG. 7A, the substrate700is fixed inside the housing71. On a top main surface of the substrate700, the operation buttons72a,72b,72c,72d,72e,72f,72gand72h, an acceleration sensor701, the LEDs702, a quartz oscillator703, a wireless module753, an antenna754and the like are provided. These elements are connected to a microcomputer751(seeFIG. 14) via lines (not shown) formed on the substrate700and the like. The wireless module753and the antenna754allow the core unit70to act as a wireless controller. The quartz oscillator703generates a reference clock of the microcomputer751described later.

As shown inFIG. 7B, at a front edge of a bottom main surface of the substrate700, the imaging information calculation section74is provided. The imaging information calculation section74includes an infrared filter741, a lens742, the image pickup element743and an image processing circuit744located in this order from the front surface of the core unit70on the bottom main surface of the substrate700. At a rear edge of the bottom main surface of the substrate700, the connector73is attached. The operation button72iis attached on the bottom main surface of the substrate700behind the imaging information calculation section74, and cells705are accommodated behind the operation button72i. On the bottom main surface of the substrate700between the cells705and the connector73, a vibrator704is attached. The vibrator704may be, for example, a vibration motor or a solenoid. The core unit70is vibrated by an actuation of the vibrator704, and the vibration is conveyed to the player's hand holding the core unit70. Thus, a so-called vibration-feedback game is realized.

With reference toFIGS. 8A,8B,8C and9, the subunit76will be described.FIG. 8Ais a top view of the subunit76.FIG. 8Bis a bottom view of the subunit76.FIG. 8Cis a left side view of the subunit76.FIG. 9is a perspective view of the subunit76as seen from the top front side thereof.

As shown inFIGS. 8A,8B,8C and9, the subunit76includes the housing77formed by, for example, plastic molding. The housing77extends in a longitudinal direction from front to rear, and has a streamline solid shape including a head which is a widest portion in the subunit76. The overall size of the subunit76is small enough to be held by one hand of an adult or even a child.

In the vicinity of the widest portion on the top surface of the housing77, a stick78ais provided. The stick78ais an operation section which includes an inclinable stick projecting from the top surface of the housing77and outputs an operation signal in accordance with the inclining direction of the stick. For example, a player can arbitrarily designate a direction and a position by inclining a stick tip in any direction of 360 degrees, whereby the player can instruct a direction in which a player character or the like appearing in a virtual game world is to move, or can instruct a direction in which a cursor is to move.

Although the stick78ais an operation section for outputting an operation signal in accordance with a direction input operation performed by the player as described above, such an operation section may be provided in another form. Hereinafter, with reference toFIGS. 10 to 13, a first through a fifth exemplary modifications, each of which includes the subunit76having an operation section for outputting an operation signal in accordance with the direction input operation, will be described.

As the first exemplary modification, as shown inFIG. 10, the subunit76may include a cross key78fsimilar to the cross key72aof the core unit70instead of the stick78a. As the second exemplary modification, as shown inFIG. 11, the subunit76may include a slide pad78gwhich includes a disc-shaped member horizontally slidable and outputs an operation signal in accordance with the sliding direction of the disc-shaped member, instead of the stick78a. As the third exemplary modification, as shown inFIG. 12, the subunit76may include a touch pad78hinstead of the stick78a. As the fourth exemplary modification, as shown inFIG. 13, the subunit76may include an operation section which has buttons78i,78j,78k, and78lrepresenting at least four directions (front, rear, right and left), respectively, and outputs an operation signal in accordance with the button (78i,78j,78k, or78l) pressed by a player, instead of the stick78a. As the fifth exemplary modification, the subunit76may include a composite switch including a push switch having a ring-shaped four-direction operation section and a center switch provided at the center thereof, instead of the stick78a.

Behind the stick78aon the top surface of the housing77and on the front surface of the housing77, a plurality of operation buttons78b,78c,78dand78eare provided. The operation buttons78b,78c,78dand78eare each an operation section for outputting a respective operation signal assigned to the operation buttons72b,72c,72d, and72ewhen the player presses ahead thereof. For example, the operation buttons78b,78c,78dand78eare assigned with functions of an X button, a Y button and the like. The operation buttons78b,78c,78dand78eare assigned with various functions in accordance with the game program executed by the game apparatus3, but this will not be described in detail because the functions are not directly relevant to the present invention. In the exemplary arrangement shown inFIGS. 8A,8B,8C and9, the operation buttons78band78care arranged in a line at the center in the left-right direction on the top surface of the housing77. The operation buttons78dand78eare arranged in a line in the front-rear direction on the front surface of the housing77.

Next, with reference toFIG. 14, an internal structure of the controller7will be described.FIG. 14is a block diagram illustrating the structure of the controller7.

As shown inFIG. 14, the core unit70includes the communication section75and the acceleration sensor701in addition to the aforementioned operation section72and the imaging information calculation section74.

The imaging information calculation section74includes the infrared filter741, the lens742, the image pickup element743and the image processing circuit744. The infrared filter741allows only infrared light to pass therethrough, among light incident on the front surface of the core unit70. The lens742collects the infrared light which has passed through the infrared filter741and outputs the infrared light to the image pickup element743. The image pickup element743is a solid-state imaging device such as, for example, a CMOS sensor or a CCD. The image pickup element743takes an image of the infrared light collected by the lens742. Accordingly, the image pickup element743takes an image of only the infrared light which has passed through the infrared filter741and generates image data. The image data generated by the image pickup element743is processed by the image processing circuit744. Specifically, the image processing circuit744processes the image data obtained from the image pickup element743, identifies a spot thereof having a high brightness, and outputs process result data representing the identified position coordinates and size of the area to the communication section75. The imaging information calculation section74is fixed to the housing71of the core unit70. The imaging direction of the imaging information calculation section74can be changed by changing the direction of the housing71. The housing71is connected to the subunit76by the flexible connecting cable79, and therefore the imaging direction of the imaging information calculation section74is not changed by changing the direction and position of the subunit76. As described later in detail, a signal can be obtained in accordance with the position and the motion of the core unit70based on the process result data outputted by the imaging information calculation section74.

The core unit70preferably includes a three-axis, linear acceleration sensor701that detects linear acceleration in three directions, i.e., the up/down direction, the left/right direction, and the forward/backward direction. Alternatively, a two axis linear accelerometer that only detects linear acceleration along each of the up/down and left/right directions (or other pair of directions) may be used in another embodiment depending on the type of control signals desired. As a non-limiting example, the three-axis or two-axis linear accelerometer701may be of the type available from Analog Devices, Inc. or STMicroelectronics N.V. Preferably, the acceleration sensor701is an electrostatic capacitance or capacitance-coupling type that is based on silicon micro-machined MEMS (microelectromechanical systems) technology. However, any other suitable accelerometer technology (e.g., piezoelectric type or piezoresistance type) now existing or later developed may be used to provide the three-axis or two-axis acceleration sensor701.

As one skilled in the art understands, linear accelerometers, as used in acceleration sensor701, are only capable of detecting acceleration along a straight line corresponding to each axis of the acceleration sensor. In other words, the direct output of the acceleration sensor701is limited to signals indicative of linear acceleration (static or dynamic) along each of the two or three axes thereof. As a result, the acceleration sensor701cannot directly detect movement along a non-linear (e.g. arcuate) path, rotation, rotational movement, angular displacement, tilt, position, attitude or any other physical characteristic.

However, through additional processing of the linear acceleration signals output from the acceleration sensor701, additional information relating to the core unit70can be inferred or calculated, as one skilled in the art will readily understand from the description herein. For example, by detecting static, linear acceleration (i.e., gravity), the linear acceleration output of the acceleration sensor701can be used to infer tilt of the object relative to the gravity vector by correlating tilt angles with detected linear acceleration. In this way, the acceleration sensor701can be used in combination with the micro-computer751(or another processor) to determine tilt, attitude or position of the core unit70. Similarly, various movements and/or positions of the core unit70can be calculated or inferred through processing of the linear acceleration signals generated by the acceleration sensor701when the core unit70containing the acceleration sensor701is subjected to dynamic accelerations by, for example, the hand of a user, as explained herein. In another embodiment, the acceleration sensor701may include an embedded signal processor or other type of dedicated processor for performing any desired processing of the acceleration signals output from the accelerometers therein prior to outputting signals to micro-computer751. For example, the embedded or dedicated processor could convert the detected acceleration signal to a corresponding tilt angle when the acceleration sensor is intended to detect static acceleration (i.e., gravity). Data representing the acceleration detected by the acceleration sensor701is outputted to the communication section75.

In another exemplary embodiment, the acceleration sensor701may be replaced with a gyro-sensor of any suitable technology incorporating, for example, a rotating or vibrating element. Exemplary MEMS gyro-sensors that may be used in this embodiment are available from Analog Devices, Inc. Unlike the linear acceleration sensor701, a gyro-sensor is capable of directly detecting rotation (or angular rate) around an axis defined by the gyroscopic element (or elements) therein. Thus, due to the fundamental differences between a gyro-sensor and an linear acceleration sensor, corresponding changes need to be made to the processing operations that are performed on the output signals from these devices depending on which device is selected for a particular application.

More specifically, when a tilt or inclination is calculated using a gyroscope instead of the acceleration sensor, significant changes are necessary. Specifically, when using a gyro-sensor, the value of inclination is initialized at the start of detection. Then, data on the angular velocity which is output from the gyroscope is integrated. Next, a change amount in inclination from the value of inclination previously initialized is calculated. In this case, the calculated inclination corresponds to an angle. In contrast, when an acceleration sensor is used, the inclination is calculated by comparing the value of the acceleration of gravity of each axial component with a predetermined reference. Therefore, the calculated inclination can be represented as a vector. Thus, without initialization, an absolute direction can be determined with an accelerometer. The type of the value calculated as an inclination is also very different between a gyroscope and an accelerometer; i.e., the value is an angle when a gyroscope is used and is a vector when an accelerometer is used. Therefore, when a gyroscope is used instead of an acceleration sensor or vice versa, data on inclination also needs to be processed by a predetermined conversion that takes into account the fundamental differences between these two devices. Due to the fact that the nature of gyroscopes is known to one skilled in the art, as well as the fundamental differences between linear accelerometers and gyroscopes, further details are not provided herein so as not to obscure the remainder of the disclosure. While gyro-sensors provide certain advantages due to their ability to directly detect rotation, linear acceleration sensors are generally more cost effective when used in connection with the controller applications described herein.

The communication section75includes the microcomputer751, a memory752, the wireless module753and the antenna754. The microcomputer751controls the wireless module753for wirelessly transmitting the transmission data while using the memory752as a storage area during the process.

Data from the core unit70including an operation signal (core key data) from the operation section72, acceleration signals (acceleration data) from the acceleration sensor701, and the process result data from the imaging information calculation section74are outputted to the microcomputer751. An operation signal (sub key data) from the operation section78of the subunit76is outputted to the microcomputer751via the connecting cable79. The microcomputer751temporarily stores the input data (core key data, sub key data, acceleration data, and process result data) in the memory752as the transmission data which is to be transmitted to the receiving unit6. The wireless transmission from the communication section75to the receiving unit6is performed periodically at a predetermined time interval. Since game process is generally performed at a cycle of 1/60 sec., data needs to be collected and transmitted at a cycle of a shorter time period. Specifically, the game process unit is 16.7 ms ( 1/60 sec.), and the transmission interval of the communication section75structured using the Bluetooth (registered trademark) technology is 5 ms. At the transmission timing to the receiving unit6, the microcomputer751outputs the transmission data stored in the memory752as a series of operation information to the wireless module753. The wireless module753uses, for example, the Bluetooth (registered trademark) technology to modulate the operation information onto a carrier wave of a predetermined frequency, and radiates the low power radio wave signal from the antenna754. Thus, the core key data from the operation section72included in the core unit70, the sub key data from the operation section78included in the subunit76, acceleration data from the acceleration sensor701, and the process result data from the imaging information calculation section74are modulated onto the low power radio wave signal by the wireless module753and radiated from the core unit70. The receiving unit6of the game apparatus3receives the low power radio wave signal, and the game apparatus3demodulates or decodes the low power radio wave signal to obtain the series of operation information (the core key data, the sub key data, the acceleration data, and the process result data). Based on the obtained operation information and the game program, the CPU30of the game apparatus3performs the game process. In the case where the communication section75is structured using the Bluetooth (registered trademark) technology, the communication section75can have a function of receiving transmission data which is wirelessly transmitted from other devices. The acceleration data and/or process result data are included in first operation data and the sub key data is included in the second operation data.

As shown inFIG. 15, in order to play a game using the controller7with the game system1, a player holds the core unit70with one hand (for example, a right hand) (seeFIGS. 16 and 17), and holds the subunit76with the other hand (for example, a left hand) (seeFIG. 19). The player holds the core unit70so as to point the front surface of the core unit70(that is, a side having an entrance through which light is incident on the imaging information calculation section74taking an image of the light) to the monitor2. On the other hand, two LED modules8L and8R are provided in the vicinity of the display screen of the monitor2. The LED modules8L and8R each outputs infrared light forward from the monitor2.

When a player holds the core unit70so as to point the front surface thereof to the monitor2, infrared lights outputted by the two LED modules8L and8R are incident on the imaging information calculation section74. The image pickup element743takes images of the infrared lights incident through the infrared filter741and the lens742, and the image processing circuit744processes the taken images. The imaging information calculation section74detects infrared components outputted by the LED modules8L and8R so as to obtain positions and area information of the LED modules8L and8R. Specifically, the imaging information calculation section74analyzes image data taken by the image pickup element743, eliminates images which do not represent the infrared lights outputted by the LED modules8L and8R from the area information, and identifies points each having a high brightness as positions of the LED modules8L and8R. The imaging information calculation section74obtains positions coordinates, coordinates of the center of gravity, and the like of each of the identified points having the high brightness and outputs the same as the process result data. When such process result data is transmitted to the game apparatus3, the game apparatus3can obtain, based on the position coordinates and the coordinates of the center of gravity, operation signals relating to the motion, posture, position and the like of the imaging information calculation section74, that is, the core unit70, with respect to the LED modules8L and8R. Specifically, the position having a high brightness in the image obtained through the communication section75is changed in accordance with the motion of the core unit70, and therefore a direction input or coordinate input is performed in accordance with the position having the high brightness being changed, thereby enabling a direction input or a coordinate input to be performed along the moving direction of the core unit70.

Thus, the imaging information calculation section74of the core unit70takes images of stationary markers (infrared lights from the two LED modules8L and8R in the present embodiment), and therefore the game apparatus3can use the process result data relating to the motion, posture, position and the like of the core unit70in the game process, whereby an operation input, which is different from an input made by pressing an operation button or using an operation key, is further intuitively performed. As described above, since the markers are provided in the vicinity of the display screen of the monitor2, the motion, posture, position and the like of the core unit70with respect to the display screen of the monitor2can be easily calculated based on positions from the markers. That is, the process result data used for obtaining the motion, posture, position and the like of the core unit70can be used as operation input immediately applied to the display screen of the monitor2.

With reference toFIGS. 16 and 17, a state of a player holding the core unit70with one hand will be described.FIG. 16shows an exemplary state of a player holding the core unit70with a right hand as seen from the front surface side of the core unit70.FIG. 17shows an exemplary state of a player holding the core unit70with a right hand as seen from the left side of the core unit70.

As shown inFIGS. 16 and 17, the overall size of the core unit70is small enough to be held by one hand of an adult or even a child. When the player puts a thumb on the top surface of the core unit70(for example, near the cross key72a), and puts an index finger in the recessed portion on the bottom surface of the core unit70(for example, near the operation button72i), the light entrance of the imaging information calculation section74on the front surface of the core unit70is exposed forward to the player. It should be understood that also when the player holds the core unit70with a left hand, the holding state is the same as that described for the right hand.

Thus, the core unit70allows a player to easily operate the operation section72such as the cross key72aor the operation button72iwhile holding the core unit70with one hand. Further, when the player holds the core unit70with one hand, the light entrance of the imaging information calculation section74on the front surface of the core unit70is exposed, whereby the light entrance can easily receive infrared lights from the aforementioned two LED modules8L and8R. That is, the player can hold the core unit70with one hand without preventing the imaging information calculation section74from functioning. That is, when the player moves his or her hand holding the core unit70with respect to the display screen, the core unit70can further perform an operation input enabling a motion of the player's hand to directly act on the display screen.

As shown inFIG. 18, the LED modules8L and8R each has a viewing angle θ1. The image pickup element743has a viewing angle θ2. For example, the viewing angle θ1of the LED modules8L and8R is 34 degrees (half-value angle), and the viewing angle θ2of the image pickup element743is 41 degrees. When both the LED modules8L and8R are in the viewing angle θ2of the image pickup element743, and the image pickup element743is in the viewing angle θ1of the LED module8L and the viewing angle θ1of the LED module8R, the game apparatus3determines a position of the core unit70using positional information relating to the point having high brightness of the two LED modules8L and8R.

When either the LED module8L or LED module8R is in the viewing angle θ2of the image pickup element743, or when the image pickup element743is in either the viewing angle θ1of the LED module8L or the viewing angle θ1of the LED module8R, the game apparatus3determines a position of the core unit70using the positional information relating to the point having high brightness of the LED module8L or the LED module8R.

Next, with reference toFIG. 19, a state of a player holding the subunit76with one hand will be described.FIG. 19shows an exemplary state of a player holding the subunit76with a left hand as seen from the right side of the subunit76.

As shown inFIG. 19, the overall size of the subunit76is small enough to be held by one hand of an adult or even a child. For example, a player can put a thumb on the top surface of the subunit76(for example, near the stick78a), put an index finger on the front surface of the subunit76(for example, near the operation buttons78dand78e), and put a middle finger, a ring finger and a little finger on the bottom surface of the subunit76so as to hold the subunit76. It should be understood that also when the player holds the subunit76with a right hand, the holding state is similar to that described for the left hand. Thus, the subunit76allows the player to easily operate the operation section78such as the stick78aand the operation buttons78dand78ewhile holding the subunit76with one hand.

Here, an exemplary game played using the aforementioned controller7will be described. As a first example, a shooting game played using the controller7will be described.FIG. 20is a diagram illustrating an exemplary game image displayed on the monitor2when the game apparatus3executes the shooting game.

As shown inFIG. 20, a portion of a three-dimensional virtual game space S is displayed on the display screen of the monitor2. As a game object acting in accordance with an operation of the controller7, a portion of the player character P and a portion of a gun G held by the player character P are displayed on the display screen. Moreover, the virtual game space S displayed on the display screen represents a field of front vision of the player character P, and for example an opponent character E is displayed as a shooting target inFIG. 20. A target indicating a position at which the player character P shoots the gun G is displayed on the display screen as the target cursor T.

In the shooting game having such a game image displayed on the monitor2, a player operates the core unit70with one hand and operates the subunit76with the other hand as shown inFIG. 15so as to play the game. For example, when the player inclines the stick78a(seeFIGS. 8A,8B,8C and9) on the subunit76, the player character P is moved in the virtual game space S in accordance with the inclining direction. Further, when the player moves his or her hand holding the core unit70with respect to the display screen, the target cursor T is moved in accordance with the motion, posture, position and the like of the core unit70with respect to the monitor2(LED modules8L and8R). When the player presses the operation button72i(shown inFIG. 6) on the core unit70, the player character P shoots the gun G at the target cursor T.

That is, while the player uses the stick78aon the subunit76so as to instruct the player character P to move, the player can operate the core unit70as if the core unit70is a gun for the shooting game, thereby enhancing enjoyment in playing a shooting game. The player can perform an operation of moving the player character P and an operation of moving the target cursor T by using respective units held by different hands, whereby the player can perform the respective operations as independent ones. For example, since the virtual game space S displayed on the display screen is changed in accordance with the movement of the player character P, it is sometimes difficult to keep the target positioned near a position observed by the player in the virtual game space S because, for example, the player may be paying attention to the opponent character E suddenly jumping into the virtual game space S. However, while the player is moving the player character P with one hand (for example, a thumb of a left hand), the player can control a motion of the arm (for example, a right arm) which is not used for moving the player character P such that the core unit70has its front surface pointed to the observed position, thereby substantially enhancing flexibility for operating the controller7and increasing the reality of the shooting game. Further, in order to move the target cursor T, the player moves the controller. However, the operation of moving the controller does not hinder the player from performing a direction instruction operation for moving the player character P, thereby enabling the player to stably perform the two direction instruction operations. That is, by using the controller7, the player can freely use his or her left and right hands and can perform a new operation having increased flexibility, which cannot be achieved using a physically single controller.

In a second example, a player inclines the stick78aon the subunit76so as to move the player character P in the virtual game space S in accordance with the inclining direction as in the first example. The player moves a hand holding the core unit70with respect to the display screen so as to move a sight point of a virtual camera in accordance with a position of the core unit70with respect to the monitor2(LED modules8L and8R). These operations allow the player to observe a position to which the core unit70is pointed in the virtual game space S while operating the stick78aon the subunit76so as to instruct the player character P to move.

In the above description, the controller7and the game apparatus3are connected to each other by wireless communication. However, the controller7and the game apparatus3may be electrically connected to each other by a cable. In this case, the cable connected to the core unit70is connected to a connection terminal of the game apparatus3.

Moreover, in the present embodiment, only the core unit70among the core unit70and the subunit76of the controller7has the communication section75. However, the subunit76may have the communication section for wirelessly transmitting the transmission data to the receiving unit6. Further, both the core unit70and the subunit76may have the respective communication sections. For example, the respective communication sections included in the core unit70and the subunit76may wirelessly transmit the transmission data to the receiving unit6, or the communication section of the subunit76may wirelessly transmit the transmission data to the communication section75of the core unit70, and the communication section75of the core unit70may wirelessly transmit, to the receiving unit6, the received transmission data from the subunit76and the transmission data of the core unit70. In these cases, the connecting cable79for electrically connecting between the core unit70and the subunit76can be eliminated.

In the above description, the receiving unit6connected to the connection terminal of the game apparatus3is used as a receiving means for receiving transmission data which is wirelessly transmitted from the controller7. Alternatively, the receiving means may be a receiving module built in the game apparatus3. In this case, the transmission data received by the receiving module is outputted to the CPU30via a predetermined bus.

Although in the present embodiment the imaging information calculation section74included in the core unit70is described as an example of a determining section for outputting a signal (process result data) in accordance with a motion of the core unit70body, the imaging information calculation section74may be provided in another form. For example, the core unit70may include the acceleration sensor701as described above, or may include a gyro sensor. The acceleration sensor or the gyro sensor can be used to determine a motion or posture of the core unit70, and, therefore, can be used as a determining section for outputting a signal in accordance with the motion of the core unit70body using the detection signal for the motion or posture. In this case, the imaging information calculation section74may be eliminated from the core unit70, or sensor and the imaging information calculation section can be used in combination.

Further, although in the present embodiment only the core unit70includes the imaging information calculation section74, the subunit76may also include a similar imaging information calculation section.

In the present embodiment, image data taken by the image pickup element743is analyzed so as to obtain position coordinates and the like of an image of infrared lights from the LED modules8L and8R, and the core unit70generates process result data from the obtained coordinates and the like and transmits the process result data to the game apparatus3. However, the core unit70may transmit data obtained in another process step to the game apparatus3. For example, the core unit70transmits to the game apparatus3image data taken by the image pickup element743, and the CPU30may perform the aforementioned analysis so as to obtain process result data. In this case, the image processing circuit744can be eliminated from the core unit70. Alternatively, the core unit70may transmit, to the game apparatus3, the image data having been analyzed halfway. For example, the core unit70transmits to the game apparatus3data indicating a brightness, a position, an area size and the like obtained from the image data, and the CPU30may perform the remaining analysis so as to obtain process result data.

Although in the present embodiment infrared lights from the two LED modules8L and8R are used as imaging targets of the imaging information calculation section74in the core unit70, the imaging target is not restricted thereto. For example, infrared light from one LED module or infrared lights from at least three LED modules provided in the vicinity of the monitor2may be used as the imaging target of the imaging information calculation section74. Alternatively, the display screen of the monitor2or another emitter (room light or the like) can be used as the imaging target of the imaging information calculation section74. When the position of the core unit70with respect to the display screen is calculated based on the positional relationship between the imaging target and the display screen of the monitor2, various emitters can be used as the imaging target of the imaging information calculation section74.

The aforementioned shapes of the core unit70and the subunit76are merely examples. Further, the shape, the number, setting position and the like of each of the operation section72of the core unit70and the operation section78of the subunit76are merely examples. Needless to say, even when the shape, the number, the setting position and the like of each of the core unit70, the subunit76, the operation section72, and the operation section78are different from those described in the embodiment, the present invention can be realized. Further, the imaging information calculation section74(light entrance of the imaging information calculation section74) of the core unit70may not be positioned on the front surface of the housing71. The imaging information calculation section74may be provided on another surface at which light can be received from the exterior of the housing71.

Thus, the controller of the present invention allows a player to operate both the core unit70and the subunit76included therein so as to enjoy a game. The core unit70has a function of outputting a signal in accordance with motion of the unit body including the imaging information calculation section74and the accelerator sensor701. The subunit76has a function of outputting a signal in accordance with a direction input operation performed by the player. For example, when used is a controller into which the core unit70and the subunit76are integrated, the whole controller has to be moved so as to output a signal in accordance with the motion of the unit body, thereby exerting some influence on the direction input operation. Further, the integration of the core unit70and the subunit76causes the opposite influence, that is, flexibility, which is realized by separation between the core unit70and the subunit76, is substantially reduced. Therefore, the core unit70and the subunit76can be separated into a right unit and a left unit as in the case of a conventional controller for the game apparatus, and simultaneously the core unit70and the subunit76allow the player to freely use his or her right and left hands, thereby providing the player with new operation, which cannot be anticipated by the integrated controller. Further, the controller can be operated with substantially enhanced flexibility, thereby providing a player with a game operation having increased reality.

The game controller and the game system according to the present invention can realize an operation having increased flexibility, and are useful as a game controller which includes two independent units and is operated by a player holding the two independent units, a game system including the game controller, and the like.

While the invention has been described in detail, the foregoing description is in all aspects illustrative and not restrictive. It is understood that numerous other modifications and variations can be devised without departing from the scope of the invention.

Claims

  1. A handheld controller arrangement for transmitting operation data to a computing device executing a program that displays information on a display, the handheld controller arrangement comprising: a first handheld unit including a first handheld body configured to be held in and supported by a user's hand in free space;a second handheld unit including a second handheld body configured to be held in and supported by a user's other hand in free space;a flexible cable for electrically connecting the first handheld unit with the second handheld unit, the first handheld unit including a first operation data generation section generating first operation data in accordance with motion of the first handheld body;the second handheld unit including a second operation data generation section generating second operation data in accordance with a direction input operation performed by the user;and one of the first handheld unit and the second handheld unit further including a wireless transmission section for wirelessly transmitting the first operation data and the second operation data to the computing device at a predetermined timing, wherein the first operation data generation section includes an image pickup section, fixed to the first handheld body, configured to take an image along a predetermined direction from the first handheld body, and output, as a part of the first operation data, an image taken by the image pickup section and/or a result of performing a predetermined calculation on the image taken by the image pickup section.
  1. The handheld controller arrangement according to claim 1 , wherein the first operation data generation section further includes a positional information calculation section for calculating positional information indicating a center of gravity, in the image taken by the image pickup section, of at least one marker image which is included in the taken image and is used as an imaging target, and outputting the positional information as part of the first operation data.
  2. The handheld controller arrangement according to claim 1 , wherein the transmission section wirelessly transmits the first operation data and the second operation data to the computing device.
  3. The handheld controller arrangement according to claim 1 , wherein the first operation data generation section has an acceleration sensor and/or a gyro sensor included in the first handheld body, and outputs data generated by the acceleration sensor and/or the gyro sensor as the first operation data.
  4. The handheld controller arrangement according to claim 1 , wherein the flexible cable is detachably connected to at least the first handheld body, and the transmission section is included in the first handheld body.
  5. The handheld controller arrangement according to claim 1 , wherein the transmission section collects and transmits to the computing device the first operation data and the second operation data at intervals shorter than 1/60 second.
  6. The handheld controller arrangement according to claim 1 , wherein the second operation data generation section includes a stick which has a tip projecting from the second handheld body the stick being inclinably movable relative to the second handheld body, and outputs data obtained in accordance with an inclining direction of the stick as the second operation data.
  7. The handheld controller arrangement according to claim 1 , wherein the second operation data generation section includes an operation button which has operation portions representing at least four directions and which is able to be pushed, by the operation portions, into the second handheld body, and outputs, as the second operation data, data corresponding to the operation portion at which the operation button is pushed.
  8. The handheld controller arrangement according to claim 1 , wherein the second operation data generation section includes a sliding member which has a top surface exposed from the second handheld body and which is horizontally movable on the second handheld body, and outputs data obtained in accordance with a horizontal moving direction of the sliding member as the second operation data.
  9. The handheld controller arrangement according to claim 1 , wherein the second handheld body has an outer surface and the second operation data generation section includes a touch pad on the outer surface of the second handheld body and outputs, as the second operation data, data obtained in accordance with a position on the touch pad at which the touch pad is touched.
  10. The handheld controller arrangement according to claim 1 , wherein the second operation data generation section includes at least four operation buttons which are able to be pushed into the second handheld body, and outputs data obtained in accordance with the pushed operation button as the second operation data.
  11. The handheld controller arrangement of claim 1 further including a first connector terminating an end of the flexible cable, and a second connector disposed in the first and/or second handheld units, the first and second connectors detachably mating to establish connectivity between the first and second handheld units while optionally detaching to permit at least one of the first and second handheld units to be operated independently of the other to transmit operation data to the computing device.
  12. The handheld controller arrangement of claim 12 wherein one of the first and second handheld units powers the other via the flexible cable.
  13. The handheld controller arrangement of claim 1 wherein the first handheld body has a generally parallelepiped shape extending in a longitudinal direction from front to rear and has an overall size small enough to be held in one hand.
  14. The handheld controller arrangement of claim 1 wherein the first handheld body has disposed thereon a cross-shaped four-direction push-switch allowing the user to select one of four directions 90° apart.
  15. The handheld controller arrangement of claim 1 wherein the first handheld unit further comprises a push switch including a ring-shaped four-direction operation section and a center switch provided at the center thereof.
  16. The handheld controller arrangement of claim 1 wherein the first handheld unit further comprises an inclinable stick projecting from the first handheld body, the inclinable stick outputting an operation signal in accordance with an inclining direction of the stick.
  17. The handheld controller arrangement of claim 1 wherein the first handheld unit further comprises a disc-shaped member horizontally slidable relative to the first handheld body, the disc-shaped member outputting an operation signal in accordance with the sliding direction of the disc-shaped member.
  18. The handheld controller arrangement of claim 1 wherein the first handheld unit further comprises plural switches representing at least four directions and which outputs a directional operation signal in accordance with a switch pressed by a user.
  19. The handheld controller arrangement of claim 1 wherein the first handheld body comprises a top surface, and the first handheld unit comprises an operation button recessed in the top surface so as not to be inadvertently pressed.
  20. The handheld controller arrangement of claim 1 wherein the first handheld unit includes a plurality of light emitting diodes which provide a visual indication distinguishing the first handheld unit from other handheld units being used simultaneously with the computing device.
  21. The handheld controller arrangement of claim 1 wherein the first handheld body comprises a recessed portion formed at a position at which an index or middle finger of the user is disposed when the user holds the first handheld body.
  22. The handheld controller arrangement of claim 22 wherein the first handheld body further comprises an operation button disposed on a rear slope of the recessed portion.
  23. The handheld controller arrangement of claim 1 wherein the first handheld unit includes a first microcomputer, the second handheld unit includes a second microcomputer, and the first and second microcomputers inter-communicate via the flexible cable.
  24. The handheld controller arrangement of claim 1 wherein the first handheld unit further comprises a vibration generator that when activated vibrates the first handheld body to convey the vibration to the user's hand.
  25. The handheld controller arrangement of claim 1 wherein the second handheld body has a streamlined solid shape including a head which is a widest portion on the second handheld unit, and which has an overall size that is small enough to be held in one hand.
  26. The handheld controller arrangement of claim 1 wherein the second handheld unit includes an inclinable stick projecting from a top surface thereof, the stick enabling a user to arbitrarily designate a direction and position by inclining the stick in any direction of 360 degrees.
  27. The handheld controller arrangement of claim 1 wherein the first operation data generation section comprises a gyroscope that senses an angular velocity that can be used to derive an inclination of the first handheld body.
  28. The handheld controller arrangement of claim 1 wherein the first handheld unit further comprises an infrared sensor that images plural infrared markers to indicate a pointing direction of the first handheld body.
  29. The handheld controller arrangement of claim 1 wherein the user is enabled to point with the first handheld unit and instructs an object to move with the second handheld unit to provide two-handed operation.
  30. The handheld controller arrangement of claim 1 wherein the first and second handheld units held in different hands allow the user to move about, use motion to provide operation data, and provide different inputs with different hands.

Disclaimer: Data collected from the USPTO and may be malformed, incomplete, and/or otherwise inaccurate.