US20070211047A1 - Persistent authenticating system and method to map real world object presence into virtual world object awareness - Google Patents
Persistent authenticating system and method to map real world object presence into virtual world object awareness Download PDFInfo
- Publication number
- US20070211047A1 US20070211047A1 US11/372,327 US37232706A US2007211047A1 US 20070211047 A1 US20070211047 A1 US 20070211047A1 US 37232706 A US37232706 A US 37232706A US 2007211047 A1 US2007211047 A1 US 2007211047A1
- Authority
- US
- United States
- Prior art keywords
- virtual
- real world
- detection
- objects
- virtual object
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 37
- 230000002085 persistent effect Effects 0.000 title abstract description 23
- 238000001514 detection method Methods 0.000 claims abstract description 376
- 230000007246 mechanism Effects 0.000 claims abstract description 119
- 238000012545 processing Methods 0.000 claims description 43
- 238000009877 rendering Methods 0.000 claims description 35
- 230000015654 memory Effects 0.000 claims description 24
- 230000004044 response Effects 0.000 claims description 19
- 238000004590 computer program Methods 0.000 claims description 16
- 230000000007 visual effect Effects 0.000 claims description 10
- 230000003213 activating effect Effects 0.000 claims description 9
- 238000012544 monitoring process Methods 0.000 claims description 4
- 230000002596 correlated effect Effects 0.000 abstract description 5
- 230000000875 corresponding effect Effects 0.000 description 58
- 230000009471 action Effects 0.000 description 36
- 230000002688 persistence Effects 0.000 description 19
- 238000013507 mapping Methods 0.000 description 17
- 238000010586 diagram Methods 0.000 description 13
- 238000004891 communication Methods 0.000 description 8
- 230000003993 interaction Effects 0.000 description 8
- 230000006870 function Effects 0.000 description 7
- 230000008439 repair process Effects 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 5
- 230000002093 peripheral effect Effects 0.000 description 4
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 230000004913 activation Effects 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 230000007123 defense Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 241000600039 Chromis punctipinnis Species 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F1/00—Card games
- A63F1/04—Card games combined with other games
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/60—Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor
- A63F13/65—Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor automatically by game devices or servers from real world data, e.g. measurement in live racing competition
-
- A63F13/12—
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/30—Interconnection arrangements between game servers and game devices; Interconnection arrangements between game devices; Interconnection arrangements between game servers
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/70—Game security or game management aspects
- A63F13/71—Game security or game management aspects using secure communication between game devices and game servers, e.g. by encrypting game data or authenticating players
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F3/00—Board games; Raffle games
- A63F3/00643—Electric board games; Electric features of board games
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/039—Accessories therefor, e.g. mouse pads
- G06F3/0393—Accessories for touch pads or touch screens, e.g. mechanical guides added to touch screens for drawing straight lines, hard keys overlaying touch screens or touch pads
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04842—Selection of displayed objects or displayed text elements
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F3/00—Board games; Raffle games
- A63F3/00643—Electric board games; Electric features of board games
- A63F2003/00662—Electric board games; Electric features of board games with an electric sensor for playing pieces
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F9/00—Games not otherwise provided for
- A63F9/24—Electric games; Games using electronic circuits not otherwise provided for
- A63F2009/2401—Detail of input, input devices
- A63F2009/2411—Input form cards, tapes, discs
- A63F2009/2419—Optical
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/40—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterised by details of platform network
- A63F2300/401—Secure communication, e.g. using encryption or authentication
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F2300/00—Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
- A63F2300/60—Methods for processing data by generating or executing the game program
- A63F2300/69—Involving elements of the real world in the game world, e.g. measurement in live races, real video
Definitions
- the present application relates generally to an improved data processing system and method. More specifically, the present application is directed to a persistent authenticating system and method to map real world object presence into virtual world object awareness.
- the ability to scan objects for a particular identification signature of the object is generally known in the art.
- the use of barcodes, dot codes, radio frequency identifier (RFID) tags, and the like, to provide an identification of an object associated with these scannable identifier devices is generally known in the art.
- RFID radio frequency identifier
- the use of such devices, such as the RFID tags, is only expected to increase in the near future.
- dot codes are represented by the E-READERTM plug-in device for the Nintendo GameBoyTM portable gaming system.
- E-READERTM dot codes may be provided on trading cards that, when scanned by the E-READERTM plug-in device unlock special features in software including mini-games, card statistics and data, special items, and the like. Once the trading card is scanned, the special features are unlocked and the trading card need not be present for the unlocked features to be utilized. In fact, the E-READERTM and GameBoyTM system may be turned off and the unlocked features may again be accessed when the system is turned back on due to a storage in the E-READERTM device.
- the unlocked features do not represent real-world items that are present at the location of the E-READERTM device. Moreover, the unlocked features do not represent how to use such a real-world item, how the real-world item's usage is persisted, or the like, so that the real-world item may be accurately modeled and integrated into the computer software. To the contrary, the trading cards used with the E-READERTM device merely serve as a recordable medium upon which the dot-codes are recorded so that they may be read by the E-READERTM device.
- the illustrative embodiments provide a persistent authenticating system and method to map real world object presence into virtual world object awareness. More specifically, the illustrative embodiments provide a mechanism by which the presence of a real world object is detected and, while the real world object's presence is continued to be detected, it is made available for use in a virtual world environment. Moreover, the detection of the presence of the real world object provides an identifier of the object which may be correlated with information regarding how to represent the real world object in the virtual world environment, how the object may be utilized in relation to other objects in the virtual world environment such that the real world object is modeled in the virtual world environment, and the like. Furthermore, the detection of multiple real world objects may be performed and identification of each of the multiple real world objects may be used to determine how these real world objects may be utilized together in the virtual world environment.
- the illustrative embodiments described herein are especially well suited for use with persistent virtual world type computer games. While the illustrative embodiments will be described in terms of computer games and persistent virtual world computer games, it should be appreciated that the illustrative embodiments are not limited to being applicable to only computer games. To the contrary, the illustrative embodiments may be used with any application in which real world objects may be rendered as virtual objects in a virtual environment.
- a computing device having an associated detection mechanism such as a detection pad upon which real world objects may be placed.
- the computing device executes operations for monitoring the detection pad to detect when one or more real world objects are placed on the detection pad or within a detection area of the detection pad.
- the operations may continue to monitor the detection pad continuously or periodically to determine a current set of one or more real world objects that are within the detection area of the detection pad and compare the current set of real world objects to a previously detected set of real world objects to determine which objects have persistent presence and which objects are no longer within the detection area.
- Only real world objects that are present within the detection area of the detection pad may be modeled and represented in the virtual world environment. Thus, when a real world object is removed from the detection area of the detection pad, the object's representation in the virtual world environment is removed. In this way, the persistent presence of a real world object as detected by the detection pad is mapped into a virtual world object awareness.
- the real world objects preferably are objects that are to be represented by similar virtual objects that are modeled in a virtual world environment within the computing device.
- the real world objects may be toys, e.g., action figures, toy vehicles, and the like, toy accessories, e.g., toy weapons, toy armor, and the like, clothing items, household items, pieces of a product for assembly, or any other type of objects that may be represented as virtual objects in the virtual world environment and used within the virtual world environment.
- the real world object e.g., the action figure, toy vehicle, toy weapon, etc.
- the virtual world object is represented in the virtual world in such a manner that is consistent with the usage of the object in the real world.
- the detection of the toy sword in the detection area of the detection pad causes a corresponding virtual sword to be represented in the virtual world environment.
- the virtual sword is represented in the virtual environment in a manner consistent with the usage of a sword in the real world.
- the virtual sword may be represented as being in a free hand of the avatar.
- the virtual sword may be represented as being sheathed and hanging on the side of the user's avatar.
- the representation of the virtual sword in the virtual world environment may be maintained as long as the toy sword is detected as being present in the detection area of the detection pad. If the user were to remove the toy sword from the detection area of the detection pad, when the absence of the toy sword is detected by the detection pad, the virtual world representation of the toy sword is discontinued.
- the detection pad may detect multiple real world objects as being present within the detection area of the detection pad.
- the identification of the various real world objects in the detection area is used to retrieve information regarding the real world objects and how to represent the real world objects as virtual objects in a virtual world environment.
- the interaction between the identified real world objects may be determined using a rules database.
- the real world objects on a detection pad comprise an action figure representing a knight and accessory objects that comprise a toy sword, a toy helmet, and a toy shield.
- Information corresponding to each of these detected items may be retrieved from a database associated with an application running in the computer system. This information includes an identification of how each item is utilized in the real world and how such usage is mapped to the virtual world environment.
- the toy sword is used in the hand of an action figure as a weapon and has certain attack and defense characteristics.
- a virtual object corresponding to the toy sword may be represented in the hand or sheath of an avatar resembling the action figure that is present on the detection pad.
- the toy shield may be represented as a virtual shield object in a hand of the virtual action figure, i.e. the avatar, or on the back of the virtual action figure in the virtual world environment.
- the toy helmet is used on the head of the action figure and thus, is represented in the virtual world environment as a helmet object placed on the head of the virtual action figure.
- the operations in the computing device may monitor a condition of the virtual world object representation of the real world object within the virtual world environment and, based on the condition of the virtual world object representation of the real world object, may disable the user's ability to use the real world object to generate a virtual world representation of the real world object.
- the real world object were a toy sword and the user, in a previous session in the virtual world environment, performed actions that resulted in his/her avatar's virtual sword, which corresponds to then real world toy sword, being damaged or broken, then the user's ability to place the toy sword on the detection pad and have a corresponding sword represented in the virtual world environment may be discontinued until the user performs appropriate actions in the virtual world environment, e.g., repairs the virtual representation of the sword in the virtual world environment.
- a persistent state of the virtual representation of the real world object may be maintained in the virtual world environment such that each time a real world object is placed in the detection area of the detection pad, the persistent state of the virtual object corresponding to that real world object may be retrieved and used to render the virtual object in the virtual world environment. For example, using the damaged sword example discussed above, rather than discontinuing the ability of the user to generate a virtual object corresponding to the real world toy sword object on the detection pad, the previous state of the virtual sword may be retrieved and used to render the virtual sword representation of the real world toy sword.
- the virtual sword object rendered in response to detecting the real world toy sword will be rendered as having the same damaged state as when the user last used that real world toy sword object with the virtual world environment.
- the detection pad may define a grid array that allows for the ability to associate coordinates with items placed within the detection field. Furthermore, it is possible to activate, highlight, etc. specific patterns of the grid to indicate active sensor regions, relevant to the specific virtual world environment, in which real world objects may be placed and their presence detected using the mechanisms of the illustrative embodiments. For example, a human pattern may be highlighted or outlined for an adventure or role playing game virtual world environment. For a trading card based game, card placement blocks may be highlighted/activated, signifying different play areas in which real world objects may be placed.
- the size of the detection pad is not limited by the illustrative embodiments described herein. To the contrary, the size of the detection pad may be determined based on the particular implementation and intended use of the detection pad. For example, if the detection pad will be used primarily with a strategy game virtual world environment, in which a large number of pieces may be placed on a large game board, the game pad may be large enough to cover a typical dining room table or coffee table. On the other hand, if the detection pad is to be used with a card game virtual environment, then the detection pad may be quite small in size relative to a dining room table or coffee table.
- a product may be purchased that requires assembly.
- the pieces of the product may have identifier tags embedded or otherwise attached to them, and may be deposited on the presence detection pad.
- An assembly program with which the detection pad interfaces may be run on an associated computing device and, for each step of the assembly instructions, the grid location corresponding to the component to be assembled/used next in the series of assembly steps may be highlighted.
- the assembly program may detect the presence of the pieces via the detection pad and represent the pieces in a virtual world environment in which the various steps for assembly of the product are virtually represented.
- the assembly program may perform a check of the complete inventory of items against the assembly program before assembly begins to ensure that all the necessary pieces are present.
- images and/or topological information may be overlaid on the detection pad surface to convey detail that enhances the application program. For example, in addition to highlighting/activating portions of the detection pad based on the particular programs with which the detection pad interfaces, images may be displayed within these highlighted/activated areas that depict the types of real world objects that should be placed in these highlighted/activated areas.
- a method, in a data processing system, for rendering virtual world objects based on a detected presence of a real world object may comprise detecting a presence of a real world object within a detection area of a detection device and generating, in response to detecting the presence of the real world object, a virtual object in a virtual world environment rendered in the data processing system.
- the virtual object may correspond to the real world object and may be rendered in the virtual world environment only while the presence of the real world object continues to be detected within the detection area of the detection device.
- the detection device may be a detection pad having one or more detectors integrated therein and upon which real world objects may be placed.
- the real world object may be a toy and the virtual world environment may be a virtual computer gaming environment, for example.
- the method may detect the presence of the real world object by detecting an identifier mechanism associated with the real world object.
- the he identifier mechanism may have associated identifier information.
- the identifier information may be correlated with virtual object characteristic information identifying a manner by which to represent the real world object in the virtual world environment.
- the virtual object characteristic information may comprise at least one of information identifying how the virtual object may be utilized in relation to other virtual objects in the virtual world environment such that the real world object is modeled in the virtual world environment, or characteristics of the virtual object with regard to particular applications in which the virtual object will be used in the virtual world environment.
- the identifier mechanism may be either affixed to or integrated with the real world object.
- the identifier mechanism may be a radio frequency identification (RFID) mechanism.
- RFID radio frequency identification
- the detection mechanism thus, may comprise one or more RFID detectors.
- the method may further comprise monitoring a condition of the virtual object corresponding to the detected real world object within the virtual world environment.
- the virtual object condition information identifying the condition of the virtual object may be stored in association with the identifier of the real world object.
- Correlating the identifier information associated with the detected identifier mechanism with virtual object characteristic information may further comprise retrieving the virtual object condition information corresponding to the identifier information and rendering the virtual object corresponding to the identifier information such that the virtual object has a condition corresponding to the virtual object condition information.
- Rendering the virtual object corresponding to the identifier information such that the virtual object has a condition corresponding to the virtual object condition information may comprise disabling rendering of the virtual object corresponding to the identifier information if the condition meets a predetermined criteria.
- the method may further comprise detecting removal of the real world object from the detection area of the detection mechanism. Rendering of the virtual object corresponding to the real world object in the virtual world environment may be automatically discontinued in response to detection of the removal of the real world object from the detection area.
- the method may detect a presence of a real world object within a detection area of a detection device by detecting a presence of a plurality of real world objects within the detection area and correlating identifiers associated with each of the real world objects with virtual object characteristic information for rendering the real world objects as virtual objects in the virtual world environment. At least one way in which at least two of the plurality of real world objects are utilized together may be determined based on the virtual object characteristic information. The virtual objects corresponding to the detected plurality of real world objects may be rendered within the virtual world environment based on the virtual object characteristic information and the determined at least one way in which at least two of the plurality of real world objects are utilized together such that at least two of the virtual objects are rendered as being utilized together.
- the method may be implemented in a distributed data processing system.
- a first computing device may detect the presence of the real world object and may transmit information corresponding to the detection of the real world object to a second computing device in the data processing system.
- the second computing device may generate, in response to the detection of the presence of the real world object, the virtual object in the virtual world environment.
- the method may further comprise activating one or more portions of the detection mechanism.
- Real world objects may only be detectable by the detection mechanism when the real world objects are placed within a detection area corresponding to the activated one or more portions of the detection mechanism.
- the one or more portions of the detection mechanism that are activated may be configured to represent a type of real world object that is to be placed within the detection area of the one or more portions of the detection mechanism.
- the method may further comprise outputting, via the detection mechanism, an output to aid a user in placing real world objects within the detection areas of the activated one or more portions of the detection mechanism.
- the detection mechanism may be comprised of a grid of detectors and activating one or more portions of the detection mechanism may comprise activating one or more grid portions. Outputting an output to aid the user may comprise outputting a visual output that highlights the activated one or more grid portions.
- a computer program product comprising a computer useable medium having a computer readable program.
- the computer readable program when executed on a computing device, causes the computing device to perform various ones, and combinations of, the operations previously described above with regard to the method illustrative embodiments.
- an apparatus for rendering virtual world objects based on a detected presence of a real world object may comprise a computing device and a detection device coupled to the computing device.
- the detection device may detect a presence of a real world object within a detection area of the detection device.
- the computing device may comprise a processor and a memory, and the memory may comprise instructions. When the instructions are executed by the processor, the instructions may cause the processor to perform various ones, and combinations of, the operations previously described above with regard to the method illustrative embodiments.
- a system for rendering virtual world objects based on a detected presence of a real world object may comprise at least one network, a first computing device coupled to the at least one network, and a second computing device coupled to the at least one network.
- the first computing device may detect the presence of a real world object within a detection area of a detection device.
- the second computing device may generate, in response to detecting the presence of the real world object, a virtual object in a virtual world environment rendered.
- the virtual object may correspond to the real world object and may be rendered in the virtual world environment only while the presence of the real world object continues to be detected within the detection area of the detection device.
- FIG. 1 is an exemplary block diagram of a distributed data processing environment in which aspects of the illustrative embodiments may be implemented;
- FIG. 2 is an exemplary block diagram of a server computing device in which aspects of the illustrative embodiments may be implemented;
- FIG. 3 is an exemplary block diagram of a client computing device in which aspects of the illustrative embodiments may be implemented;
- FIG. 4 is an exemplary block diagram illustrating a mechanism for detecting a persistent presence of real world objects such that the real world objects may be represented as virtual world objects in a virtual world environment in accordance with one illustrative embodiment
- FIG. 5A is pictorial diagram illustrating the rendering of virtual objects in a virtual world environment that correspond to real world objects that are detected as being present in a detection area of a detection pad in accordance with an illustrative embodiment
- FIG. 5B is pictorial diagram illustrating the discontinuing of the rendering of a virtual object in a virtual world environment when its corresponding real world object is no longer detected as being present in a detection area of a detection pad in accordance with an illustrative embodiment
- FIG. 6 is a flowchart outlining an exemplary operation of a real world object persistence to virtual object awareness mapping mechanism in accordance with an illustrative embodiment
- FIG. 7 is a diagram illustrating an exemplary implementation of the illustrative embodiments in which the detection mechanism has a grid array that may be activated in a customized manner depending upon the virtual world environment with which the detection mechanism operates;
- FIG. 8A-8C are diagrams illustrating various implementations of the illustrative embodiments for different types of applications.
- FIGS. 1-3 hereafter are provided as examples of a distributed data processing environment and computing devices in which exemplary aspects of the illustrative embodiments may be implemented.
- FIGS. 1-3 are only exemplary and are not intended to state or imply any limitation with regard to the types and/or configurations of computing devices in which the illustrative embodiments may be implemented. Many modifications to the computing devices and environments depicted in FIGS. 1-3 may be made without departing from the spirit and scope of the present invention.
- FIG. 1 depicts a pictorial representation of a network of data processing systems in which the present invention may be implemented.
- Network data processing system 100 is a network of computers in which the present invention may be implemented.
- Network data processing system 100 contains a network 102 , which is the medium used to provide communications links between various devices and computers connected together within network data processing system 100 .
- Network 102 may include connections, such as wire, wireless communication links, or fiber optic cables.
- server 104 is connected to network 102 along with storage unit 106 .
- clients 108 , 110 , and 112 are connected to network 102 .
- These clients 108 , 110 , and 112 may be, for example, personal computers or network computers.
- server 104 provides data, such as boot files, operating system images, and applications to clients 108 - 112 .
- Clients 108 , 110 , and 112 are clients to server 104 .
- Network data processing system 100 may include additional servers, clients, and other devices not shown.
- network data processing system 100 is the Internet with network 102 representing a worldwide collection of networks and gateways that use the Transmission Control Protocol/Internet Protocol (TCP/IP) suite of protocols to communicate with one another.
- TCP/IP Transmission Control Protocol/Internet Protocol
- At the heart of the Internet is a backbone of high-speed data communication lines between major nodes or host computers, consisting of thousands of commercial, government, educational and other computer systems that route data and messages.
- network data processing system 100 also may be implemented as a number of different types of networks, such as for example, an intranet, a local area network (LAN), or a wide area network (WAN).
- FIG. 1 is intended as an example, and not as an architectural limitation for the present invention.
- Data processing system 200 may be a symmetric multiprocessor (SMP) system including a plurality of processors 202 and 204 connected to system bus 206 . Alternatively, a single processor system may be employed. Also connected to system bus 206 is memory controller/cache 208 , which provides an interface to local memory 209 . I/O Bus Bridge 210 is connected to system bus 206 and provides an interface to I/O bus 212 . Memory controller/cache 208 and I/O Bus Bridge 210 may be integrated as depicted.
- SMP symmetric multiprocessor
- Peripheral component interconnect (PCI) bus bridge 214 connected to I/O bus 212 provides an interface to PCI local bus 216 .
- PCI Peripheral component interconnect
- a number of modems may be connected to PCI local bus 216 .
- Typical PCI bus implementations will support four PCI expansion slots or add-in connectors.
- Communications links to clients 108 - 112 in FIG. 1 may be provided through modem 218 and network adapter 220 connected to PCI local bus 216 through add-in connectors.
- Additional PCI bus bridges 222 and 224 provide interfaces for additional PCI local buses 226 and 228 , from which additional modems or network adapters may be supported. In this manner, data processing system 200 allows connections to multiple network computers.
- a memory-mapped graphics adapter 230 and hard disk 232 may also be connected to I/O bus 212 as depicted, either directly or indirectly.
- FIG. 2 may vary.
- other peripheral devices such as optical disk drives and the like, also may be used in addition to or in place of the hardware depicted.
- the depicted example is not meant to imply architectural limitations with respect to the present invention.
- the data processing system depicted in FIG. 2 may be, for example, an IBM eServer pSeries system, a product of International Business Machines Corporation in Armonk, N.Y., running the Advanced Interactive Executive (AIX) operating system or LINUX operating system.
- AIX Advanced Interactive Executive
- Data processing system 300 is an example of a stand-alone or client computing device.
- Data processing system 300 employs a peripheral component interconnect (PCI) local bus architecture.
- PCI peripheral component interconnect
- AGP Accelerated Graphics Port
- ISA Industry Standard Architecture
- Processor 302 and main memory 304 are connected to PCI local bus 306 through PCI Bridge 308 .
- PCI Bridge 308 also may include an integrated memory controller and cache memory for processor 302 . Additional connections to PCI local bus 306 may be made through direct component interconnection or through add-in boards.
- local area network (LAN) adapter 310 small computer system interface (SCSI) host bus adapter 312 , and expansion bus interface 314 are connected to PCI local bus 306 by direct component connection.
- audio adapter 316 graphics adapter 318 , and audio/video adapter 319 are connected to PCI local bus 306 by add-in boards inserted into expansion slots.
- Expansion bus interface 314 provides a connection for a keyboard and mouse adapter 320 , modem 322 , and additional memory 324 .
- SCSI host bus adapter 312 provides a connection for hard disk drive 326 , tape drive 328 , and CD-ROM drive 330 .
- Typical PCI local bus implementations will support three or four PCI expansion slots or add-in connectors.
- An operating system runs on processor 302 and is used to coordinate and provide control of various components within data processing system 300 in FIG. 3 .
- the operating system may be a commercially available operating system, such as Windows XP, which is available from Microsoft Corporation.
- An object oriented programming system such as Java may run in conjunction with the operating system and provide calls to the operating system from Java programs or applications executing on data processing system 300 . “Java” is a trademark of Sun Microsystems, Inc. Instructions for the operating system, the object-oriented programming system, and applications or programs are located on storage devices, such as hard disk drive 326 , and may be loaded into main memory 304 for execution by processor 302 .
- FIG. 3 may vary depending on the implementation.
- Other internal hardware or peripheral devices such as flash read-only memory (ROM), equivalent nonvolatile memory, or optical disk drives and the like, may be used in addition to or in place of the hardware depicted in FIG. 3 .
- the processes of the present invention may be applied to a multiprocessor data processing system.
- data processing system 300 may be a stand-alone system configured to be bootable without relying on some type of network communication interfaces
- data processing system 300 may be a personal digital assistant (PDA) device, which is configured with ROM and/or flash ROM in order to provide non-volatile memory for storing operating system files and/or user-generated data.
- PDA personal digital assistant
- data processing system 300 also may be a notebook computer or hand held computer in addition to taking the form of a PDA.
- data processing system 300 also may be a kiosk or a Web appliance.
- the client/stand-alone computing device or server computing device executes software applications which provide a virtual world environment which is accessible by a user.
- a user may interact with the virtual world via one or more input devices connected to the computing devices or connected to a client device that communicates with the computing device.
- One of these input devices is a detection mechanism for detecting the presence of real world objects.
- a detection mechanism may use any detection scheme suitable to the particular implementation including bar code scanners, radio frequency identifier (RFID) detection devices, dot code detection devices, and the like.
- the detection mechanism is provided as a detection pad upon which real world objects may be placed and whose presence may be detected using a RFID detection device integrated into the detection pad.
- the detection pad has an associated detection area that comprises the pad itself and a limited range of space around the pad in which real world objects having associated RFID tags may be detected.
- the detection mechanism e.g., the detection pad
- the detection of the presence of the real world object provides an identifier of the object which may be correlated with information regarding how to represent the real world object in the virtual world environment, how the object may be utilized in relation to other objects in the virtual world environment such that the real world object is modeled in the virtual world environment, characteristics of the virtual world object representation of the real world object with regard to the particular applications in which the virtual world object will be used in the virtual world environment, and the like.
- the detection of multiple real world objects may be performed and identification of each of the multiple real world objects may be used to determine how these real world objects may be utilized together in the virtual world environment.
- the illustrative embodiments described herein are especially well suited for use with persistent virtual world type computer games. While the illustrative embodiments will be described in terms of computer games and persistent virtual world computer games, it should be appreciated that the illustrative embodiments are not limited to being applicable to only computer games. To the contrary, the illustrative embodiments may be used with any application in which real world objects may be rendered as virtual objects in a virtual environment.
- a computing device executes operations for monitoring a detection mechanism, which in the illustrative embodiments set forth herein is a detection pad, to detect when one or more real world objects are placed on the detection pad or within the detection area of the detection pad.
- the operations may continue to monitor the detection pad continuously or periodically to determine a current set of one or more real world objects that are within the detection area of the detection pad and compare the current set of real world objects to a previously detected set of real world objects to determine which objects have persistent presence and which objects are no longer within the detection area.
- Only real world objects that are present within the detection area of the detection pad may be modeled and represented in the virtual world environment. Thus, when a real world object is removed from the detection area of the detection pad, the object's representation in the virtual world environment is removed. In this way, the persistent presence of a real world object as detected by the detection pad is mapped into a virtual world object awareness.
- the detection pad is coupled to a stand-alone or client computing device.
- the virtual world environment is provided either directly by the stand-alone or client computing device, by one or more server computing devices with which the client computing device communicates, or by a combination of software applications running on the client computing device and one or more applications running on one or more server computing devices.
- Processing of information from the detection pad is first performed by the stand-alone or client computing device, using driver software and the like, for example, which converts inputs from the detection pad into data that is understandable by one or more software applications running on the client and/or server computing devices.
- the driver software receives inputs from the detection mechanisms of the detection pad, such as RFID detectors of the detection pad, and provides data to one or more software applications corresponding to the inputs received from the detection pad.
- the inputs received from the detection pad are representative of the detection of particular real world objects being within a detection area of the detection pad.
- the RFID detectors of the detection pad may detect RFID tags associated with real world objects that are placed within the detection area of the detection pad and provide as inputs to the stand-alone or client computing device, the corresponding identifiers associated with the RFID tags detected. These identifiers may then be correlated with object information for associated real world objects, such as through a database lookup operation.
- the real world objects preferably are objects that are to be represented by similar virtual objects that are modeled in a virtual world environment within the computing device.
- the real world objects may be toys, e.g., action figures, toy vehicles, and the like, toy accessories, e.g., toy weapons, toy armor, and the like, models, etc. of objects that may be represented as virtual objects in the virtual world environment and used within the virtual world environment.
- the real world object is present within the detection area of the detection pad, the real world object, e.g., the action figure, toy vehicle, toy weapon, etc., is represented in the virtual world in such a manner that is consistent with the usage of the object in the real world.
- the detection of the toy sword in the detection area of the detection pad causes a corresponding virtual sword to be represented in the virtual world environment.
- the virtual sword is represented in the virtual environment in a manner consistent with the usage of a sword in the real world.
- the virtual sword may be represented as being in a free hand of the avatar.
- the virtual sword may be represented as being sheathed and hanging on the side of the user's avatar.
- the representation of the virtual sword in the virtual world environment may be maintained as long as the toy sword is detected as being present in the detection area of the detection pad. If the user were to remove the toy sword from the detection area of the detection pad, when the absence of the toy sword is detected by the detection pad, the virtual world representation of the toy sword is discontinued.
- the detection pad may detect multiple real world objects as being present within the detection area of the detection pad.
- the identification of the various real world objects in the detection area is used to retrieve information regarding the real world objects and how to represent the real world objects as virtual objects in a virtual world environment.
- the interaction between the identified real world objects may be determined using a rules database.
- the real world objects on a detection pad comprise an action figure representing a knight and accessory objects that comprise a toy sword, a toy helmet, and a toy shield.
- Information corresponding to each of these detected items may be retrieved from a database associated with an application running in the computer system. This information may include an identification of how each item is utilized in the real world and how such usage is mapped to the virtual environment.
- a rules database may be provided that provides rules based on the detected real world object identifiers. For example, a rule may state that if a first real world object and a second real world object are detected on the detection pad, then the virtual object corresponding to the second real world object should be rendered as being in a particular association with the first real world object.
- rules which may be simpler or more complicated than the example rule given above, may be provided within a rules database that operate on the detected real world object identifiers and retrieved information corresponding to the real world object identifiers to determine how the virtual objects should be rendered in the virtual world environment.
- the toy sword (e.g., a second real world object) is used in the hand of an action figure as a weapon and has certain attack and defense characteristics.
- a virtual object corresponding to the toy sword may be represented in the hand or sheath of an avatar resembling the action figure that is present on the detection pad.
- the toy shield may be represented as a virtual shield object in a hand of the virtual action figure, i.e. the avatar, or on the back of the virtual action figure in the virtual world environment.
- the toy helmet is used on the head of the action figure and thus, is represented in the virtual world environment as a helmet object placed on the head of the virtual action figure.
- the proper interaction of the detected real world objects in the detection area of the detection pad is automatically determined and used to represent the real world objects as virtual objects in a manner consistent with the way in which these real world objects interact with each other in the real world.
- the operations in the computing device may monitor a condition of the virtual world object representation of the real world object within the virtual world environment and, based on the condition of the virtual world object representation of the real world object, may disable the user's ability to use the real world object to generate a virtual world representation of the real world object.
- the real world object were a toy sword and the user, in a previous session in the virtual world environment, performed actions that resulted in his/her avatar's virtual sword, which corresponds to then real world toy sword, being damaged or broken, then the user's ability to place the toy sword on the detection pad and have a corresponding sword represented in the virtual world environment may be discontinued until the user performs appropriate actions in the virtual world environment, e.g., repairs the virtual representation of the sword in the virtual world environment.
- a persistent state of the virtual representation of the real world object may be maintained in the virtual world environment such that each time a real world object is placed in the detection area of the detection pad, the persistent state of the virtual object corresponding to that real world object may be retrieved and used to render the virtual object in the virtual world environment. For example, using the damaged sword example discussed above, rather than discontinuing the ability of the user to generate a virtual object corresponding to the real world toy sword object on the detection pad, the previous state of the virtual sword may be retrieved and used to render the virtual sword representation of the real world toy sword.
- the virtual sword object rendered in response to detecting the real world toy sword will be rendered as having the same damaged state as when the user last used that real world toy sword object with the virtual world environment.
- FIG. 4 is an exemplary block diagram illustrating a mechanism for detecting a persistent presence of real world objects such that the real world objects may be represented as virtual world objects in a virtual world environment in accordance with one illustrative embodiment.
- the mechanism illustrative in FIG. 4 is an example of a stand-alone computing device. It should be appreciated that the stand-alone computing device in FIG. 4 is only exemplary of one illustrative embodiment and is not intended to state or imply any limitation with regard to the types of computing devices in which the present invention may be implemented.
- the mechanisms of the illustrative embodiments may be distributed amongst a plurality of computing devices in a distributed data processing environment.
- the mechanism includes a detection pad 410 and a computing device 420 .
- Elements 410 and 420 are preferably implemented as hardware units while the remaining elements shown in FIG. 4 may be implemented in hardware, software, or any combination of hardware and software without departing from the spirit and scope of the present invention.
- the detection pad 410 includes one or more detectors, which may be integrated into the detection pad 410 , for detecting the presence of real world objects within a detection area of the detection pad 410 .
- the detection pad 410 transmits information to the computing device 420 either through a wired or wireless communication link 415 .
- the information transmitted to the computing device corresponds to real world object identifiers, corresponding to identification devices that are affixed to or integrated with the real world objects, and which are detected by the detectors in the detection pad 410 . This information may be detected by way of a barcodes provided on the real world objects, RFIDs, dot codes, or the like.
- the real world objects are provided with RFIDs and the detector in the detection pad 410 is configured to scan a detection area and detect the presence of RFIDs within the detection area.
- RFIDs and RFID detectors are generally known in the art and thus, a detailed description is not provided herein.
- the information transmitted to the computing device 420 from the detection pad 410 is input to the input/output interface 430 .
- the input/output interface 430 preferably utilizes detection pad driver(s) 435 to interpret the inputs from the detection pad 410 and provide the corresponding data to the real world object persistence module 440 .
- the real world object persistence module 440 is responsible for determining what real world objects are currently present in the detection area of the detection pad 410 based on inputs from the detection pad 410 , what real world objects have been removed from the detection area of the detection pad 410 , as well as updating information in a detected objects database 450 .
- the real world object persistence module 440 communicates information regarding the real world objects that are currently present within the detection area of the detection pad 410 with the virtual world environment object awareness module 460 .
- the virtual world environment object awareness module 460 is responsible for rendering virtual objects corresponding to the detected real world objects that are currently in the detection area of the detection pad 410 in the virtual world environment.
- the information communicated to the virtual world environment object awareness module 460 from the real world object persistence module 440 may be used to retrieve mapping information from the real world object/virtual object mapping database 470 .
- the virtual world environment object awareness module 460 may utilize this mapping information to render virtual objects corresponding to the detected real world objects in the virtual world environment generated by the virtual world environment application(s) 480 .
- elements 430 - 480 are shown as being part of the computing device 420 .
- the present invention is not limited to such a configuration. Rather, many of the elements 430 - 480 may be provided in one or more other computing devices with which the computing device 420 may communicate via one or more networks.
- elements 460 - 480 may be provided by a server computing device with which the computing device 420 is a client.
- the client computing device 420 is responsible for determining the real world aspects of the real world persistence to virtual world awareness mapping while the server computing device is responsible for the virtual world aspects of this mapping.
- the identifier device is permanently affixed to, or integrated with, the real world object, although this is not required for the functioning of the illustrative embodiments. Such permanent affixing or integration is intended to ensure that the real world object is within the detection area and not just the identifier device.
- the detected identifier associated with the real world object is transmitted from the detection pad 410 to the computing device 420 via the communication link 415 .
- the signals transmitted by the detection pad 410 are processed via the I/O interface 430 using the detection pad driver(s) 435 .
- the resulting identifier information is provided to real world object persistence module 440 .
- the real world object persistence module 440 receives the identifier information and performs a lookup of the identifier information in the detected objects database 450 .
- the detected objects database 450 preferably has two primary portions. In a first portion 452 , the detected objects database 450 maintains a current real world objects data structure that has entries for each real world object that is detected as being currently present within the detection area of the detection pad 410 . In a second portion 454 , the detected objects database 450 maintains a current status of all real world objects that have been detected using the detection pad 410 . The second portion 454 maintains information regarding real world objects that are currently present within the detection area of the detection pad 410 as well as real world objects that are not currently detected as being within the detection area but have been detected by the detection pad 410 previously.
- the entry may be indexed by the identifier associated with the real world object and may contain, for example, information identifying a current state of the virtual world object that corresponds to the real world object, which may be determined based upon previous state information retrieved from the second portion 454 . This state information may be updated periodically, or when a change in state of the virtual world object that corresponds to the real world object occurs in the virtual world environment provided by application(s) 480 .
- the real world object persistence module 440 sends a message to the virtual world environment object awareness module 460 indicating that a new real world object has been detected and provides the identification information for the new real world object.
- the virtual world environment object awareness module 460 maps the real world object identifier to a virtual object that is to be represented in the virtual world environment. This mapping is performed using mapping information retrieved from the real world object/virtual object mapping database 470 .
- the information maintained in the real world object/virtual object mapping database 470 includes pointer information identifying the data, files, etc. corresponding to a virtual world representation of the real world object, i.e.
- the virtual object and pointer information for data, files, etc. that identify the manner by which the virtual object is utilized within the virtual world environment.
- These data, files, etc. preferably model the real world object as a virtual object in such a manner that the virtual world object emulates the real world object not only in appearance but in the manner by which the real world object is utilized with other objects in the real world.
- the virtual world environment object awareness module 460 interfaces with the virtual world environment provided by the virtual world environment application(s) 480 running on the computing device 420 .
- the virtual world environment object awareness module 460 informs the virtual world environment application(s) of the virtual objects that correspond to real world objects whose presence is currently detected by the detection pad 410 as well as the necessary information for rendering the virtual objects and modeling the virtual objects with regard to the virtual objects' interaction with the virtual world. In this way, a virtual object corresponding to a detected real world object may be generated and utilized in a virtual world environment.
- the real world object persistence module 440 In addition to registering real world objects that are detected using the detection pad 410 , the real world object persistence module 440 also checks the first portion 452 of the detected objects database 450 to determine if there are any entries for real world objects that are no longer detected as being within the detection area of the detection pad 410 . That is, the identifier information obtained based on the inputs from the detection pad 410 are compared to the entries in the first portion 452 of the detected objects database 450 as previously described above.
- Entries that are created or already existing in the first portion 452 that match the identifiers of the real world objects detected using the detection pad 410 are temporarily marked, e.g., a bit flag is set, to thereby identify the entry as being associated with a real world object that is currently present in the detection area of the detection pad 410 .
- the real world object persistence module 440 may then check the entries in the first portion 452 of the detected objects database 450 and thereby identify which entries are not marked as being associated with real world objects that are currently detected in the detection area of the detection pad 410 .
- the real world objects associated with these entries are real world objects that have been removed from the detection area of the detection pad 410 .
- the real world object persistence module 440 may communicate with the virtual world environment object awareness module 460 to inform the virtual world environment object awareness module 460 that the virtual objects corresponding to the real world objects that were removed from the detection area should be removed from the virtual world environment.
- the virtual world environment object awareness module 460 sends a message to the virtual world environment application(s) 480 to discontinue the rendering of the virtual objects corresponding to the removed read world objects.
- the above described operation for detecting real world objects and identifying which real world objects have been removed from a detection area may be performed continuously or periodically. In this way, only those real world objects whose presence is persistent as detected by the detection pad may have their corresponding virtual objects continue to be rendered and utilized in the virtual world environment. Thus, a real world object persistence is mapped to a virtual world object awareness.
- the virtual world environment object awareness module 460 determines the manner by which a plurality of real world objects detected within the detection area of the detection pad 410 interact with one another or are utilized with each other, if at all, so that they may be accurately represented as a combination of virtual objects in the virtual world environment. For example, if the real world object persistence module 440 receives identifier information for an action figure, a helmet for the action figure, a sword for the action figure, and a shield for the action figure, the virtual world environment object awareness module 460 determines how these objects may be utilized together based on information maintained within the real world object/virtual object mapping database 470 .
- the virtual world environment object awareness module 460 may determine through a lookup of the real world object identifiers in the real world object/virtual object mapping database 470 that the sword should be placed in a right hand of the action figure, the helmet should be placed on a head of the action figure, and the shield should be placed in the left hand of the action figure.
- the virtual world environment object awareness module 460 may generate a virtual avatar in the virtual world environment, based on the information retrieved from the real world object/virtual object mapping database 470 , and may place a sword virtual object corresponding to the real world sword object in the right hand of the virtual avatar, a shield virtual object corresponding to the real world shield object in the left hand of the virtual avatar, and a helmet virtual object corresponding to the helmet real world object on the head of the virtual avatar. The user may then utilize these objects within the virtual world environment in a manner corresponding to the models of these objects within the virtual world.
- the modeling of virtual objects in a virtual world environment is generally known in the art and thus, a detailed description is not provided herein.
- the virtual world environment object awareness module 460 is able to determine the manner by which each of the accessory real world objects, i.e. the sword, shield and helmet, are utilized with or interact with the action figure real world object. This interaction is then modeled in the virtual world environment by providing a combination of virtual objects that are modeled after the manner by which the detected real world objects are utilized in the real world.
- the utilization of the objects in the virtual world may be completely based upon the modeling specified by the programmer which may or may not have any relation to the real world.
- the current status of a virtual object in the virtual world environment that corresponds to a real world object detected using the detection pad 410 may be persisted and associated with the real world object such that when the real world object is detected again by the detection pad 410 , the last state of the virtual object corresponding to the real world object may be retrieved.
- a real world object that is detected by the detection pad 410 corresponds to a toy sword for an action figure.
- the toy sword is rendered in the virtual world environment as a virtual sword object that may be utilized by a virtual avatar to battle enemies in the virtual world environment.
- the virtual sword object may sustain damage that causes the virtual sword object to be less effective or need repair. Such a status may be determined and maintained in the virtual world environment.
- the current status of the virtual sword object may be reported back to the real world object persistence module 440 .
- the real world object persistence module 440 may then store this current status in the second portion 454 of the detected objects database 450 in associated with the identifier of the real world toy sword object.
- the real world object persistence module 440 may retrieve the current status information from the second portion 454 of the detected objects database 450 and provide the current status information to the virtual world environment object awareness module 460 .
- the virtual world environment object awareness module 460 may then use this current status information when rendering and modeling the virtual sword object in the virtual world environment.
- the current status information may identify any status that may be modeled within the virtual world environment, e.g., damaged, needing repair, ineffective, destroyed, etc.
- the current status information may simply identify real world objects that may be rendered as virtual objects within the virtual world environment and real world objects that cannot until some action in the virtual world environment is taken. For example, if virtual sword object in the virtual world environment is broken within the virtual world environment, this status may be stored in association with the real world toy sword object and thus, when the real world toy sword object is placed within the detection area of the detection pad, the virtual sword object is not rendered until the user visits a blacksmith in the virtual world environment and has the virtual sword object repaired. Thus, not only is the presence of the real world objects within the detection area of the detection pad made persistent, but the virtual status of the virtual objects corresponding to the real world objects is persisted in association with the real world objects.
- FIG. 5A is pictorial diagram illustrating the rendering of virtual objects in a virtual world environment that correspond to real world objects that are detected as being present in a detection area of a detection pad in accordance with an illustrative embodiment.
- an action figure 510 representing a knight is placed on a detection pad 550 along with accessories for the action figure that include a toy shield 520 , a toy sword 530 , and a toy helmet 540 .
- Each of these real world objects has an associated RFID tag attached or integrated into them that identifies the real world objects.
- the detection pad 550 detects the presence of these real world objects and transmits the identification information to the computing device 560 .
- the computing device 560 runs software modules corresponding to elements 430 - 480 of FIG. 4 .
- the computing device 560 uses the various operations described above with regard to FIG. 4 , determines how to render the virtual avatar 570 in the virtual world environment 580 output using the display of the computing device 560 .
- the virtual avatar 570 includes a helmet virtual object 572 placed on a head location of the virtual avatar 570 , a sword virtual object 574 in a right hand of the virtual avatar 570 , and a shield virtual object 576 in a left hand of the virtual avatar 570 . Because the shield virtual object 576 had been used in a previous session and was damaged, the rendering of the shield object 576 shows the shield virtual object 576 as being damaged.
- FIG. 5B is pictorial diagram illustrating the discontinuing of the rendering of a virtual object in a virtual world environment when its corresponding real world object is no longer detected as being present in a detection area of a detection pad in accordance with an illustrative embodiment.
- the toy shield real world object 520 has been removed from the detection area of the detection pad 550 .
- the computing device 560 detects the removal of the toy shield real world object 520 and discontinues rendering of the shield virtual object 576 in the virtual world environment 580 .
- the virtual avatar 570 is shown in the virtual world environment 580 as having only the helmet virtual object 572 and the sword virtual object 574 .
- FIG. 6 is a flowchart outlining an exemplary operation of a real world object persistence to virtual object awareness mapping mechanism in accordance with an illustrative embodiment. It will be understood that each block of the flowchart illustration, and combinations of blocks in the flowchart illustration, can be implemented by computer program instructions. These computer program instructions may be provided to a processor or other programmable data processing apparatus to produce a machine, such that the instructions which execute on the processor or other programmable data processing apparatus create means for implementing the functions specified in the flowchart block or blocks.
- These computer program instructions may also be stored in a computer-readable memory or storage medium that can direct a processor or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory or storage medium produce an article of manufacture including instruction means which implement the functions specified in the flowchart block or blocks.
- blocks of the flowchart illustration support combinations of means for performing the specified functions, combinations of steps for performing the specified functions and program instruction means for performing the specified functions. It will also be understood that each block of the flowchart illustration, and combinations of blocks in the flowchart illustration, can be implemented by special purpose hardware-based computer systems which perform the specified functions or steps, or by combinations of special purpose hardware and computer instructions.
- the operation starts with the detection of real world objects using the detection mechanism, e.g., the detection pad (step 610 ).
- the identifiers of the real world objects that are detected are provided to the real world object persistence module (step 620 ).
- a determination is made as to whether an entries are already existing in the current detected objects data structure for the detected real world objects (step 630 ). If not, an entry for each detected real world object that does not already have an entry in the current detected objects data structure is generated (step 640 ).
- a previous state associated with the detected real world objects that did not have an entry is retrieved from the detected objects data structure if such a previous state exists (step 650 ).
- the real world objects that have been removed from the detection area of the detection mechanism are identified (step 660 ).
- One or more messages are sent to the virtual world environment object awareness module identifying detected real world objects and their associated states, if any (step 670 ).
- a determination is made as to whether any real world objects were removed from the detection area of the detection mechanism (step 680 ). If so, then the current state of the virtual objects corresponding to the removed real world objects is stored in association with the identifiers of the real world objects that were removed (step 690 ).
- the representation of the virtual objects corresponding to the removed real world objects is then discontinued in the virtual world environment (step 700 ).
- step 710 information is retrieved from the real world object/virtual object mapping database for the currently detected real world objects.
- Virtual objects corresponding to the currently detected real world objects are generated in the virtual world environment based on information from the mapping database, state information associated with the real world objects, and the like (step 720 ).
- the generation of such virtual objects may further involve determining the manner by which the real world objects interact with each other or otherwise are related to each other and then depicting the virtual objects in a manner consistent with such interaction or relation.
- the detection mechanism may further be used as an output device as well and may be configured to include output mechanisms for outputting information to a user in a visual, auditory, and/or tactile manner.
- the detection mechanism may be configured to provide output to the user so as to guide the user in the use of the detection mechanism with the real world objects that are placed within the detection area of the detection mechanism.
- the detection mechanism of the illustrative embodiments may have a plurality of detectors or sensors that are used to detect or sense the presence of a real world object within a detection area of the detection mechanism.
- These detectors or sensors may be configured in any manner suitable to the particular implementation of the illustrative embodiments selected.
- these detectors/sensors may be configured in a grid-like manner within the detection mechanism.
- the detection mechanism may have visual output elements that are arranged in a similar grid-like manner that may be selectively activated so as to inform a user of areas of the detection mechanism where certain real world objects may be placed.
- various ones of the visual output elements may be energized or otherwise made more noticeable (hereafter referred to as being “highlighted”) in accordance with a desired output pattern 750 on the detection pad.
- This output pattern may be representative, for example, of the types of real world objects that may be placed on the detection pad and used with the virtual world environment.
- the visual output elements may be light emitting diodes (LEDs) or other types of light generating devices that may be selectively activated so as to display a pattern on the detection pad.
- LEDs light emitting diodes
- the particular shapes that may be generated using such visual output elements are not limited by the illustrative embodiments herein and may encompass any shape that may be represented using a grid of visual output elements.
- a human pattern may be highlighted or outlined for an adventure or role playing game virtual world environment.
- card placement blocks may be highlighted signifying different play areas in which real world objects may be placed.
- the particular areas of the detection pad that are highlighted may be associated with activated detectors or sensors such that portions of the detection pad that are not highlighted are not active. In this way, selected portions of the detection pad may be activated, depending upon the particular virtual world environment with which the detection pad is being utilized, and may be used to detect the presence of real world objects while other areas are de-activated and will not detect the presence of real world objects even if the real world objects are present in these areas.
- a coordinate system may be established with the detectors/sensors of the detection pad, such as a lower left corner of the detection pad being at coordinates (0,0), and this coordinate system may be used to determine the relative placement of real world objects on the detection pad.
- the ability to associated coordinates with real world objects aids in the relative rendering of virtual objects in the virtual world environment that correspond to real world objects placed on the detection pad. For example, if the virtual world environment represents a strategy board game, the relative placement of real world pieces may be determined using the coordinate system of the detection pad and used to render virtual objects corresponding to those pieces with the same relative placement in the virtual world environment.
- FIGS. 8A-8C illustrate some exemplary applications of the detection mechanism of the illustrative embodiments.
- the detection mechanism 800 of the illustrative embodiments may be used to highlight active areas 810 - 850 of the detection mechanism 800 in which playing cards for a collectable card game may be placed.
- active areas 810 - 840 represent cards that are in play while active area 850 represent the active deck for the card game. Cards that have been played and that are no longer in play may be removed from the detection pad 800 , as shown by played deck 860 .
- the cards that are in play in active areas 810 - 850 are detected by the detectors associated with these active areas and corresponding virtual objects may be generated in a virtual world environment represented in a computing device. These virtual objects may be virtual cards or may be virtual objects corresponding to items, people, creatures, or the like, that are represented by the real world cards.
- FIG. 8B illustrates another detection mechanism 860 that is configured for use with large board games such as strategy games or the like.
- the detection mechanism 860 has a first portion 862 that comprises the detection array, i.e. the array or detectors/sensors for detecting the presence of real world objects.
- the first portion 862 may further comprise the visual output elements described previously above.
- a second portion 864 may comprise a display device that is capable of generating graphical displays, e.g., a liquid crystal display, plasma display, or the like.
- the second portion 864 may be used to output a representation of a playing field upon which real world objects may be placed while the first portion 862 detects the presence of these real world objects on the detection mechanism.
- the playing field may be represented as graphical images, such as of rivers, mountains, and other terrain, and may be generated based on outputs from the virtual world environment software running on the computing device to which the detection mechanism 860 is connected, or the like.
- the presence of the real world objects on the detection mechanism 860 may be translated into a virtual object awareness in the virtual world environment generated by such software.
- the detection mechanism 860 may represent a two-dimensional playing field
- the computing device in which the virtual world environment is rendered may represent the two-dimensional playing field in three dimensions.
- the second portion 864 provides a mechanism by which images and/or topological information may be overlaid on the detection mechanism's surface to convey detail that enhances the software program with which the detection mechanism is being used.
- images may be displayed within these highlighted/activated areas that depict the types of real world objects that should be placed in these highlighted/activated areas. For example, an image of a playing card may be displayed within the highlighted areas 810 - 840 of FIG. 8A .
- a parts list for products requiring assembly may be displayed and the presence of these parts may be detected.
- a product may be purchased that requires assembly.
- the pieces of the product may have identifier tags embedded or otherwise attached to them and may be deposited on the presence detection mechanism.
- An assembly program with which the detection mechanism interfaces may be run on an associated computing device and, for each step of the assembly instructions, the grid location corresponding to the component to be assembled/used next in the series of assembly steps may be highlighted.
- the assembly program may detect the presence of the pieces via the detection mechanism and represent the pieces in a virtual world environment in which the various steps for assembly of the product are virtually represented.
- the assembly program may perform a check of the complete inventory of items against the assembly program before assembly begins to ensure that all the necessary pieces are present.
- the detection mechanism may be used to represent clothing in a virtual world environment, such as on a virtual mannequin, as illustrated in FIG. 8C .
- a user may register with a computing system and provide general dimensions for the user.
- the user may register with the computing system either directly in the clothing store where the real world clothing objects are present or may register on-line from another location, for example.
- the user may select various pieces of clothing 870 - 880 and place them on the detection mechanism 890 , which may be oriented horizontally, vertically, or the like.
- the detection mechanism 890 may detect the presence of these clothing items 870 - 880 and, through the mechanisms previously described above, determine how to represent these real world objects as virtual objects in the virtual world environment in which the virtual mannequin is present.
- the detected pieces of clothing may be mapped as virtual clothing onto the virtual mannequin that has the dimensions of the user.
- the virtual mannequin with the virtual clothing mapped thereon may be displayed to the user via a display device 895 so that the user may view the results of the mapping.
- the illustrative embodiments help to alleviate the cumbersome nature of tying on clothes and reduce the work to merely matching clothing items and selecting appropriate sizes.
- the size of the detection mechanism e.g., the detection pad
- the size of the detection pad may be determined based on the particular implementation and intended use of the detection pad. For example, if the detection pad will be used primarily with a strategy game virtual world environment, in which a large number of pieces may be placed on a large game board, the game pad may be large enough to cover a typical dining room table or coffee table. On the other hand, if the detection pad is to be used with a card game virtual environment, then the detection pad may be quite small in size relative to a dining room table or coffee table.
- the illustrative embodiments provide a mechanism for detecting the persistent presence of real world objects and translating this persistent presence into a virtual object awareness in a virtual world environment.
- virtual objects corresponding to currently detected real world objects may be generated based on detected identifiers of the real world objects. These virtual objects may be generated in such a manner as to be consistent with the intended interaction or relationships of the real world objects.
- the virtual object corresponding to that real world object is removed from the virtual world environment.
- virtual objects corresponding to real world objects are only able to be utilized in the virtual world environment while their corresponding real world objects are detected within a detection area of a detection mechanism.
- the illustrative embodiments described above may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment containing both hardware and software elements.
- the invention is implemented in software, which includes but is not limited to firmware, resident software, microcode, etc.
- the illustrative embodiments may take the form of a computer program product accessible from a computer-usable or computer-readable medium providing program code for use by or in connection with a computer or any instruction execution system.
- a computer-usable or computer readable medium can be any apparatus that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
- the medium may be an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system (or apparatus or device) or a propagation medium.
- Examples of a computer-readable medium include a semiconductor or solid state memory, magnetic tape, a removable computer diskette, a random access memory (RAM), a read-only memory (ROM), a rigid magnetic disk and an optical disk.
- Current examples of optical disks include compact disk—read only memory (CD-ROM), compact disk—read/write (CD-R/W) and DVD.
- a data processing system suitable for storing and/or executing program code will include at least one processor coupled directly or indirectly to memory elements through a system bus.
- the memory elements can include local memory employed during actual execution of the program code, bulk storage, and cache memories which provide temporary storage of at least some program code in order to reduce the number of times code must be retrieved from bulk storage during execution.
- I/O devices may be coupled to the system either directly or through intervening I/O controllers.
- Network adapters may also be coupled to the system to enable the data processing system to become coupled to other data processing systems or remote printers or storage devices through intervening private or public networks. Modems, cable modem and Ethernet cards are just a few of the currently available types of network adapters.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Computer Security & Cryptography (AREA)
- General Business, Economics & Management (AREA)
- Processing Or Creating Images (AREA)
Abstract
Description
- 1. Technical Field
- The present application relates generally to an improved data processing system and method. More specifically, the present application is directed to a persistent authenticating system and method to map real world object presence into virtual world object awareness.
- 2. Description of Related Art
- The ability to scan objects for a particular identification signature of the object is generally known in the art. For example, the use of barcodes, dot codes, radio frequency identifier (RFID) tags, and the like, to provide an identification of an object associated with these scannable identifier devices is generally known in the art. The use of such devices, such as the RFID tags, is only expected to increase in the near future.
- One application of scannable identification devices is the use of dot codes on scannable trading cards. For example, the use of dot codes is represented by the E-READER™ plug-in device for the Nintendo GameBoy™ portable gaming system. With the E-READER™, dot codes may be provided on trading cards that, when scanned by the E-READER™ plug-in device unlock special features in software including mini-games, card statistics and data, special items, and the like. Once the trading card is scanned, the special features are unlocked and the trading card need not be present for the unlocked features to be utilized. In fact, the E-READER™ and GameBoy™ system may be turned off and the unlocked features may again be accessed when the system is turned back on due to a storage in the E-READER™ device.
- While the trading cards and the E-READER™ device provide a mechanism for unlocking features of computer software, the unlocked features do not represent real-world items that are present at the location of the E-READER™ device. Moreover, the unlocked features do not represent how to use such a real-world item, how the real-world item's usage is persisted, or the like, so that the real-world item may be accurately modeled and integrated into the computer software. To the contrary, the trading cards used with the E-READER™ device merely serve as a recordable medium upon which the dot-codes are recorded so that they may be read by the E-READER™ device.
- The illustrative embodiments provide a persistent authenticating system and method to map real world object presence into virtual world object awareness. More specifically, the illustrative embodiments provide a mechanism by which the presence of a real world object is detected and, while the real world object's presence is continued to be detected, it is made available for use in a virtual world environment. Moreover, the detection of the presence of the real world object provides an identifier of the object which may be correlated with information regarding how to represent the real world object in the virtual world environment, how the object may be utilized in relation to other objects in the virtual world environment such that the real world object is modeled in the virtual world environment, and the like. Furthermore, the detection of multiple real world objects may be performed and identification of each of the multiple real world objects may be used to determine how these real world objects may be utilized together in the virtual world environment.
- One significant application of the illustrative embodiments described herein is to the field of computer gaming. Moreover, within the field of computer gaming, the illustrative embodiments are especially well suited for use with persistent virtual world type computer games. While the illustrative embodiments will be described in terms of computer games and persistent virtual world computer games, it should be appreciated that the illustrative embodiments are not limited to being applicable to only computer games. To the contrary, the illustrative embodiments may be used with any application in which real world objects may be rendered as virtual objects in a virtual environment.
- In one illustrative embodiment, a computing device having an associated detection mechanism, such as a detection pad upon which real world objects may be placed, is provided. The computing device executes operations for monitoring the detection pad to detect when one or more real world objects are placed on the detection pad or within a detection area of the detection pad. The operations may continue to monitor the detection pad continuously or periodically to determine a current set of one or more real world objects that are within the detection area of the detection pad and compare the current set of real world objects to a previously detected set of real world objects to determine which objects have persistent presence and which objects are no longer within the detection area. Only real world objects that are present within the detection area of the detection pad may be modeled and represented in the virtual world environment. Thus, when a real world object is removed from the detection area of the detection pad, the object's representation in the virtual world environment is removed. In this way, the persistent presence of a real world object as detected by the detection pad is mapped into a virtual world object awareness.
- The real world objects preferably are objects that are to be represented by similar virtual objects that are modeled in a virtual world environment within the computing device. For example, the real world objects may be toys, e.g., action figures, toy vehicles, and the like, toy accessories, e.g., toy weapons, toy armor, and the like, clothing items, household items, pieces of a product for assembly, or any other type of objects that may be represented as virtual objects in the virtual world environment and used within the virtual world environment. When the real world object is present within the detection area of the detection pad, the real world object, e.g., the action figure, toy vehicle, toy weapon, etc., is represented in the virtual world in such a manner that is consistent with the usage of the object in the real world.
- For example, if the real world object is a toy representation of a sword, the detection of the toy sword in the detection area of the detection pad causes a corresponding virtual sword to be represented in the virtual world environment. Moreover, the virtual sword is represented in the virtual environment in a manner consistent with the usage of a sword in the real world. Thus, if a user's virtual world avatar currently is involved in a combat situation, the virtual sword may be represented as being in a free hand of the avatar. If the user's virtual world avatar is currently not involved in combat, the virtual sword may be represented as being sheathed and hanging on the side of the user's avatar.
- The representation of the virtual sword in the virtual world environment may be maintained as long as the toy sword is detected as being present in the detection area of the detection pad. If the user were to remove the toy sword from the detection area of the detection pad, when the absence of the toy sword is detected by the detection pad, the virtual world representation of the toy sword is discontinued.
- As mentioned above, the detection pad may detect multiple real world objects as being present within the detection area of the detection pad. The identification of the various real world objects in the detection area is used to retrieve information regarding the real world objects and how to represent the real world objects as virtual objects in a virtual world environment. As part of the determination as to how to represent the real world objects as virtual objects in the virtual world environment, the interaction between the identified real world objects may be determined using a rules database.
- For example, assume that the real world objects on a detection pad comprise an action figure representing a knight and accessory objects that comprise a toy sword, a toy helmet, and a toy shield. Information corresponding to each of these detected items may be retrieved from a database associated with an application running in the computer system. This information includes an identification of how each item is utilized in the real world and how such usage is mapped to the virtual world environment. For example, the toy sword is used in the hand of an action figure as a weapon and has certain attack and defense characteristics. As a result, when representing the toy sword in conjunction with the action figure as virtual objects in the virtual world environment, a virtual object corresponding to the toy sword may be represented in the hand or sheath of an avatar resembling the action figure that is present on the detection pad. Similarly, the toy shield, may be represented as a virtual shield object in a hand of the virtual action figure, i.e. the avatar, or on the back of the virtual action figure in the virtual world environment. The toy helmet is used on the head of the action figure and thus, is represented in the virtual world environment as a helmet object placed on the head of the virtual action figure. Thus, the proper interaction of the detected real world objects in the detection area of the detection pad is automatically determined and used to represent the real world objects as virtual objects in a manner consistent with the way in which these real world objects interact with each other in the real world.
- Furthermore, the operations in the computing device may monitor a condition of the virtual world object representation of the real world object within the virtual world environment and, based on the condition of the virtual world object representation of the real world object, may disable the user's ability to use the real world object to generate a virtual world representation of the real world object. For example, if the real world object were a toy sword and the user, in a previous session in the virtual world environment, performed actions that resulted in his/her avatar's virtual sword, which corresponds to then real world toy sword, being damaged or broken, then the user's ability to place the toy sword on the detection pad and have a corresponding sword represented in the virtual world environment may be discontinued until the user performs appropriate actions in the virtual world environment, e.g., repairs the virtual representation of the sword in the virtual world environment.
- Alternatively, a persistent state of the virtual representation of the real world object may be maintained in the virtual world environment such that each time a real world object is placed in the detection area of the detection pad, the persistent state of the virtual object corresponding to that real world object may be retrieved and used to render the virtual object in the virtual world environment. For example, using the damaged sword example discussed above, rather than discontinuing the ability of the user to generate a virtual object corresponding to the real world toy sword object on the detection pad, the previous state of the virtual sword may be retrieved and used to render the virtual sword representation of the real world toy sword. Thus, if a user, during a previous session in the virtual world environment, damages his virtual sword but fails to repair it, when the user next places the real world toy sword on the detection pad, the virtual sword object rendered in response to detecting the real world toy sword will be rendered as having the same damaged state as when the user last used that real world toy sword object with the virtual world environment.
- Additionally, the detection pad may define a grid array that allows for the ability to associate coordinates with items placed within the detection field. Furthermore, it is possible to activate, highlight, etc. specific patterns of the grid to indicate active sensor regions, relevant to the specific virtual world environment, in which real world objects may be placed and their presence detected using the mechanisms of the illustrative embodiments. For example, a human pattern may be highlighted or outlined for an adventure or role playing game virtual world environment. For a trading card based game, card placement blocks may be highlighted/activated, signifying different play areas in which real world objects may be placed.
- The size of the detection pad is not limited by the illustrative embodiments described herein. To the contrary, the size of the detection pad may be determined based on the particular implementation and intended use of the detection pad. For example, if the detection pad will be used primarily with a strategy game virtual world environment, in which a large number of pieces may be placed on a large game board, the game pad may be large enough to cover a typical dining room table or coffee table. On the other hand, if the detection pad is to be used with a card game virtual environment, then the detection pad may be quite small in size relative to a dining room table or coffee table.
- As a non-gaming related example, using the highlighting/activation of grid patterns or locations on the detection pad described above may be exemplified in a parts list for products requiring assembly. For example a product may be purchased that requires assembly. The pieces of the product may have identifier tags embedded or otherwise attached to them, and may be deposited on the presence detection pad. An assembly program with which the detection pad interfaces may be run on an associated computing device and, for each step of the assembly instructions, the grid location corresponding to the component to be assembled/used next in the series of assembly steps may be highlighted. The assembly program may detect the presence of the pieces via the detection pad and represent the pieces in a virtual world environment in which the various steps for assembly of the product are virtually represented. Moreover, the assembly program may perform a check of the complete inventory of items against the assembly program before assembly begins to ensure that all the necessary pieces are present.
- In a further illustrative embodiment, images and/or topological information may be overlaid on the detection pad surface to convey detail that enhances the application program. For example, in addition to highlighting/activating portions of the detection pad based on the particular programs with which the detection pad interfaces, images may be displayed within these highlighted/activated areas that depict the types of real world objects that should be placed in these highlighted/activated areas.
- In one illustrative embodiment, a method, in a data processing system, for rendering virtual world objects based on a detected presence of a real world object is provided. The method may comprise detecting a presence of a real world object within a detection area of a detection device and generating, in response to detecting the presence of the real world object, a virtual object in a virtual world environment rendered in the data processing system. The virtual object may correspond to the real world object and may be rendered in the virtual world environment only while the presence of the real world object continues to be detected within the detection area of the detection device. The detection device may be a detection pad having one or more detectors integrated therein and upon which real world objects may be placed. The real world object may be a toy and the virtual world environment may be a virtual computer gaming environment, for example.
- The method may detect the presence of the real world object by detecting an identifier mechanism associated with the real world object. The he identifier mechanism may have associated identifier information. The identifier information may be correlated with virtual object characteristic information identifying a manner by which to represent the real world object in the virtual world environment. The virtual object characteristic information may comprise at least one of information identifying how the virtual object may be utilized in relation to other virtual objects in the virtual world environment such that the real world object is modeled in the virtual world environment, or characteristics of the virtual object with regard to particular applications in which the virtual object will be used in the virtual world environment. The identifier mechanism may be either affixed to or integrated with the real world object.
- In one illustrative embodiment, the identifier mechanism may be a radio frequency identification (RFID) mechanism. The detection mechanism thus, may comprise one or more RFID detectors.
- The method may further comprise monitoring a condition of the virtual object corresponding to the detected real world object within the virtual world environment. The virtual object condition information identifying the condition of the virtual object may be stored in association with the identifier of the real world object. Correlating the identifier information associated with the detected identifier mechanism with virtual object characteristic information may further comprise retrieving the virtual object condition information corresponding to the identifier information and rendering the virtual object corresponding to the identifier information such that the virtual object has a condition corresponding to the virtual object condition information. Rendering the virtual object corresponding to the identifier information such that the virtual object has a condition corresponding to the virtual object condition information may comprise disabling rendering of the virtual object corresponding to the identifier information if the condition meets a predetermined criteria.
- The method may further comprise detecting removal of the real world object from the detection area of the detection mechanism. Rendering of the virtual object corresponding to the real world object in the virtual world environment may be automatically discontinued in response to detection of the removal of the real world object from the detection area.
- The method may detect a presence of a real world object within a detection area of a detection device by detecting a presence of a plurality of real world objects within the detection area and correlating identifiers associated with each of the real world objects with virtual object characteristic information for rendering the real world objects as virtual objects in the virtual world environment. At least one way in which at least two of the plurality of real world objects are utilized together may be determined based on the virtual object characteristic information. The virtual objects corresponding to the detected plurality of real world objects may be rendered within the virtual world environment based on the virtual object characteristic information and the determined at least one way in which at least two of the plurality of real world objects are utilized together such that at least two of the virtual objects are rendered as being utilized together.
- The method may be implemented in a distributed data processing system. A first computing device may detect the presence of the real world object and may transmit information corresponding to the detection of the real world object to a second computing device in the data processing system. The second computing device may generate, in response to the detection of the presence of the real world object, the virtual object in the virtual world environment.
- In a further illustrative embodiment, the method may further comprise activating one or more portions of the detection mechanism. Real world objects may only be detectable by the detection mechanism when the real world objects are placed within a detection area corresponding to the activated one or more portions of the detection mechanism. The one or more portions of the detection mechanism that are activated may be configured to represent a type of real world object that is to be placed within the detection area of the one or more portions of the detection mechanism.
- The method may further comprise outputting, via the detection mechanism, an output to aid a user in placing real world objects within the detection areas of the activated one or more portions of the detection mechanism. The detection mechanism may be comprised of a grid of detectors and activating one or more portions of the detection mechanism may comprise activating one or more grid portions. Outputting an output to aid the user may comprise outputting a visual output that highlights the activated one or more grid portions.
- In a further illustrative embodiment, a computer program product comprising a computer useable medium having a computer readable program is provided. The computer readable program, when executed on a computing device, causes the computing device to perform various ones, and combinations of, the operations previously described above with regard to the method illustrative embodiments.
- In yet another illustrative embodiment, an apparatus for rendering virtual world objects based on a detected presence of a real world object is provided. The apparatus may comprise a computing device and a detection device coupled to the computing device. The detection device may detect a presence of a real world object within a detection area of the detection device. The computing device may comprise a processor and a memory, and the memory may comprise instructions. When the instructions are executed by the processor, the instructions may cause the processor to perform various ones, and combinations of, the operations previously described above with regard to the method illustrative embodiments.
- In another illustrative embodiment, a system for rendering virtual world objects based on a detected presence of a real world object is provided. The system may comprise at least one network, a first computing device coupled to the at least one network, and a second computing device coupled to the at least one network. The first computing device may detect the presence of a real world object within a detection area of a detection device. The second computing device may generate, in response to detecting the presence of the real world object, a virtual object in a virtual world environment rendered. The virtual object may correspond to the real world object and may be rendered in the virtual world environment only while the presence of the real world object continues to be detected within the detection area of the detection device.
- These and other features and advantages of the present invention will be described in, or will become apparent to those of ordinary skill in the art in view of, the following detailed description of the exemplary embodiments of the present invention.
- The novel features believed characteristic of the invention are set forth in the appended claims. The invention itself, however, as well as a preferred mode of use, further objectives and advantages thereof, will best be understood by reference to the following detailed description of an illustrative embodiment when read in conjunction with the accompanying drawings, wherein:
-
FIG. 1 is an exemplary block diagram of a distributed data processing environment in which aspects of the illustrative embodiments may be implemented; -
FIG. 2 is an exemplary block diagram of a server computing device in which aspects of the illustrative embodiments may be implemented; -
FIG. 3 is an exemplary block diagram of a client computing device in which aspects of the illustrative embodiments may be implemented; -
FIG. 4 is an exemplary block diagram illustrating a mechanism for detecting a persistent presence of real world objects such that the real world objects may be represented as virtual world objects in a virtual world environment in accordance with one illustrative embodiment; -
FIG. 5A is pictorial diagram illustrating the rendering of virtual objects in a virtual world environment that correspond to real world objects that are detected as being present in a detection area of a detection pad in accordance with an illustrative embodiment; -
FIG. 5B is pictorial diagram illustrating the discontinuing of the rendering of a virtual object in a virtual world environment when its corresponding real world object is no longer detected as being present in a detection area of a detection pad in accordance with an illustrative embodiment; -
FIG. 6 is a flowchart outlining an exemplary operation of a real world object persistence to virtual object awareness mapping mechanism in accordance with an illustrative embodiment; -
FIG. 7 is a diagram illustrating an exemplary implementation of the illustrative embodiments in which the detection mechanism has a grid array that may be activated in a customized manner depending upon the virtual world environment with which the detection mechanism operates; and -
FIG. 8A-8C are diagrams illustrating various implementations of the illustrative embodiments for different types of applications. - The illustrative embodiments described hereafter provide a persistent authenticating system and method to map real world object presence into virtual world object awareness. As such, the illustrative embodiments may be implemented in a stand-alone computing device or in a distributed data processing environment in which multiple computing devices are utilized along with one or more data networks. Accordingly,
FIGS. 1-3 hereafter are provided as examples of a distributed data processing environment and computing devices in which exemplary aspects of the illustrative embodiments may be implemented.FIGS. 1-3 are only exemplary and are not intended to state or imply any limitation with regard to the types and/or configurations of computing devices in which the illustrative embodiments may be implemented. Many modifications to the computing devices and environments depicted inFIGS. 1-3 may be made without departing from the spirit and scope of the present invention. - With reference now to the figures,
FIG. 1 depicts a pictorial representation of a network of data processing systems in which the present invention may be implemented. Networkdata processing system 100 is a network of computers in which the present invention may be implemented. Networkdata processing system 100 contains anetwork 102, which is the medium used to provide communications links between various devices and computers connected together within networkdata processing system 100.Network 102 may include connections, such as wire, wireless communication links, or fiber optic cables. - In the depicted example,
server 104 is connected to network 102 along withstorage unit 106. In addition,clients clients server 104 provides data, such as boot files, operating system images, and applications to clients 108-112.Clients server 104. Networkdata processing system 100 may include additional servers, clients, and other devices not shown. In the depicted example, networkdata processing system 100 is the Internet withnetwork 102 representing a worldwide collection of networks and gateways that use the Transmission Control Protocol/Internet Protocol (TCP/IP) suite of protocols to communicate with one another. At the heart of the Internet is a backbone of high-speed data communication lines between major nodes or host computers, consisting of thousands of commercial, government, educational and other computer systems that route data and messages. Of course, networkdata processing system 100 also may be implemented as a number of different types of networks, such as for example, an intranet, a local area network (LAN), or a wide area network (WAN).FIG. 1 is intended as an example, and not as an architectural limitation for the present invention. - Referring to
FIG. 2 , a block diagram of a data processing system that may be implemented as a server, such asserver 104 inFIG. 1 , is depicted in accordance with a preferred embodiment of the present invention.Data processing system 200 may be a symmetric multiprocessor (SMP) system including a plurality ofprocessors system bus 206. Alternatively, a single processor system may be employed. Also connected tosystem bus 206 is memory controller/cache 208, which provides an interface tolocal memory 209. I/O Bus Bridge 210 is connected tosystem bus 206 and provides an interface to I/O bus 212. Memory controller/cache 208 and I/O Bus Bridge 210 may be integrated as depicted. - Peripheral component interconnect (PCI)
bus bridge 214 connected to I/O bus 212 provides an interface to PCIlocal bus 216. A number of modems may be connected to PCIlocal bus 216. Typical PCI bus implementations will support four PCI expansion slots or add-in connectors. Communications links to clients 108-112 inFIG. 1 may be provided throughmodem 218 andnetwork adapter 220 connected to PCIlocal bus 216 through add-in connectors. - Additional
PCI bus bridges local buses data processing system 200 allows connections to multiple network computers. A memory-mappedgraphics adapter 230 andhard disk 232 may also be connected to I/O bus 212 as depicted, either directly or indirectly. - Those of ordinary skill in the art will appreciate that the hardware depicted in
FIG. 2 may vary. For example, other peripheral devices, such as optical disk drives and the like, also may be used in addition to or in place of the hardware depicted. The depicted example is not meant to imply architectural limitations with respect to the present invention. - The data processing system depicted in
FIG. 2 may be, for example, an IBM eServer pSeries system, a product of International Business Machines Corporation in Armonk, N.Y., running the Advanced Interactive Executive (AIX) operating system or LINUX operating system. - With reference now to
FIG. 3 , a block diagram illustrating a data processing system is depicted in which the present invention may be implemented.Data processing system 300 is an example of a stand-alone or client computing device.Data processing system 300 employs a peripheral component interconnect (PCI) local bus architecture. Although the depicted example employs a PCI bus, other bus architectures such as Accelerated Graphics Port (AGP) and Industry Standard Architecture (ISA) may be used.Processor 302 andmain memory 304 are connected to PCIlocal bus 306 throughPCI Bridge 308.PCI Bridge 308 also may include an integrated memory controller and cache memory forprocessor 302. Additional connections to PCIlocal bus 306 may be made through direct component interconnection or through add-in boards. In the depicted example, local area network (LAN)adapter 310, small computer system interface (SCSI)host bus adapter 312, andexpansion bus interface 314 are connected to PCIlocal bus 306 by direct component connection. In contrast,audio adapter 316,graphics adapter 318, and audio/video adapter 319 are connected to PCIlocal bus 306 by add-in boards inserted into expansion slots.Expansion bus interface 314 provides a connection for a keyboard andmouse adapter 320,modem 322, andadditional memory 324. SCSIhost bus adapter 312 provides a connection forhard disk drive 326,tape drive 328, and CD-ROM drive 330. Typical PCI local bus implementations will support three or four PCI expansion slots or add-in connectors. - An operating system runs on
processor 302 and is used to coordinate and provide control of various components withindata processing system 300 inFIG. 3 . The operating system may be a commercially available operating system, such as Windows XP, which is available from Microsoft Corporation. An object oriented programming system such as Java may run in conjunction with the operating system and provide calls to the operating system from Java programs or applications executing ondata processing system 300. “Java” is a trademark of Sun Microsystems, Inc. Instructions for the operating system, the object-oriented programming system, and applications or programs are located on storage devices, such ashard disk drive 326, and may be loaded intomain memory 304 for execution byprocessor 302. - Those of ordinary skill in the art will appreciate that the hardware in
FIG. 3 may vary depending on the implementation. Other internal hardware or peripheral devices, such as flash read-only memory (ROM), equivalent nonvolatile memory, or optical disk drives and the like, may be used in addition to or in place of the hardware depicted inFIG. 3 . Also, the processes of the present invention may be applied to a multiprocessor data processing system. - As another example,
data processing system 300 may be a stand-alone system configured to be bootable without relying on some type of network communication interfaces As a further example,data processing system 300 may be a personal digital assistant (PDA) device, which is configured with ROM and/or flash ROM in order to provide non-volatile memory for storing operating system files and/or user-generated data. - The depicted example in
FIG. 3 and above-described examples are not meant to imply architectural limitations. For example,data processing system 300 also may be a notebook computer or hand held computer in addition to taking the form of a PDA.Data processing system 300 also may be a kiosk or a Web appliance. - With regard to the illustrative embodiments of the present invention, the client/stand-alone computing device or server computing device executes software applications which provide a virtual world environment which is accessible by a user. For example, a user may interact with the virtual world via one or more input devices connected to the computing devices or connected to a client device that communicates with the computing device. One of these input devices, in accordance with the illustrative embodiments, is a detection mechanism for detecting the presence of real world objects. Such a detection mechanism may use any detection scheme suitable to the particular implementation including bar code scanners, radio frequency identifier (RFID) detection devices, dot code detection devices, and the like.
- In a preferred embodiment, the detection mechanism is provided as a detection pad upon which real world objects may be placed and whose presence may be detected using a RFID detection device integrated into the detection pad. The detection pad has an associated detection area that comprises the pad itself and a limited range of space around the pad in which real world objects having associated RFID tags may be detected.
- With regard to the illustrative embodiments herein, the detection mechanism, e.g., the detection pad, provides a mechanism by which the presence of a real world object is detected and, while the real world object's presence is continued to be detected, it is made available for use in a virtual world environment. Moreover, the detection of the presence of the real world object provides an identifier of the object which may be correlated with information regarding how to represent the real world object in the virtual world environment, how the object may be utilized in relation to other objects in the virtual world environment such that the real world object is modeled in the virtual world environment, characteristics of the virtual world object representation of the real world object with regard to the particular applications in which the virtual world object will be used in the virtual world environment, and the like. Furthermore, the detection of multiple real world objects may be performed and identification of each of the multiple real world objects may be used to determine how these real world objects may be utilized together in the virtual world environment.
- One significant application of the illustrative embodiments described herein is to the field of computer gaming. Moreover, within the field of computer gaming, the illustrative embodiments are especially well suited for use with persistent virtual world type computer games. While the illustrative embodiments will be described in terms of computer games and persistent virtual world computer games, it should be appreciated that the illustrative embodiments are not limited to being applicable to only computer games. To the contrary, the illustrative embodiments may be used with any application in which real world objects may be rendered as virtual objects in a virtual environment.
- In one illustrative embodiment, a computing device executes operations for monitoring a detection mechanism, which in the illustrative embodiments set forth herein is a detection pad, to detect when one or more real world objects are placed on the detection pad or within the detection area of the detection pad. The operations may continue to monitor the detection pad continuously or periodically to determine a current set of one or more real world objects that are within the detection area of the detection pad and compare the current set of real world objects to a previously detected set of real world objects to determine which objects have persistent presence and which objects are no longer within the detection area. Only real world objects that are present within the detection area of the detection pad may be modeled and represented in the virtual world environment. Thus, when a real world object is removed from the detection area of the detection pad, the object's representation in the virtual world environment is removed. In this way, the persistent presence of a real world object as detected by the detection pad is mapped into a virtual world object awareness.
- In the illustrative embodiments, the detection pad is coupled to a stand-alone or client computing device. The virtual world environment is provided either directly by the stand-alone or client computing device, by one or more server computing devices with which the client computing device communicates, or by a combination of software applications running on the client computing device and one or more applications running on one or more server computing devices. Processing of information from the detection pad is first performed by the stand-alone or client computing device, using driver software and the like, for example, which converts inputs from the detection pad into data that is understandable by one or more software applications running on the client and/or server computing devices. In particular, the driver software receives inputs from the detection mechanisms of the detection pad, such as RFID detectors of the detection pad, and provides data to one or more software applications corresponding to the inputs received from the detection pad.
- In the illustrative embodiments, the inputs received from the detection pad are representative of the detection of particular real world objects being within a detection area of the detection pad. For example, the RFID detectors of the detection pad may detect RFID tags associated with real world objects that are placed within the detection area of the detection pad and provide as inputs to the stand-alone or client computing device, the corresponding identifiers associated with the RFID tags detected. These identifiers may then be correlated with object information for associated real world objects, such as through a database lookup operation.
- The real world objects preferably are objects that are to be represented by similar virtual objects that are modeled in a virtual world environment within the computing device. For example, the real world objects may be toys, e.g., action figures, toy vehicles, and the like, toy accessories, e.g., toy weapons, toy armor, and the like, models, etc. of objects that may be represented as virtual objects in the virtual world environment and used within the virtual world environment. When the real world object is present within the detection area of the detection pad, the real world object, e.g., the action figure, toy vehicle, toy weapon, etc., is represented in the virtual world in such a manner that is consistent with the usage of the object in the real world.
- For example, if the real world object is a toy representation of a sword, the detection of the toy sword in the detection area of the detection pad causes a corresponding virtual sword to be represented in the virtual world environment. Moreover, the virtual sword is represented in the virtual environment in a manner consistent with the usage of a sword in the real world. Thus, if a user's virtual world avatar currently is involved in a combat situation, the virtual sword may be represented as being in a free hand of the avatar. If the user's virtual world avatar is currently not involved in combat, the virtual sword may be represented as being sheathed and hanging on the side of the user's avatar. The representation of the virtual sword in the virtual world environment may be maintained as long as the toy sword is detected as being present in the detection area of the detection pad. If the user were to remove the toy sword from the detection area of the detection pad, when the absence of the toy sword is detected by the detection pad, the virtual world representation of the toy sword is discontinued.
- As mentioned above, the detection pad may detect multiple real world objects as being present within the detection area of the detection pad. The identification of the various real world objects in the detection area is used to retrieve information regarding the real world objects and how to represent the real world objects as virtual objects in a virtual world environment. As part of the determination as to how to represent the real world objects as virtual objects in the virtual world environment, the interaction between the identified real world objects may be determined using a rules database.
- For example, assume that the real world objects on a detection pad comprise an action figure representing a knight and accessory objects that comprise a toy sword, a toy helmet, and a toy shield. Information corresponding to each of these detected items may be retrieved from a database associated with an application running in the computer system. This information may include an identification of how each item is utilized in the real world and how such usage is mapped to the virtual environment. Alternatively, a rules database may be provided that provides rules based on the detected real world object identifiers. For example, a rule may state that if a first real world object and a second real world object are detected on the detection pad, then the virtual object corresponding to the second real world object should be rendered as being in a particular association with the first real world object. Other such rules, which may be simpler or more complicated than the example rule given above, may be provided within a rules database that operate on the detected real world object identifiers and retrieved information corresponding to the real world object identifiers to determine how the virtual objects should be rendered in the virtual world environment.
- For example, the toy sword (e.g., a second real world object) is used in the hand of an action figure as a weapon and has certain attack and defense characteristics. As a result, when representing the toy sword in conjunction with the action figure (e.g., a first real world object) as virtual objects in the virtual world environment, a virtual object corresponding to the toy sword may be represented in the hand or sheath of an avatar resembling the action figure that is present on the detection pad. Similarly, the toy shield, may be represented as a virtual shield object in a hand of the virtual action figure, i.e. the avatar, or on the back of the virtual action figure in the virtual world environment. The toy helmet is used on the head of the action figure and thus, is represented in the virtual world environment as a helmet object placed on the head of the virtual action figure. Thus, the proper interaction of the detected real world objects in the detection area of the detection pad is automatically determined and used to represent the real world objects as virtual objects in a manner consistent with the way in which these real world objects interact with each other in the real world.
- Furthermore, the operations in the computing device may monitor a condition of the virtual world object representation of the real world object within the virtual world environment and, based on the condition of the virtual world object representation of the real world object, may disable the user's ability to use the real world object to generate a virtual world representation of the real world object. For example, if the real world object were a toy sword and the user, in a previous session in the virtual world environment, performed actions that resulted in his/her avatar's virtual sword, which corresponds to then real world toy sword, being damaged or broken, then the user's ability to place the toy sword on the detection pad and have a corresponding sword represented in the virtual world environment may be discontinued until the user performs appropriate actions in the virtual world environment, e.g., repairs the virtual representation of the sword in the virtual world environment.
- Alternatively, a persistent state of the virtual representation of the real world object may be maintained in the virtual world environment such that each time a real world object is placed in the detection area of the detection pad, the persistent state of the virtual object corresponding to that real world object may be retrieved and used to render the virtual object in the virtual world environment. For example, using the damaged sword example discussed above, rather than discontinuing the ability of the user to generate a virtual object corresponding to the real world toy sword object on the detection pad, the previous state of the virtual sword may be retrieved and used to render the virtual sword representation of the real world toy sword. Thus, if a user, during a previous session in the virtual world environment, damages his virtual sword but fails to repair it, when the user next places the real world toy sword on the detection pad, the virtual sword object rendered in response to detecting the real world toy sword will be rendered as having the same damaged state as when the user last used that real world toy sword object with the virtual world environment.
-
FIG. 4 is an exemplary block diagram illustrating a mechanism for detecting a persistent presence of real world objects such that the real world objects may be represented as virtual world objects in a virtual world environment in accordance with one illustrative embodiment. The mechanism illustrative inFIG. 4 is an example of a stand-alone computing device. It should be appreciated that the stand-alone computing device inFIG. 4 is only exemplary of one illustrative embodiment and is not intended to state or imply any limitation with regard to the types of computing devices in which the present invention may be implemented. As mentioned previously above, the mechanisms of the illustrative embodiments may be distributed amongst a plurality of computing devices in a distributed data processing environment. - As shown in
FIG. 4 , the mechanism includes adetection pad 410 and acomputing device 420.Elements FIG. 4 may be implemented in hardware, software, or any combination of hardware and software without departing from the spirit and scope of the present invention. - The
detection pad 410 includes one or more detectors, which may be integrated into thedetection pad 410, for detecting the presence of real world objects within a detection area of thedetection pad 410. Thedetection pad 410 transmits information to thecomputing device 420 either through a wired orwireless communication link 415. The information transmitted to the computing device corresponds to real world object identifiers, corresponding to identification devices that are affixed to or integrated with the real world objects, and which are detected by the detectors in thedetection pad 410. This information may be detected by way of a barcodes provided on the real world objects, RFIDs, dot codes, or the like. In a preferred embodiment, the real world objects are provided with RFIDs and the detector in thedetection pad 410 is configured to scan a detection area and detect the presence of RFIDs within the detection area. The use of RFIDs and RFID detectors is generally known in the art and thus, a detailed description is not provided herein. - The information transmitted to the
computing device 420 from thedetection pad 410 is input to the input/output interface 430. The input/output interface 430 preferably utilizes detection pad driver(s) 435 to interpret the inputs from thedetection pad 410 and provide the corresponding data to the real worldobject persistence module 440. - The real world
object persistence module 440 is responsible for determining what real world objects are currently present in the detection area of thedetection pad 410 based on inputs from thedetection pad 410, what real world objects have been removed from the detection area of thedetection pad 410, as well as updating information in a detected objectsdatabase 450. The real worldobject persistence module 440 communicates information regarding the real world objects that are currently present within the detection area of thedetection pad 410 with the virtual world environment objectawareness module 460. - The virtual world environment object
awareness module 460 is responsible for rendering virtual objects corresponding to the detected real world objects that are currently in the detection area of thedetection pad 410 in the virtual world environment. The information communicated to the virtual world environment objectawareness module 460 from the real worldobject persistence module 440 may be used to retrieve mapping information from the real world object/virtualobject mapping database 470. The virtual world environment objectawareness module 460 may utilize this mapping information to render virtual objects corresponding to the detected real world objects in the virtual world environment generated by the virtual world environment application(s) 480. - It should be noted that, because the depicted illustrative embodiment utilizes a stand-
alone computing device 420, all of the elements 430-480 are shown as being part of thecomputing device 420. However, the present invention is not limited to such a configuration. Rather, many of the elements 430-480 may be provided in one or more other computing devices with which thecomputing device 420 may communicate via one or more networks. For example, elements 460-480 may be provided by a server computing device with which thecomputing device 420 is a client. In such an embodiment, theclient computing device 420 is responsible for determining the real world aspects of the real world persistence to virtual world awareness mapping while the server computing device is responsible for the virtual world aspects of this mapping. - When a real world object having an associated identifier device associated with it, e.g., a barcode, dot code, RFID tag, or the like, is placed within the detection area of the
detection pad 410, the presence of the real world object is detected by the detectors in thedetection pad 410. In a preferred embodiment, the identifier device is permanently affixed to, or integrated with, the real world object, although this is not required for the functioning of the illustrative embodiments. Such permanent affixing or integration is intended to ensure that the real world object is within the detection area and not just the identifier device. - The detected identifier associated with the real world object is transmitted from the
detection pad 410 to thecomputing device 420 via thecommunication link 415. The signals transmitted by thedetection pad 410 are processed via the I/O interface 430 using the detection pad driver(s) 435. The resulting identifier information is provided to real worldobject persistence module 440. - The real world
object persistence module 440 receives the identifier information and performs a lookup of the identifier information in the detectedobjects database 450. The detected objectsdatabase 450 preferably has two primary portions. In afirst portion 452, the detectedobjects database 450 maintains a current real world objects data structure that has entries for each real world object that is detected as being currently present within the detection area of thedetection pad 410. In asecond portion 454, the detectedobjects database 450 maintains a current status of all real world objects that have been detected using thedetection pad 410. Thesecond portion 454 maintains information regarding real world objects that are currently present within the detection area of thedetection pad 410 as well as real world objects that are not currently detected as being within the detection area but have been detected by thedetection pad 410 previously. - In response to receiving the real world object identification information, a determination is made as to whether an entry for that particular real world object is currently present in the
first portion 452 of the detectedobjects database 450. If not, a new entry in thefirst portion 452 is generated. The entry may be indexed by the identifier associated with the real world object and may contain, for example, information identifying a current state of the virtual world object that corresponds to the real world object, which may be determined based upon previous state information retrieved from thesecond portion 454. This state information may be updated periodically, or when a change in state of the virtual world object that corresponds to the real world object occurs in the virtual world environment provided by application(s) 480. - If an entry does not already exist in the
first portion 452 of the detectedobjects database 450, then the real worldobject persistence module 440 sends a message to the virtual world environment objectawareness module 460 indicating that a new real world object has been detected and provides the identification information for the new real world object. In response, the virtual world environment objectawareness module 460 maps the real world object identifier to a virtual object that is to be represented in the virtual world environment. This mapping is performed using mapping information retrieved from the real world object/virtualobject mapping database 470. The information maintained in the real world object/virtualobject mapping database 470 includes pointer information identifying the data, files, etc. corresponding to a virtual world representation of the real world object, i.e. the virtual object, and pointer information for data, files, etc. that identify the manner by which the virtual object is utilized within the virtual world environment. These data, files, etc. preferably model the real world object as a virtual object in such a manner that the virtual world object emulates the real world object not only in appearance but in the manner by which the real world object is utilized with other objects in the real world. - The virtual world environment object
awareness module 460 interfaces with the virtual world environment provided by the virtual world environment application(s) 480 running on thecomputing device 420. The virtual world environment objectawareness module 460 informs the virtual world environment application(s) of the virtual objects that correspond to real world objects whose presence is currently detected by thedetection pad 410 as well as the necessary information for rendering the virtual objects and modeling the virtual objects with regard to the virtual objects' interaction with the virtual world. In this way, a virtual object corresponding to a detected real world object may be generated and utilized in a virtual world environment. - In addition to registering real world objects that are detected using the
detection pad 410, the real worldobject persistence module 440 also checks thefirst portion 452 of the detectedobjects database 450 to determine if there are any entries for real world objects that are no longer detected as being within the detection area of thedetection pad 410. That is, the identifier information obtained based on the inputs from thedetection pad 410 are compared to the entries in thefirst portion 452 of the detectedobjects database 450 as previously described above. - Entries that are created or already existing in the
first portion 452 that match the identifiers of the real world objects detected using thedetection pad 410 are temporarily marked, e.g., a bit flag is set, to thereby identify the entry as being associated with a real world object that is currently present in the detection area of thedetection pad 410. The real worldobject persistence module 440 may then check the entries in thefirst portion 452 of the detectedobjects database 450 and thereby identify which entries are not marked as being associated with real world objects that are currently detected in the detection area of thedetection pad 410. The real world objects associated with these entries are real world objects that have been removed from the detection area of thedetection pad 410. - The real world
object persistence module 440 may communicate with the virtual world environment objectawareness module 460 to inform the virtual world environment objectawareness module 460 that the virtual objects corresponding to the real world objects that were removed from the detection area should be removed from the virtual world environment. In response to such a communication, the virtual world environment objectawareness module 460 sends a message to the virtual world environment application(s) 480 to discontinue the rendering of the virtual objects corresponding to the removed read world objects. - The above described operation for detecting real world objects and identifying which real world objects have been removed from a detection area may be performed continuously or periodically. In this way, only those real world objects whose presence is persistent as detected by the detection pad may have their corresponding virtual objects continue to be rendered and utilized in the virtual world environment. Thus, a real world object persistence is mapped to a virtual world object awareness.
- In a further illustrative embodiment, the virtual world environment object
awareness module 460 determines the manner by which a plurality of real world objects detected within the detection area of thedetection pad 410 interact with one another or are utilized with each other, if at all, so that they may be accurately represented as a combination of virtual objects in the virtual world environment. For example, if the real worldobject persistence module 440 receives identifier information for an action figure, a helmet for the action figure, a sword for the action figure, and a shield for the action figure, the virtual world environment objectawareness module 460 determines how these objects may be utilized together based on information maintained within the real world object/virtualobject mapping database 470. For example, the virtual world environment objectawareness module 460 may determine through a lookup of the real world object identifiers in the real world object/virtualobject mapping database 470 that the sword should be placed in a right hand of the action figure, the helmet should be placed on a head of the action figure, and the shield should be placed in the left hand of the action figure. - Based on this determination, the virtual world environment object
awareness module 460 may generate a virtual avatar in the virtual world environment, based on the information retrieved from the real world object/virtualobject mapping database 470, and may place a sword virtual object corresponding to the real world sword object in the right hand of the virtual avatar, a shield virtual object corresponding to the real world shield object in the left hand of the virtual avatar, and a helmet virtual object corresponding to the helmet real world object on the head of the virtual avatar. The user may then utilize these objects within the virtual world environment in a manner corresponding to the models of these objects within the virtual world. The modeling of virtual objects in a virtual world environment is generally known in the art and thus, a detailed description is not provided herein. - Thus, the virtual world environment object
awareness module 460 is able to determine the manner by which each of the accessory real world objects, i.e. the sword, shield and helmet, are utilized with or interact with the action figure real world object. This interaction is then modeled in the virtual world environment by providing a combination of virtual objects that are modeled after the manner by which the detected real world objects are utilized in the real world. Of course for objects that do not actually operate in the real world, e.g., toys that represent fantasy type items, such as wands, magical creatures, and the like, the utilization of the objects in the virtual world may be completely based upon the modeling specified by the programmer which may or may not have any relation to the real world. - In yet a further illustrative embodiment, the current status of a virtual object in the virtual world environment that corresponds to a real world object detected using the
detection pad 410 may be persisted and associated with the real world object such that when the real world object is detected again by thedetection pad 410, the last state of the virtual object corresponding to the real world object may be retrieved. As an example, assume that a real world object that is detected by thedetection pad 410 corresponds to a toy sword for an action figure. The toy sword is rendered in the virtual world environment as a virtual sword object that may be utilized by a virtual avatar to battle enemies in the virtual world environment. During the process, the virtual sword object may sustain damage that causes the virtual sword object to be less effective or need repair. Such a status may be determined and maintained in the virtual world environment. - In response to a determination that the real world toy sword object is removed from the detection area of the
detection pad 410, or the user discontinuing a session within the virtual world environment, the current status of the virtual sword object may be reported back to the real worldobject persistence module 440. The real worldobject persistence module 440 may then store this current status in thesecond portion 454 of the detectedobjects database 450 in associated with the identifier of the real world toy sword object. In this way, when the real world toy sword object is again placed within the detection area of thedetection pad 410, the real worldobject persistence module 440 may retrieve the current status information from thesecond portion 454 of the detectedobjects database 450 and provide the current status information to the virtual world environment objectawareness module 460. The virtual world environment objectawareness module 460 may then use this current status information when rendering and modeling the virtual sword object in the virtual world environment. - In one illustrative embodiment, the current status information may identify any status that may be modeled within the virtual world environment, e.g., damaged, needing repair, ineffective, destroyed, etc. In another illustrative embodiment, the current status information may simply identify real world objects that may be rendered as virtual objects within the virtual world environment and real world objects that cannot until some action in the virtual world environment is taken. For example, if virtual sword object in the virtual world environment is broken within the virtual world environment, this status may be stored in association with the real world toy sword object and thus, when the real world toy sword object is placed within the detection area of the detection pad, the virtual sword object is not rendered until the user visits a blacksmith in the virtual world environment and has the virtual sword object repaired. Thus, not only is the presence of the real world objects within the detection area of the detection pad made persistent, but the virtual status of the virtual objects corresponding to the real world objects is persisted in association with the real world objects.
-
FIG. 5A is pictorial diagram illustrating the rendering of virtual objects in a virtual world environment that correspond to real world objects that are detected as being present in a detection area of a detection pad in accordance with an illustrative embodiment. As shown inFIG. 5A , an action figure 510 representing a knight is placed on adetection pad 550 along with accessories for the action figure that include atoy shield 520, atoy sword 530, and atoy helmet 540. Each of these real world objects has an associated RFID tag attached or integrated into them that identifies the real world objects. Thedetection pad 550 detects the presence of these real world objects and transmits the identification information to thecomputing device 560. - In an illustrative embodiment, the
computing device 560 runs software modules corresponding to elements 430-480 ofFIG. 4 . Thus, thecomputing device 560, using the various operations described above with regard toFIG. 4 , determines how to render thevirtual avatar 570 in thevirtual world environment 580 output using the display of thecomputing device 560. As shown, thevirtual avatar 570 includes a helmetvirtual object 572 placed on a head location of thevirtual avatar 570, a swordvirtual object 574 in a right hand of thevirtual avatar 570, and a shieldvirtual object 576 in a left hand of thevirtual avatar 570. Because the shieldvirtual object 576 had been used in a previous session and was damaged, the rendering of theshield object 576 shows the shieldvirtual object 576 as being damaged. -
FIG. 5B is pictorial diagram illustrating the discontinuing of the rendering of a virtual object in a virtual world environment when its corresponding real world object is no longer detected as being present in a detection area of a detection pad in accordance with an illustrative embodiment. As shown inFIG. 5B , the toy shieldreal world object 520 has been removed from the detection area of thedetection pad 550. In response, thecomputing device 560 detects the removal of the toy shieldreal world object 520 and discontinues rendering of the shieldvirtual object 576 in thevirtual world environment 580. Thus, thevirtual avatar 570 is shown in thevirtual world environment 580 as having only the helmetvirtual object 572 and the swordvirtual object 574. -
FIG. 6 is a flowchart outlining an exemplary operation of a real world object persistence to virtual object awareness mapping mechanism in accordance with an illustrative embodiment. It will be understood that each block of the flowchart illustration, and combinations of blocks in the flowchart illustration, can be implemented by computer program instructions. These computer program instructions may be provided to a processor or other programmable data processing apparatus to produce a machine, such that the instructions which execute on the processor or other programmable data processing apparatus create means for implementing the functions specified in the flowchart block or blocks. These computer program instructions may also be stored in a computer-readable memory or storage medium that can direct a processor or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory or storage medium produce an article of manufacture including instruction means which implement the functions specified in the flowchart block or blocks. - Accordingly, blocks of the flowchart illustration support combinations of means for performing the specified functions, combinations of steps for performing the specified functions and program instruction means for performing the specified functions. It will also be understood that each block of the flowchart illustration, and combinations of blocks in the flowchart illustration, can be implemented by special purpose hardware-based computer systems which perform the specified functions or steps, or by combinations of special purpose hardware and computer instructions.
- As shown in
FIG. 6 , the operation starts with the detection of real world objects using the detection mechanism, e.g., the detection pad (step 610). The identifiers of the real world objects that are detected are provided to the real world object persistence module (step 620). A determination is made as to whether an entries are already existing in the current detected objects data structure for the detected real world objects (step 630). If not, an entry for each detected real world object that does not already have an entry in the current detected objects data structure is generated (step 640). A previous state associated with the detected real world objects that did not have an entry is retrieved from the detected objects data structure if such a previous state exists (step 650). - Thereafter, or if all of the detected real world objects have associated entries in the current detected objects data structure, the real world objects that have been removed from the detection area of the detection mechanism are identified (step 660). One or more messages are sent to the virtual world environment object awareness module identifying detected real world objects and their associated states, if any (step 670). A determination is made as to whether any real world objects were removed from the detection area of the detection mechanism (step 680). If so, then the current state of the virtual objects corresponding to the removed real world objects is stored in association with the identifiers of the real world objects that were removed (step 690). The representation of the virtual objects corresponding to the removed real world objects is then discontinued in the virtual world environment (step 700).
- Thereafter, or if no real world objects were removed from the detection area of the detection mechanism, information is retrieved from the real world object/virtual object mapping database for the currently detected real world objects (step 710). Virtual objects corresponding to the currently detected real world objects are generated in the virtual world environment based on information from the mapping database, state information associated with the real world objects, and the like (step 720). As previously noted above, the generation of such virtual objects may further involve determining the manner by which the real world objects interact with each other or otherwise are related to each other and then depicting the virtual objects in a manner consistent with such interaction or relation.
- A determination is made as to whether the session with the virtual world environment has been terminated (step 730). If not, the operation returns to step 610 and is repeated. If the session has been terminated, the current state of the virtual objects corresponding to the currently detected real world objects is stored in association with the identifiers of the real world objects (step 740). The operation then terminates.
- The above illustrative embodiments describe the detection mechanism as primarily being an input device to a computing system. It should be noted that the detection mechanism may further be used as an output device as well and may be configured to include output mechanisms for outputting information to a user in a visual, auditory, and/or tactile manner. In particular, as an output device, the detection mechanism may be configured to provide output to the user so as to guide the user in the use of the detection mechanism with the real world objects that are placed within the detection area of the detection mechanism.
- For example, as mentioned above, the detection mechanism of the illustrative embodiments, e.g., the detection pad, may have a plurality of detectors or sensors that are used to detect or sense the presence of a real world object within a detection area of the detection mechanism. These detectors or sensors may be configured in any manner suitable to the particular implementation of the illustrative embodiments selected. In one illustrative embodiment, these detectors/sensors may be configured in a grid-like manner within the detection mechanism. Moreover, as shown in
FIG. 7 , the detection mechanism may have visual output elements that are arranged in a similar grid-like manner that may be selectively activated so as to inform a user of areas of the detection mechanism where certain real world objects may be placed. - For example, based on inputs to the detection pad generated by a virtual world environment software program, various ones of the visual output elements may be energized or otherwise made more noticeable (hereafter referred to as being “highlighted”) in accordance with a desired
output pattern 750 on the detection pad. This output pattern may be representative, for example, of the types of real world objects that may be placed on the detection pad and used with the virtual world environment. In one illustrative embodiment, the visual output elements may be light emitting diodes (LEDs) or other types of light generating devices that may be selectively activated so as to display a pattern on the detection pad. The particular shapes that may be generated using such visual output elements are not limited by the illustrative embodiments herein and may encompass any shape that may be represented using a grid of visual output elements. - For example, a human pattern may be highlighted or outlined for an adventure or role playing game virtual world environment. For a trading card based game, card placement blocks may be highlighted signifying different play areas in which real world objects may be placed. Moreover, the particular areas of the detection pad that are highlighted may be associated with activated detectors or sensors such that portions of the detection pad that are not highlighted are not active. In this way, selected portions of the detection pad may be activated, depending upon the particular virtual world environment with which the detection pad is being utilized, and may be used to detect the presence of real world objects while other areas are de-activated and will not detect the presence of real world objects even if the real world objects are present in these areas.
- The use of a grid-like array of detectors/sensors allows the possibility to associate coordinates with real world objects placed within the detection field. That is, a coordinate system may be established with the detectors/sensors of the detection pad, such as a lower left corner of the detection pad being at coordinates (0,0), and this coordinate system may be used to determine the relative placement of real world objects on the detection pad. The ability to associated coordinates with real world objects aids in the relative rendering of virtual objects in the virtual world environment that correspond to real world objects placed on the detection pad. For example, if the virtual world environment represents a strategy board game, the relative placement of real world pieces may be determined using the coordinate system of the detection pad and used to render virtual objects corresponding to those pieces with the same relative placement in the virtual world environment.
-
FIGS. 8A-8C illustrate some exemplary applications of the detection mechanism of the illustrative embodiments. As shown inFIG. 8A , thedetection mechanism 800 of the illustrative embodiments may be used to highlight active areas 810-850 of thedetection mechanism 800 in which playing cards for a collectable card game may be placed. For example, active areas 810-840 represent cards that are in play whileactive area 850 represent the active deck for the card game. Cards that have been played and that are no longer in play may be removed from thedetection pad 800, as shown by playeddeck 860. The cards that are in play in active areas 810-850 are detected by the detectors associated with these active areas and corresponding virtual objects may be generated in a virtual world environment represented in a computing device. These virtual objects may be virtual cards or may be virtual objects corresponding to items, people, creatures, or the like, that are represented by the real world cards. -
FIG. 8B illustrates anotherdetection mechanism 860 that is configured for use with large board games such as strategy games or the like. As shown, thedetection mechanism 860 has afirst portion 862 that comprises the detection array, i.e. the array or detectors/sensors for detecting the presence of real world objects. Thefirst portion 862 may further comprise the visual output elements described previously above. Asecond portion 864 may comprise a display device that is capable of generating graphical displays, e.g., a liquid crystal display, plasma display, or the like. Thesecond portion 864 may be used to output a representation of a playing field upon which real world objects may be placed while thefirst portion 862 detects the presence of these real world objects on the detection mechanism. The playing field may be represented as graphical images, such as of rivers, mountains, and other terrain, and may be generated based on outputs from the virtual world environment software running on the computing device to which thedetection mechanism 860 is connected, or the like. The presence of the real world objects on thedetection mechanism 860 may be translated into a virtual object awareness in the virtual world environment generated by such software. For example, while thedetection mechanism 860 may represent a two-dimensional playing field, the computing device in which the virtual world environment is rendered may represent the two-dimensional playing field in three dimensions. - Thus, the
second portion 864 provides a mechanism by which images and/or topological information may be overlaid on the detection mechanism's surface to convey detail that enhances the software program with which the detection mechanism is being used. In a further example, in addition to highlighting/activating portions of the detection mechanism based on the particular programs with which the detection mechanism interfaces as described above, images may be displayed within these highlighted/activated areas that depict the types of real world objects that should be placed in these highlighted/activated areas. For example, an image of a playing card may be displayed within the highlighted areas 810-840 ofFIG. 8A . - The various illustrative embodiments described above are primarily directed to computer game virtual world applications, however the illustrative embodiments are not limited to computer gaming. To the contrary, there are numerous non-gaming applications with which the illustrative embodiments may be utilized. For example,
- As a non-gaming related example, using the highlighting/activation of grid patterns or locations on the detection mechanism, as described above, a parts list for products requiring assembly may be displayed and the presence of these parts may be detected. For example a product may be purchased that requires assembly. The pieces of the product may have identifier tags embedded or otherwise attached to them and may be deposited on the presence detection mechanism. An assembly program with which the detection mechanism interfaces may be run on an associated computing device and, for each step of the assembly instructions, the grid location corresponding to the component to be assembled/used next in the series of assembly steps may be highlighted. The assembly program may detect the presence of the pieces via the detection mechanism and represent the pieces in a virtual world environment in which the various steps for assembly of the product are virtually represented. Moreover, the assembly program may perform a check of the complete inventory of items against the assembly program before assembly begins to ensure that all the necessary pieces are present.
- As a further non-gaming related example implementation of the illustrative embodiments, the detection mechanism may be used to represent clothing in a virtual world environment, such as on a virtual mannequin, as illustrated in
FIG. 8C . For example, a user may register with a computing system and provide general dimensions for the user. The user may register with the computing system either directly in the clothing store where the real world clothing objects are present or may register on-line from another location, for example. - Having registered with the computing system, the user may select various pieces of clothing 870-880 and place them on the
detection mechanism 890, which may be oriented horizontally, vertically, or the like. Thedetection mechanism 890 may detect the presence of these clothing items 870-880 and, through the mechanisms previously described above, determine how to represent these real world objects as virtual objects in the virtual world environment in which the virtual mannequin is present. In essence, the detected pieces of clothing may be mapped as virtual clothing onto the virtual mannequin that has the dimensions of the user. The virtual mannequin with the virtual clothing mapped thereon may be displayed to the user via adisplay device 895 so that the user may view the results of the mapping. In this way, the user may be able to visualize the way in which the clothing may look on their frame without having to actually try on the clothing. Thus, with this implementation, the illustrative embodiments help to alleviate the cumbersome nature of tying on clothes and reduce the work to merely matching clothing items and selecting appropriate sizes. - As can be seen from the various implementations of the illustrative embodiments described above, the size of the detection mechanism, e.g., the detection pad, is not limited by the illustrative embodiments described herein. To the contrary, the size of the detection pad may be determined based on the particular implementation and intended use of the detection pad. For example, if the detection pad will be used primarily with a strategy game virtual world environment, in which a large number of pieces may be placed on a large game board, the game pad may be large enough to cover a typical dining room table or coffee table. On the other hand, if the detection pad is to be used with a card game virtual environment, then the detection pad may be quite small in size relative to a dining room table or coffee table.
- Thus, the illustrative embodiments provide a mechanism for detecting the persistent presence of real world objects and translating this persistent presence into a virtual object awareness in a virtual world environment. With the mechanisms of the illustrative embodiments virtual objects corresponding to currently detected real world objects may be generated based on detected identifiers of the real world objects. These virtual objects may be generated in such a manner as to be consistent with the intended interaction or relationships of the real world objects. When a real world object is detected as having been removed from the detection area of the detection mechanism, the virtual object corresponding to that real world object is removed from the virtual world environment. Thus, virtual objects corresponding to real world objects are only able to be utilized in the virtual world environment while their corresponding real world objects are detected within a detection area of a detection mechanism.
- The illustrative embodiments described above may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment containing both hardware and software elements. In a preferred embodiment, the invention is implemented in software, which includes but is not limited to firmware, resident software, microcode, etc.
- Furthermore, the illustrative embodiments may take the form of a computer program product accessible from a computer-usable or computer-readable medium providing program code for use by or in connection with a computer or any instruction execution system. For the purposes of this description, a computer-usable or computer readable medium can be any apparatus that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
- The medium may be an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system (or apparatus or device) or a propagation medium. Examples of a computer-readable medium include a semiconductor or solid state memory, magnetic tape, a removable computer diskette, a random access memory (RAM), a read-only memory (ROM), a rigid magnetic disk and an optical disk. Current examples of optical disks include compact disk—read only memory (CD-ROM), compact disk—read/write (CD-R/W) and DVD.
- As previously described above with regard to
FIGS. 1-3 , a data processing system suitable for storing and/or executing program code will include at least one processor coupled directly or indirectly to memory elements through a system bus. The memory elements can include local memory employed during actual execution of the program code, bulk storage, and cache memories which provide temporary storage of at least some program code in order to reduce the number of times code must be retrieved from bulk storage during execution. - Input/output or I/O devices (including but not limited to keyboards, displays, pointing devices, etc.) may be coupled to the system either directly or through intervening I/O controllers. Network adapters may also be coupled to the system to enable the data processing system to become coupled to other data processing systems or remote printers or storage devices through intervening private or public networks. Modems, cable modem and Ethernet cards are just a few of the currently available types of network adapters.
- The description of the present invention has been presented for purposes of illustration and description, and is not intended to be exhaustive or limited to the invention in the form disclosed. Many modifications and variations will be apparent to those of ordinary skill in the art. The embodiment was chosen and described in order to best explain the principles of the invention, the practical application, and to enable others of ordinary skill in the art to understand the invention for various embodiments with various modifications as are suited to the particular use contemplated.
Claims (35)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/372,327 US7843471B2 (en) | 2006-03-09 | 2006-03-09 | Persistent authenticating mechanism to map real world object presence into virtual world object awareness |
PCT/EP2007/051617 WO2007101785A1 (en) | 2006-03-09 | 2007-02-20 | Persistent authenticating system and method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/372,327 US7843471B2 (en) | 2006-03-09 | 2006-03-09 | Persistent authenticating mechanism to map real world object presence into virtual world object awareness |
Publications (2)
Publication Number | Publication Date |
---|---|
US20070211047A1 true US20070211047A1 (en) | 2007-09-13 |
US7843471B2 US7843471B2 (en) | 2010-11-30 |
Family
ID=37847809
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US11/372,327 Active 2028-11-13 US7843471B2 (en) | 2006-03-09 | 2006-03-09 | Persistent authenticating mechanism to map real world object presence into virtual world object awareness |
Country Status (2)
Country | Link |
---|---|
US (1) | US7843471B2 (en) |
WO (1) | WO2007101785A1 (en) |
Cited By (73)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090100352A1 (en) * | 2007-10-15 | 2009-04-16 | Yunwu Huang | Method and apparatus for bridging real-world web applications and 3d virtual worlds |
US20090164916A1 (en) * | 2007-12-21 | 2009-06-25 | Samsung Electronics Co., Ltd. | Method and system for creating mixed world that reflects real state |
JP2009279347A (en) * | 2008-05-26 | 2009-12-03 | Taito Corp | Game apparatus, and pen inputting device |
US20100005007A1 (en) * | 2008-07-07 | 2010-01-07 | Aaron Roger Cox | Methods of associating real world items with virtual world representations |
US20100011312A1 (en) * | 2008-07-11 | 2010-01-14 | International Business Machines Corporation | Rfid reader integration to virtual world monitoring |
US20100045660A1 (en) * | 2008-08-21 | 2010-02-25 | Dettinger Richard D | Recommendations for Object Use in Virtual Environment |
US20100145769A1 (en) * | 2008-12-09 | 2010-06-10 | International Business Machines Corporation | System and method for product inquiry and incentive for virtual universes |
US20100162136A1 (en) * | 2008-12-19 | 2010-06-24 | International Business Machines Corporation | Degrading avatar appearances in a virtual universe |
US20100192110A1 (en) * | 2009-01-23 | 2010-07-29 | International Business Machines Corporation | Method for making a 3-dimensional virtual world accessible for the blind |
WO2010093995A1 (en) * | 2009-02-13 | 2010-08-19 | Social Gaming Network | Apparatuses, methods and systems for an interworld feedback platform bridge |
US20100220097A1 (en) * | 2009-02-28 | 2010-09-02 | International Business Machines Corporation | Altering avatar appearances based on avatar population in a virtual universe |
US20110055927A1 (en) * | 2009-08-27 | 2011-03-03 | International Business Machines Corporation | Updating assets rendered in a virtual world environment based on detected user interactions in another world |
US20110098092A1 (en) * | 2009-10-27 | 2011-04-28 | Reiche Iii Paul | Video game with representative physical object related content |
US20110165939A1 (en) * | 2010-01-05 | 2011-07-07 | Ganz | Method and system for providing a 3d activity in a virtual presentation |
US20110231433A1 (en) * | 2010-03-17 | 2011-09-22 | Ricoh Company, Limited | Management system, management method, and temporary storage document server |
US20120016761A1 (en) * | 2010-07-15 | 2012-01-19 | Enyama, Inc. | Techniques For Provisioning Content |
US20120038667A1 (en) * | 2010-08-11 | 2012-02-16 | International Business Machines Corporation | Replicating Changes Between Corresponding Objects |
US8128500B1 (en) * | 2007-07-13 | 2012-03-06 | Ganz | System and method for generating a virtual environment for land-based and underwater virtual characters |
US20120254779A1 (en) * | 2011-04-01 | 2012-10-04 | Arthur Austin Ollivierre | System and method for displaying objects in a user interface based on a visual acuity of a viewer |
US20120295699A1 (en) * | 2011-05-17 | 2012-11-22 | Paul Reiche | Conditional access to areas in a video game |
US20120295700A1 (en) * | 2011-05-17 | 2012-11-22 | Paul Reiche | Conditional access to areas in a video game |
US8495078B2 (en) * | 2009-12-02 | 2013-07-23 | International Business Machines Corporation | System and method for abstraction of objects for cross virtual universe deployment |
US20130288563A1 (en) * | 2005-10-21 | 2013-10-31 | Patent Category Corporation | Interactive toy system |
US20130296058A1 (en) * | 2012-05-01 | 2013-11-07 | Activision Publishing, Inc. | Server based interactive video game with toys |
US20130346239A1 (en) * | 2007-08-23 | 2013-12-26 | Ebay Inc. | Methods and systems to facilitate a purchase of an item on a network-based marketplace |
US20140100029A1 (en) * | 2012-10-10 | 2014-04-10 | Paul Reiche | Interactive video game with different sized toys having different abilities within the video game |
US8726172B1 (en) | 2012-12-26 | 2014-05-13 | Disney Enterprises, Inc. | Managing an environment of a virtual space based on characters made accessible responsive to corresponding tokens being detected |
US20140181820A1 (en) * | 2012-12-26 | 2014-06-26 | Disney Enterprises, Inc. | Facilitating customization of a virtual space based on accessible virtual items |
US20140256430A1 (en) * | 2013-03-08 | 2014-09-11 | Nintendo Co., Ltd. | Information processing apparatus, information processing system, information processing method and recording medium |
US20140253591A1 (en) * | 2013-03-05 | 2014-09-11 | Nintendo Co., Ltd. | Information processing system, information processing apparatus, information processing method, and computer-readable recording medium recording information processing program |
US8836719B2 (en) | 2010-04-23 | 2014-09-16 | Ganz | Crafting system in a virtual environment |
US20140274313A1 (en) * | 2013-03-15 | 2014-09-18 | Karthik Bala | System and method for purchasing physical toys and corresponding virtual toys |
US20140273717A1 (en) * | 2013-03-13 | 2014-09-18 | Hasbro, Inc. | Three way multidirectional interactive toy |
US8904298B2 (en) | 2013-03-15 | 2014-12-02 | Disney Enterprises, Inc. | Effectuating modifications within an instance of a virtual space presented via multiple disparate client computing platforms responsive to detection of a token associated with a single client computing platform |
US8909920B2 (en) | 2012-12-26 | 2014-12-09 | Disney Enterprises, Inc. | Linking token detection at a single computing platform with a user identification to effectuate modifications in virtual space instances presented via multiple computing platforms |
US8910266B2 (en) | 2012-12-26 | 2014-12-09 | Disney Enterprises, Inc. | Linking token detection at a single computing platform with a user identification to unlock content in virtual space instances presented via multiple computing platforms |
US20150009190A1 (en) * | 2013-07-08 | 2015-01-08 | Masato Kuwahara | Display device, storage medium, display method and display system |
CN104281787A (en) * | 2013-07-01 | 2015-01-14 | 迪士尼企业公司 | Implementing encrypted content in a game |
US8972369B2 (en) | 2012-12-26 | 2015-03-03 | Disney Enterprises, Inc. | Providing a common virtual item repository in a virtual space |
US20150178516A1 (en) * | 2013-12-23 | 2015-06-25 | Dropbox, Inc. | Restricting access to content |
US20150187108A1 (en) * | 2013-12-31 | 2015-07-02 | Daqri, Llc | Augmented reality content adapted to changes in real world space geometry |
US20150258435A1 (en) * | 2014-03-11 | 2015-09-17 | Microsoft Corporation | Modular construction for interacting with software |
US20150375134A1 (en) * | 2014-06-30 | 2015-12-31 | Microsoft Corporation | Interactive play sets |
USD748200S1 (en) | 2013-01-15 | 2016-01-26 | Disney Enterprises, Inc. | Power disk reader |
USD748199S1 (en) | 2013-01-15 | 2016-01-26 | Disney Enterprises, Inc. | Multi-sided power disk |
US9327200B2 (en) | 2012-12-26 | 2016-05-03 | Disney Enterprises, Inc. | Managing a theme of a virtual space based on characters made accessible responsive to corresponding tokens being detected |
US20160171770A1 (en) * | 2014-12-10 | 2016-06-16 | Sixense Entertainment, Inc. | System and Method for Assisting a User in Locating Physical Objects While the User is in a Virtual Reality Environment |
US9381430B2 (en) * | 2011-05-17 | 2016-07-05 | Activision Publishing, Inc. | Interactive video game using game-related physical objects for conducting gameplay |
US9387407B2 (en) | 2012-12-26 | 2016-07-12 | Disney Enterprises, Inc. | Managing objectives associated with a virtual space based on characters made accessible responsive to corresponding tokens being detected |
US9457263B2 (en) | 2012-12-26 | 2016-10-04 | Disney Enterprises, Inc. | Unlocking virtual items in a virtual space responsive to physical token detection |
US9517404B2 (en) | 2012-12-26 | 2016-12-13 | Disney Enterprises, Inc. | Apparatus, system, and method for effectuating modifications to a virtual space responsive to token detection |
US9776071B2 (en) | 2013-05-09 | 2017-10-03 | Mattel, Inc. | Resonant coils for use with games and toys |
KR20170134644A (en) * | 2015-04-08 | 2017-12-06 | 레고 에이/에스 | Game system |
US10089253B2 (en) | 2014-03-11 | 2018-10-02 | Microsoft Technology Licensing, Llc | Data store for a modular assembly system |
US10150043B2 (en) | 2014-03-11 | 2018-12-11 | Microsoft Technology Licensing, Llc | Interactive smart beads |
US10157487B2 (en) * | 2015-07-30 | 2018-12-18 | International Business Machines Corporation | VR biometric integration |
US10159894B2 (en) | 2014-03-11 | 2018-12-25 | Microsoft Technology Licensing, Llc | Gaming system for modular toys |
US10315119B2 (en) * | 2011-05-17 | 2019-06-11 | Activision Publishing, Inc. | Video game with concurrent processing of game-related physical objects |
US10339613B2 (en) | 2007-08-23 | 2019-07-02 | Ebay Inc. | Viewing shopping information on a network based social platform |
WO2019183485A1 (en) * | 2018-03-22 | 2019-09-26 | Infinite Kingdoms Llc | Connected avatar technology |
US10445437B2 (en) | 2014-03-11 | 2019-10-15 | Microsoft Technology Licensing, Llc | Generation of custom modular objects |
US20200108311A1 (en) * | 2018-10-09 | 2020-04-09 | Ivan Mauricio Buritica Suarez | System and method for emulating an electronic game to play in a physical gaming environment |
US10984126B2 (en) | 2007-08-23 | 2021-04-20 | Ebay Inc. | Sharing information on a network-based social platform |
CN113476857A (en) * | 2020-08-27 | 2021-10-08 | 万代股份有限公司 | Game fighting system, program, terminal and connection device |
CN113521756A (en) * | 2020-08-27 | 2021-10-22 | 万代股份有限公司 | Game support system, program, information communication terminal, and connection device |
US11173401B1 (en) * | 2020-01-21 | 2021-11-16 | Mythical, Inc. | Systems and methods for adjusting previously-made modifications of game assets based on changes in the ownership of tradeable items |
US11224814B2 (en) * | 2018-01-30 | 2022-01-18 | Netease (Hangzhou) Network Co., Ltd. | Processing method and apparatus for virtual resource in game scene |
US11247127B1 (en) | 2020-01-21 | 2022-02-15 | Mythical, Inc. | Systems and methods for modifying game assets of an online gaming platform based on tradeable items |
US11369877B2 (en) * | 2019-12-24 | 2022-06-28 | Nintendo Co., Ltd. | Computer-readable non-transitory storage medium having game program stored therein, game apparatus, game system, and game processing method for providing user interface showing change in parameter corresponding to an additional item |
US11376508B1 (en) | 2020-04-03 | 2022-07-05 | Mythical, Inc. | Systems and methods for enabling exchange of tradeable items |
CN115272630A (en) * | 2022-09-29 | 2022-11-01 | 南方科技大学 | Data processing method, device, virtual reality glasses and storage medium |
US11517809B2 (en) * | 2017-06-12 | 2022-12-06 | Bandai Co., Ltd. | Game device, program product, game system and article for games |
US20230226448A1 (en) * | 2019-12-09 | 2023-07-20 | Jeremy Noah Anderson | Game player created object and video game play |
Families Citing this family (143)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
NZ564006A (en) | 2006-12-06 | 2009-03-31 | 2121200 Ontario Inc | System and method for product marketing using feature codes |
FR2917931A1 (en) * | 2007-06-22 | 2008-12-26 | France Telecom | METHOD AND SYSTEM FOR CONNECTING PEOPLE IN A TELECOMMUNICATIONS SYSTEM. |
EP2051208A3 (en) * | 2007-10-17 | 2017-05-24 | Sony Interactive Entertainment Inc. | Generating an asset for interactive entertainment using digital image capture |
US8127235B2 (en) | 2007-11-30 | 2012-02-28 | International Business Machines Corporation | Automatic increasing of capacity of a virtual space in a virtual world |
US20090164919A1 (en) | 2007-12-24 | 2009-06-25 | Cary Lee Bates | Generating data for managing encounters in a virtual world environment |
JP5159375B2 (en) | 2008-03-07 | 2013-03-06 | インターナショナル・ビジネス・マシーンズ・コーポレーション | Object authenticity determination system and method in metaverse, and computer program thereof |
US10155156B2 (en) | 2008-06-03 | 2018-12-18 | Tweedletech, Llc | Multi-dimensional game comprising interactive physical and virtual components |
US9649551B2 (en) | 2008-06-03 | 2017-05-16 | Tweedletech, Llc | Furniture and building structures comprising sensors for determining the position of one or more objects |
EP2328662A4 (en) | 2008-06-03 | 2013-05-29 | Tweedletech Llc | An intelligent game system for putting intelligence into board and tabletop games including miniatures |
US8602857B2 (en) | 2008-06-03 | 2013-12-10 | Tweedletech, Llc | Intelligent board game system with visual marker based game object tracking and identification |
US8974295B2 (en) | 2008-06-03 | 2015-03-10 | Tweedletech, Llc | Intelligent game system including intelligent foldable three-dimensional terrain |
US8161397B2 (en) * | 2008-06-05 | 2012-04-17 | Samsung Electronics Co., Ltd. | Interaction between real-world digital environments and virtual worlds |
US20090315257A1 (en) * | 2008-06-24 | 2009-12-24 | Three Of Hearts, Llc | Musical playing card board game |
US20100030804A1 (en) * | 2008-07-31 | 2010-02-04 | International Business Machines Corporation | Synchronization of Locations in Real and Virtual Worlds |
US8472670B2 (en) * | 2008-08-08 | 2013-06-25 | Panasonic Corporation | Target detection device and target detection method |
US8255807B2 (en) | 2008-12-23 | 2012-08-28 | Ganz | Item customization and website customization |
CA2772135C (en) | 2009-08-31 | 2013-12-31 | Ganz | System and method for limiting the number of characters displayed in a common area |
US8738533B2 (en) * | 2009-10-29 | 2014-05-27 | Disney Enterprises, Inc. | Real-world items unlocking virtual items online and in video games |
US9205328B2 (en) | 2010-02-18 | 2015-12-08 | Activision Publishing, Inc. | Videogame system and method that enables characters to earn virtual fans by completing secondary objectives |
WO2011119552A2 (en) | 2010-03-22 | 2011-09-29 | Mattel, Inc. | Electronic device and the input and output of data |
US9682324B2 (en) | 2010-05-12 | 2017-06-20 | Activision Publishing, Inc. | System and method for enabling players to participate in asynchronous, competitive challenges |
JP5993856B2 (en) | 2010-09-09 | 2016-09-14 | トウィードルテック リミテッド ライアビリティ カンパニー | Board game with dynamic feature tracking |
WO2012105836A2 (en) * | 2011-01-31 | 2012-08-09 | Studio Sophisti V.O.F. | Game board |
US8937620B1 (en) | 2011-04-07 | 2015-01-20 | Google Inc. | System and methods for generation and control of story animation |
US10238977B2 (en) | 2011-05-17 | 2019-03-26 | Activision Publishing, Inc. | Collection of marketing information developed during video game play |
USD662949S1 (en) | 2011-05-17 | 2012-07-03 | Joby-Rome Otero | Video game peripheral detection device |
US10796494B2 (en) | 2011-06-06 | 2020-10-06 | Microsoft Technology Licensing, Llc | Adding attributes to virtual representations of real-world objects |
US8773467B2 (en) * | 2011-06-13 | 2014-07-08 | International Business Machines Corporation | Enhanced asset management and planning system |
AU2011205223C1 (en) | 2011-08-09 | 2013-03-28 | Microsoft Technology Licensing, Llc | Physical interaction with virtual objects for DRM |
US9153195B2 (en) | 2011-08-17 | 2015-10-06 | Microsoft Technology Licensing, Llc | Providing contextual personal information by a mixed reality device |
WO2013028908A1 (en) | 2011-08-24 | 2013-02-28 | Microsoft Corporation | Touch and social cues as inputs into a computer |
US8894462B2 (en) | 2011-12-22 | 2014-11-25 | Activision Publishing, Inc. | Interactive video game with visual lighting effects |
US8737035B2 (en) * | 2012-05-18 | 2014-05-27 | Disney Enterprises, Inc. | Magnetically movable objects over a display of an electronic device |
US9539498B1 (en) * | 2012-07-31 | 2017-01-10 | Niantic, Inc. | Mapping real world actions to a virtual world associated with a location-based game |
US8858339B2 (en) | 2012-12-11 | 2014-10-14 | Activision Publishing, Inc. | Interactive video game system comprising toys with rewritable memories |
EP2839868A3 (en) * | 2012-12-26 | 2015-11-25 | Disney Enterprises, Inc. | Providing a common virtual item repository in a virtual space; unlocking virtual items in a virtual space responsive to physical token detection; and facilitating customization of a virtual space based on accessible virtual items |
EP2749327A1 (en) * | 2012-12-26 | 2014-07-02 | Disney Enterprises, Inc. | Managing objectives, an environment, and/or a theme associated with a virtual space based on characters made accessible responsive to corresponding tokens being detected |
EP2749328A1 (en) * | 2012-12-26 | 2014-07-02 | Disney Enterprises, Inc. | Apparatus, system, and method for effectuating modifications to characters and/or virtual items within a virtual space, and/or to the virtual space itself, responsive to token detection |
US10137376B2 (en) | 2012-12-31 | 2018-11-27 | Activision Publishing, Inc. | System and method for creating and streaming augmented game sessions |
FR3000818B1 (en) * | 2013-01-04 | 2016-03-04 | Epawn | CONTROL OF A REMOTE DEVICE BY A HOST DEVICE FOR MOBILE DEVICES |
WO2014143776A2 (en) | 2013-03-15 | 2014-09-18 | Bodhi Technology Ventures Llc | Providing remote interactions with host device using a wireless device |
US9011194B2 (en) * | 2013-03-15 | 2015-04-21 | Disney Enterprises, Inc. | Managing virtual content based on information associated with toy objects |
US9610500B2 (en) | 2013-03-15 | 2017-04-04 | Disney Enterprise, Inc. | Managing virtual content based on information associated with toy objects |
EP2777786A3 (en) * | 2013-03-15 | 2014-12-10 | Disney Enterprises, Inc. | Managing virtual content based on information associated with toy objects |
GB2513845B (en) * | 2013-05-02 | 2015-06-10 | Sony Comp Entertainment Europe | Identification apparatus and method |
US9214987B2 (en) * | 2014-05-18 | 2015-12-15 | Auden Techno Corp. | Near field antenna for object detecting device |
US9764228B2 (en) * | 2014-05-21 | 2017-09-19 | Activision Publishing, Inc. | Contextual play pattern switching system and method |
US9483763B2 (en) | 2014-05-29 | 2016-11-01 | Apple Inc. | User interface for payments |
US9967401B2 (en) | 2014-05-30 | 2018-05-08 | Apple Inc. | User interface for phone call routing among devices |
WO2015185629A2 (en) | 2014-06-06 | 2015-12-10 | Lego A/S | Interactive game apparatus and toy construction system |
US20160175698A1 (en) * | 2014-06-13 | 2016-06-23 | Zheng Shi | System and method for directing a targeted object on an interactive surface to produce a response |
US10286326B2 (en) | 2014-07-03 | 2019-05-14 | Activision Publishing, Inc. | Soft reservation system and method for multiplayer video games |
EP3195098B1 (en) | 2014-07-21 | 2024-10-23 | Apple Inc. | Remote user interface |
US10339293B2 (en) | 2014-08-15 | 2019-07-02 | Apple Inc. | Authenticated device used to unlock another device |
CN106605196B (en) | 2014-09-02 | 2018-11-09 | 苹果公司 | remote camera user interface |
WO2016036552A1 (en) | 2014-09-02 | 2016-03-10 | Apple Inc. | User interactions for a mapping application |
WO2016036603A1 (en) * | 2014-09-02 | 2016-03-10 | Apple Inc. | Reduced size configuration interface |
EP3200886B1 (en) | 2014-10-02 | 2023-09-20 | Lego A/S | Game system |
US11351466B2 (en) | 2014-12-05 | 2022-06-07 | Activision Publishing, Ing. | System and method for customizing a replay of one or more game events in a video game |
US10118099B2 (en) | 2014-12-16 | 2018-11-06 | Activision Publishing, Inc. | System and method for transparently styling non-player characters in a multiplayer video game |
US9901827B2 (en) | 2015-01-06 | 2018-02-27 | Spin Master Ltd. | Methods and system relating to physical constructions and virtual representations |
US9574896B2 (en) | 2015-02-13 | 2017-02-21 | Apple Inc. | Navigation user interface |
US9474964B2 (en) | 2015-02-13 | 2016-10-25 | Jumo, Inc. | System and method for providing state information of an action figure |
US9259651B1 (en) | 2015-02-13 | 2016-02-16 | Jumo, Inc. | System and method for providing relevant notifications via an action figure |
US9266027B1 (en) | 2015-02-13 | 2016-02-23 | Jumo, Inc. | System and method for providing an enhanced marketing, sale, or order fulfillment experience related to action figures or action figure accessories having corresponding virtual counterparts |
US9833695B2 (en) | 2015-02-13 | 2017-12-05 | Jumo, Inc. | System and method for presenting a virtual counterpart of an action figure based on action figure state information |
US9205336B1 (en) | 2015-03-02 | 2015-12-08 | Jumo, Inc. | System and method for providing secured wireless communication with an action figure or action figure accessory |
US10216351B2 (en) | 2015-03-08 | 2019-02-26 | Apple Inc. | Device configuration user interface |
US20160284245A1 (en) * | 2015-03-26 | 2016-09-29 | Martin Suarez | Apparatus and method to motivate performance |
DE112016001653T5 (en) | 2015-04-09 | 2017-12-21 | Warner Bros. Entertainment Inc. | PORTAL DEVICE AND CO-OPERATING VIDEO GAME |
US10486068B2 (en) | 2015-05-14 | 2019-11-26 | Activision Publishing, Inc. | System and method for providing dynamically variable maps in a video game |
US10315113B2 (en) | 2015-05-14 | 2019-06-11 | Activision Publishing, Inc. | System and method for simulating gameplay of nonplayer characters distributed across networked end user devices |
US10286314B2 (en) | 2015-05-14 | 2019-05-14 | Activision Publishing, Inc. | System and method for providing continuous gameplay in a multiplayer video game through an unbounded gameplay session |
US20160358133A1 (en) | 2015-06-05 | 2016-12-08 | Apple Inc. | User interface for loyalty accounts and private label accounts for a wearable device |
US10086266B2 (en) | 2015-06-15 | 2018-10-02 | Activision Publishing, Inc. | System and method for uniquely identifying physical trading cards and incorporating trading card game items in a video game |
US10471348B2 (en) | 2015-07-24 | 2019-11-12 | Activision Publishing, Inc. | System and method for creating and sharing customized video game weapon configurations in multiplayer video games via one or more social networks |
US10099140B2 (en) | 2015-10-08 | 2018-10-16 | Activision Publishing, Inc. | System and method for generating personalized messaging campaigns for video game players |
US11185784B2 (en) | 2015-10-08 | 2021-11-30 | Activision Publishing, Inc. | System and method for generating personalized messaging campaigns for video game players |
US10245509B2 (en) | 2015-10-21 | 2019-04-02 | Activision Publishing, Inc. | System and method of inferring user interest in different aspects of video game streams |
US10376781B2 (en) | 2015-10-21 | 2019-08-13 | Activision Publishing, Inc. | System and method of generating and distributing video game streams |
US10232272B2 (en) | 2015-10-21 | 2019-03-19 | Activision Publishing, Inc. | System and method for replaying video game streams |
US10694352B2 (en) | 2015-10-28 | 2020-06-23 | Activision Publishing, Inc. | System and method of using physical objects to control software access |
US10603580B2 (en) | 2015-12-09 | 2020-03-31 | Activision Publishing, Inc. | Videogame portal game play |
US10969748B1 (en) | 2015-12-28 | 2021-04-06 | Disney Enterprises, Inc. | Systems and methods for using a vehicle as a motion base for a simulated experience |
US10376776B2 (en) | 2016-01-20 | 2019-08-13 | Disney Enterprises, Inc. | Systems and methods for providing a video game using information stored by physical objects |
US11524242B2 (en) | 2016-01-20 | 2022-12-13 | Disney Enterprises, Inc. | Systems and methods for providing customized instances of a game within a virtual space |
US10226703B2 (en) | 2016-04-01 | 2019-03-12 | Activision Publishing, Inc. | System and method of generating and providing interactive annotation items based on triggering events in a video game |
KR102474088B1 (en) | 2016-04-22 | 2022-12-05 | 인터디지털 씨이 페이튼트 홀딩스, 에스에이에스 | Method and device for compositing an image |
US10226701B2 (en) | 2016-04-29 | 2019-03-12 | Activision Publishing, Inc. | System and method for identifying spawn locations in a video game |
US10179289B2 (en) | 2016-06-21 | 2019-01-15 | Activision Publishing, Inc. | System and method for reading graphically-encoded identifiers from physical trading cards through image-based template matching |
US10143918B2 (en) * | 2016-07-22 | 2018-12-04 | Play Fusion Limited | Apparatus, system and method for enhancing a gaming experience |
US10573065B2 (en) | 2016-07-29 | 2020-02-25 | Activision Publishing, Inc. | Systems and methods for automating the personalization of blendshape rigs based on performance capture data |
US10709981B2 (en) | 2016-11-17 | 2020-07-14 | Activision Publishing, Inc. | Systems and methods for the real-time generation of in-game, locally accessible barrier-aware heatmaps |
US10463964B2 (en) | 2016-11-17 | 2019-11-05 | Activision Publishing, Inc. | Systems and methods for the real-time generation of in-game, locally accessible heatmaps |
US10500498B2 (en) | 2016-11-29 | 2019-12-10 | Activision Publishing, Inc. | System and method for optimizing virtual games |
US10055880B2 (en) | 2016-12-06 | 2018-08-21 | Activision Publishing, Inc. | Methods and systems to modify a two dimensional facial image to increase dimensional depth and generate a facial image that appears three dimensional |
US10861079B2 (en) | 2017-02-23 | 2020-12-08 | Activision Publishing, Inc. | Flexible online pre-ordering system for media |
US10818060B2 (en) | 2017-09-05 | 2020-10-27 | Activision Publishing, Inc. | Systems and methods for guiding motion capture actors using a motion reference system |
US10974150B2 (en) | 2017-09-27 | 2021-04-13 | Activision Publishing, Inc. | Methods and systems for improved content customization in multiplayer gaming environments |
US10561945B2 (en) | 2017-09-27 | 2020-02-18 | Activision Publishing, Inc. | Methods and systems for incentivizing team cooperation in multiplayer gaming environments |
US11040286B2 (en) | 2017-09-27 | 2021-06-22 | Activision Publishing, Inc. | Methods and systems for improved content generation in multiplayer gaming environments |
US10463971B2 (en) | 2017-12-06 | 2019-11-05 | Activision Publishing, Inc. | System and method for validating video gaming data |
US10537809B2 (en) | 2017-12-06 | 2020-01-21 | Activision Publishing, Inc. | System and method for validating video gaming data |
US10981051B2 (en) | 2017-12-19 | 2021-04-20 | Activision Publishing, Inc. | Synchronized, fully programmable game controllers |
US10864443B2 (en) | 2017-12-22 | 2020-12-15 | Activision Publishing, Inc. | Video game content aggregation, normalization, and publication systems and methods |
US11278813B2 (en) | 2017-12-22 | 2022-03-22 | Activision Publishing, Inc. | Systems and methods for enabling audience participation in bonus game play sessions |
US10596471B2 (en) | 2017-12-22 | 2020-03-24 | Activision Publishing, Inc. | Systems and methods for enabling audience participation in multi-player video game play sessions |
US10970560B2 (en) | 2018-01-12 | 2021-04-06 | Disney Enterprises, Inc. | Systems and methods to trigger presentation of in-vehicle content |
US10887193B2 (en) | 2018-06-03 | 2021-01-05 | Apple Inc. | User interfaces for updating network connection settings of external devices |
US10841632B2 (en) | 2018-08-08 | 2020-11-17 | Disney Enterprises, Inc. | Sequential multiplayer storytelling in connected vehicles |
US11192028B2 (en) | 2018-11-19 | 2021-12-07 | Activision Publishing, Inc. | Systems and methods for the real-time customization of video game content based on player data |
US11263670B2 (en) | 2018-11-19 | 2022-03-01 | Activision Publishing, Inc. | Systems and methods for dynamically modifying video game content based on non-video gaming content being concurrently experienced by a user |
US11115712B2 (en) | 2018-12-15 | 2021-09-07 | Activision Publishing, Inc. | Systems and methods for indexing, searching for, and retrieving digital media |
US11679330B2 (en) | 2018-12-18 | 2023-06-20 | Activision Publishing, Inc. | Systems and methods for generating improved non-player characters |
US11305191B2 (en) | 2018-12-20 | 2022-04-19 | Activision Publishing, Inc. | Systems and methods for controlling camera perspectives, movements, and displays of video game gameplay |
KR102354046B1 (en) | 2019-05-06 | 2022-01-25 | 애플 인크. | Limited operation of electronic devices |
DK201970533A1 (en) | 2019-05-31 | 2021-02-15 | Apple Inc | Methods and user interfaces for sharing audio |
US12033271B2 (en) * | 2019-06-18 | 2024-07-09 | The Calany Holding S. À R.L. | 3D structure engine-based computation platform |
US12039354B2 (en) | 2019-06-18 | 2024-07-16 | The Calany Holding S. À R.L. | System and method to operate 3D applications through positional virtualization technology |
US12040993B2 (en) | 2019-06-18 | 2024-07-16 | The Calany Holding S. À R.L. | Software engine virtualization and dynamic resource and task distribution across edge and cloud |
US11344808B2 (en) | 2019-06-28 | 2022-05-31 | Activision Publishing, Inc. | Systems and methods for dynamically generating and modulating music based on gaming events, player profiles and/or player reactions |
US10785621B1 (en) | 2019-07-30 | 2020-09-22 | Disney Enterprises, Inc. | Systems and methods to provide an interactive space based on vehicle-to-vehicle communications |
US11097193B2 (en) | 2019-09-11 | 2021-08-24 | Activision Publishing, Inc. | Methods and systems for increasing player engagement in multiplayer gaming environments |
US11423605B2 (en) | 2019-11-01 | 2022-08-23 | Activision Publishing, Inc. | Systems and methods for remastering a game space while maintaining the underlying game simulation |
US11712627B2 (en) | 2019-11-08 | 2023-08-01 | Activision Publishing, Inc. | System and method for providing conditional access to virtual gaming items |
US11537209B2 (en) | 2019-12-17 | 2022-12-27 | Activision Publishing, Inc. | Systems and methods for guiding actors using a motion capture reference system |
US11420122B2 (en) | 2019-12-23 | 2022-08-23 | Activision Publishing, Inc. | Systems and methods for controlling camera perspectives, movements, and displays of video game gameplay |
US11563774B2 (en) | 2019-12-27 | 2023-01-24 | Activision Publishing, Inc. | Systems and methods for tracking and identifying phishing website authors |
US11076276B1 (en) | 2020-03-13 | 2021-07-27 | Disney Enterprises, Inc. | Systems and methods to provide wireless communication between computing platforms and articles |
US11524234B2 (en) | 2020-08-18 | 2022-12-13 | Activision Publishing, Inc. | Multiplayer video games with virtual characters having dynamically modified fields of view |
US11351459B2 (en) | 2020-08-18 | 2022-06-07 | Activision Publishing, Inc. | Multiplayer video games with virtual characters having dynamically generated attribute profiles unconstrained by predefined discrete values |
US11724188B2 (en) | 2020-09-29 | 2023-08-15 | Activision Publishing, Inc. | Methods and systems for selecting a level of detail visual asset during the execution of a video game |
US11717753B2 (en) | 2020-09-29 | 2023-08-08 | Activision Publishing, Inc. | Methods and systems for generating modified level of detail visual assets in a video game |
US11833423B2 (en) | 2020-09-29 | 2023-12-05 | Activision Publishing, Inc. | Methods and systems for generating level of detail visual assets in a video game |
US11439904B2 (en) | 2020-11-11 | 2022-09-13 | Activision Publishing, Inc. | Systems and methods for imparting dynamic and realistic movement to player-controlled avatars in video games |
US12179113B2 (en) | 2020-12-09 | 2024-12-31 | Activision Publishing, Inc. | Methods and systems for dynamically selecting and modifying digital assets within game elements of a multiplayer video game |
US12097430B2 (en) | 2020-12-28 | 2024-09-24 | Activision Publishing, Inc. | Methods and systems for generating and managing active objects in video games |
US11853439B2 (en) | 2020-12-30 | 2023-12-26 | Activision Publishing, Inc. | Distributed data storage system providing enhanced security |
US11794107B2 (en) | 2020-12-30 | 2023-10-24 | Activision Publishing, Inc. | Systems and methods for improved collision detection in video games |
US12064688B2 (en) | 2020-12-30 | 2024-08-20 | Activision Publishing, Inc. | Methods and systems for determining decal projections intersecting spatial units in a frame of a game space |
US12211031B2 (en) | 2021-08-11 | 2025-01-28 | Disney Enterprises, Inc. | Systems and methods to compilate an experience summary based on real-world experiences |
US11756447B1 (en) | 2021-09-13 | 2023-09-12 | Christine Hoffman | System for teaching music notation and associated method of use |
US12141791B2 (en) | 2022-11-07 | 2024-11-12 | Disney Enterprises, Inc. | Systems and methods to adjust a unit of experience based on digital assets of users |
Citations (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5853327A (en) * | 1994-07-28 | 1998-12-29 | Super Dimension, Inc. | Computerized game board |
US5953686A (en) * | 1995-08-03 | 1999-09-14 | Interval Research Corporation | Video camera based computer input system with interchangeable physical interface |
US6165068A (en) * | 1997-01-22 | 2000-12-26 | Tomy Company, Ltd. | Connection-fighting type game machine and connection-fighting type game methods |
US6167353A (en) * | 1996-07-03 | 2000-12-26 | Interval Research Corporation | Computer method and apparatus for interacting with a physical system |
US6190174B1 (en) * | 1999-06-03 | 2001-02-20 | Kader Industrial Company Limited | Electronic story board |
US20020090985A1 (en) * | 2000-09-07 | 2002-07-11 | Ilan Tochner | Coexistent interaction between a virtual character and the real world |
US6460851B1 (en) * | 1996-05-10 | 2002-10-08 | Dennis H. Lee | Computer interface apparatus for linking games to personal computers |
US20020185532A1 (en) * | 2001-06-07 | 2002-12-12 | Berquist David T. | RFID data collection and use |
US20030134679A1 (en) * | 2001-11-19 | 2003-07-17 | Radica China Ltd. | Electronic gaming device using coded input data |
US20030171984A1 (en) * | 2002-03-06 | 2003-09-11 | Wodka Joseph F. | Customization of promotional material through use of programmable radio frequency identification technology |
US6690156B1 (en) * | 2000-07-28 | 2004-02-10 | N-Trig Ltd. | Physical object location apparatus and method and a graphic display device using the same |
US20040111335A1 (en) * | 2002-12-04 | 2004-06-10 | Black Charles Ronald | RFID space monitoring and asset inventory system |
US20040127277A1 (en) * | 2002-10-09 | 2004-07-01 | Walker Jay S. | Method and apparatus for authenticating data relating to usage of a gaming device |
US6761634B1 (en) * | 2001-06-07 | 2004-07-13 | Hasbro, Inc. | Arcade table |
US20040145474A1 (en) * | 2002-11-15 | 2004-07-29 | Schmidtberg Rupert A. | RF identification reader for communicating condition information associated with the reader |
US6835131B1 (en) * | 1998-11-26 | 2004-12-28 | Innovision Research & Technology Plc | Game apparatus for supplying power to playing elements |
US20050069171A1 (en) * | 1999-05-19 | 2005-03-31 | Rhoads Geoffrey B. | Embedded data in gaming objects for authentication and association of behavior information |
US20050099268A1 (en) * | 2003-11-12 | 2005-05-12 | Ari Juels | Radio frequency identification system with privacy policy implementation based on device classification |
US20050174236A1 (en) * | 2004-01-29 | 2005-08-11 | Brookner George M. | RFID device tracking and information gathering |
US20050199723A1 (en) * | 2001-11-30 | 2005-09-15 | Allen Lubow | Method for identifying and authenticating goods using codes, barcodes and radio frequency identification |
US20050219204A1 (en) * | 2004-04-05 | 2005-10-06 | Wyatt Huddleston | Interactive display system |
US20050237194A1 (en) * | 2004-04-26 | 2005-10-27 | Microsoft Corporation | Self-monitored active rack |
US20050239546A1 (en) * | 2001-08-03 | 2005-10-27 | Igt | Player tracking communication mechanisms in a gaming machine |
US20050245302A1 (en) * | 2004-04-29 | 2005-11-03 | Microsoft Corporation | Interaction between objects and a virtual environment display |
US20050289590A1 (en) * | 2004-05-28 | 2005-12-29 | Cheok Adrian D | Marketing platform |
US7081033B1 (en) * | 2000-03-07 | 2006-07-25 | Hasbro, Inc. | Toy figure for use with multiple, different game systems |
US7397464B1 (en) * | 2004-04-30 | 2008-07-08 | Microsoft Corporation | Associating application states with a physical object |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004337502A (en) | 2003-05-19 | 2004-12-02 | Namco Ltd | Input device and game system |
-
2006
- 2006-03-09 US US11/372,327 patent/US7843471B2/en active Active
-
2007
- 2007-02-20 WO PCT/EP2007/051617 patent/WO2007101785A1/en active Application Filing
Patent Citations (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5853327A (en) * | 1994-07-28 | 1998-12-29 | Super Dimension, Inc. | Computerized game board |
US5953686A (en) * | 1995-08-03 | 1999-09-14 | Interval Research Corporation | Video camera based computer input system with interchangeable physical interface |
US6460851B1 (en) * | 1996-05-10 | 2002-10-08 | Dennis H. Lee | Computer interface apparatus for linking games to personal computers |
US6167353A (en) * | 1996-07-03 | 2000-12-26 | Interval Research Corporation | Computer method and apparatus for interacting with a physical system |
US6165068A (en) * | 1997-01-22 | 2000-12-26 | Tomy Company, Ltd. | Connection-fighting type game machine and connection-fighting type game methods |
US6835131B1 (en) * | 1998-11-26 | 2004-12-28 | Innovision Research & Technology Plc | Game apparatus for supplying power to playing elements |
US20050069171A1 (en) * | 1999-05-19 | 2005-03-31 | Rhoads Geoffrey B. | Embedded data in gaming objects for authentication and association of behavior information |
US6190174B1 (en) * | 1999-06-03 | 2001-02-20 | Kader Industrial Company Limited | Electronic story board |
US7081033B1 (en) * | 2000-03-07 | 2006-07-25 | Hasbro, Inc. | Toy figure for use with multiple, different game systems |
US6690156B1 (en) * | 2000-07-28 | 2004-02-10 | N-Trig Ltd. | Physical object location apparatus and method and a graphic display device using the same |
US20020090985A1 (en) * | 2000-09-07 | 2002-07-11 | Ilan Tochner | Coexistent interaction between a virtual character and the real world |
US20020185532A1 (en) * | 2001-06-07 | 2002-12-12 | Berquist David T. | RFID data collection and use |
US6761634B1 (en) * | 2001-06-07 | 2004-07-13 | Hasbro, Inc. | Arcade table |
US20050239546A1 (en) * | 2001-08-03 | 2005-10-27 | Igt | Player tracking communication mechanisms in a gaming machine |
US20030134679A1 (en) * | 2001-11-19 | 2003-07-17 | Radica China Ltd. | Electronic gaming device using coded input data |
US20050199723A1 (en) * | 2001-11-30 | 2005-09-15 | Allen Lubow | Method for identifying and authenticating goods using codes, barcodes and radio frequency identification |
US20030171984A1 (en) * | 2002-03-06 | 2003-09-11 | Wodka Joseph F. | Customization of promotional material through use of programmable radio frequency identification technology |
US20040127277A1 (en) * | 2002-10-09 | 2004-07-01 | Walker Jay S. | Method and apparatus for authenticating data relating to usage of a gaming device |
US20040145474A1 (en) * | 2002-11-15 | 2004-07-29 | Schmidtberg Rupert A. | RF identification reader for communicating condition information associated with the reader |
US20040111335A1 (en) * | 2002-12-04 | 2004-06-10 | Black Charles Ronald | RFID space monitoring and asset inventory system |
US20050099268A1 (en) * | 2003-11-12 | 2005-05-12 | Ari Juels | Radio frequency identification system with privacy policy implementation based on device classification |
US20050174236A1 (en) * | 2004-01-29 | 2005-08-11 | Brookner George M. | RFID device tracking and information gathering |
US20050219204A1 (en) * | 2004-04-05 | 2005-10-06 | Wyatt Huddleston | Interactive display system |
US20050237194A1 (en) * | 2004-04-26 | 2005-10-27 | Microsoft Corporation | Self-monitored active rack |
US20050245302A1 (en) * | 2004-04-29 | 2005-11-03 | Microsoft Corporation | Interaction between objects and a virtual environment display |
US7397464B1 (en) * | 2004-04-30 | 2008-07-08 | Microsoft Corporation | Associating application states with a physical object |
US20050289590A1 (en) * | 2004-05-28 | 2005-12-29 | Cheok Adrian D | Marketing platform |
Cited By (136)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130288563A1 (en) * | 2005-10-21 | 2013-10-31 | Patent Category Corporation | Interactive toy system |
US8353767B1 (en) | 2007-07-13 | 2013-01-15 | Ganz | System and method for a virtual character in a virtual world to interact with a user |
US8128500B1 (en) * | 2007-07-13 | 2012-03-06 | Ganz | System and method for generating a virtual environment for land-based and underwater virtual characters |
US11803659B2 (en) | 2007-08-23 | 2023-10-31 | Ebay Inc. | Sharing information on a network-based social platform |
US11869097B2 (en) | 2007-08-23 | 2024-01-09 | Ebay Inc. | Viewing shopping information on a network based social platform |
US20130346239A1 (en) * | 2007-08-23 | 2013-12-26 | Ebay Inc. | Methods and systems to facilitate a purchase of an item on a network-based marketplace |
US11106819B2 (en) | 2007-08-23 | 2021-08-31 | Ebay Inc. | Sharing information on a network-based social platform |
US9092819B2 (en) * | 2007-08-23 | 2015-07-28 | Ebay Inc. | Method, system, and medium of controlling a visually perceptible output in relation to a virtual gift |
US11080797B2 (en) | 2007-08-23 | 2021-08-03 | Ebay Inc. | Viewing shopping information on a network based social platform |
US10037564B2 (en) | 2007-08-23 | 2018-07-31 | Paypal, Inc. | Methods and systems to facilitate a purchase of an item on a network-based marketplace |
US10339613B2 (en) | 2007-08-23 | 2019-07-02 | Ebay Inc. | Viewing shopping information on a network based social platform |
US10984126B2 (en) | 2007-08-23 | 2021-04-20 | Ebay Inc. | Sharing information on a network-based social platform |
JP2011501291A (en) * | 2007-10-15 | 2011-01-06 | インターナショナル・ビジネス・マシーンズ・コーポレーション | Method and apparatus for bridging real world web applications and 3D virtual worlds |
US20090100352A1 (en) * | 2007-10-15 | 2009-04-16 | Yunwu Huang | Method and apparatus for bridging real-world web applications and 3d virtual worlds |
US8370370B2 (en) * | 2007-10-15 | 2013-02-05 | International Business Machines Corporation | Bridging real-world web applications and 3D virtual worlds |
US20090164916A1 (en) * | 2007-12-21 | 2009-06-25 | Samsung Electronics Co., Ltd. | Method and system for creating mixed world that reflects real state |
JP2009279347A (en) * | 2008-05-26 | 2009-12-03 | Taito Corp | Game apparatus, and pen inputting device |
US20100005007A1 (en) * | 2008-07-07 | 2010-01-07 | Aaron Roger Cox | Methods of associating real world items with virtual world representations |
US20100011312A1 (en) * | 2008-07-11 | 2010-01-14 | International Business Machines Corporation | Rfid reader integration to virtual world monitoring |
US8510681B2 (en) * | 2008-07-11 | 2013-08-13 | International Business Machines Corporation | RFID reader integration to virtual world monitoring |
US20100045660A1 (en) * | 2008-08-21 | 2010-02-25 | Dettinger Richard D | Recommendations for Object Use in Virtual Environment |
US8605115B2 (en) * | 2008-08-21 | 2013-12-10 | International Business Machines Corporation | Recommendations for object use in virtual environment |
US20100145769A1 (en) * | 2008-12-09 | 2010-06-10 | International Business Machines Corporation | System and method for product inquiry and incentive for virtual universes |
US8898574B2 (en) | 2008-12-19 | 2014-11-25 | International Business Machines Corporation | Degrading avatar appearances in a virtual universe |
US20100162136A1 (en) * | 2008-12-19 | 2010-06-24 | International Business Machines Corporation | Degrading avatar appearances in a virtual universe |
US20100192110A1 (en) * | 2009-01-23 | 2010-07-29 | International Business Machines Corporation | Method for making a 3-dimensional virtual world accessible for the blind |
US8271888B2 (en) * | 2009-01-23 | 2012-09-18 | International Business Machines Corporation | Three-dimensional virtual world accessible for the blind |
WO2010093995A1 (en) * | 2009-02-13 | 2010-08-19 | Social Gaming Network | Apparatuses, methods and systems for an interworld feedback platform bridge |
US9633465B2 (en) | 2009-02-28 | 2017-04-25 | International Business Machines Corporation | Altering avatar appearances based on avatar population in a virtual universe |
US20100220097A1 (en) * | 2009-02-28 | 2010-09-02 | International Business Machines Corporation | Altering avatar appearances based on avatar population in a virtual universe |
US20110055927A1 (en) * | 2009-08-27 | 2011-03-03 | International Business Machines Corporation | Updating assets rendered in a virtual world environment based on detected user interactions in another world |
US9122380B2 (en) | 2009-08-27 | 2015-09-01 | International Business Machines Corporation | Updating assets rendered in a virtual world environment based on detected user interactions in another world |
US8307308B2 (en) | 2009-08-27 | 2012-11-06 | International Business Machines Corporation | Updating assets rendered in a virtual world environment based on detected user interactions in another world |
US9904442B2 (en) | 2009-08-27 | 2018-02-27 | International Business Machines Corporation | Updating assets rendered in a virtual world environment based on detected user interactions in another world |
US8719731B2 (en) | 2009-08-27 | 2014-05-06 | International Business Machines Corporation | Updating assets rendered in a virtual world environment based on detected user interactions in another world |
US10754513B2 (en) | 2009-08-27 | 2020-08-25 | International Business Machines Corporation | Updating assets rendered in a virtual world environment based on detected user interactions in another world |
US8864589B2 (en) * | 2009-10-27 | 2014-10-21 | Activision Publishing, Inc. | Video game with representative physical object related content |
US20110098092A1 (en) * | 2009-10-27 | 2011-04-28 | Reiche Iii Paul | Video game with representative physical object related content |
US10673932B2 (en) | 2009-12-02 | 2020-06-02 | International Business Machines Corporation | System and method for abstraction of objects for cross virtual universe deployment |
US9882961B2 (en) * | 2009-12-02 | 2018-01-30 | International Business Machines Corporation | System and method for abstraction of objects for cross virtual universe deployment |
US8495078B2 (en) * | 2009-12-02 | 2013-07-23 | International Business Machines Corporation | System and method for abstraction of objects for cross virtual universe deployment |
US20110165939A1 (en) * | 2010-01-05 | 2011-07-07 | Ganz | Method and system for providing a 3d activity in a virtual presentation |
US20110231433A1 (en) * | 2010-03-17 | 2011-09-22 | Ricoh Company, Limited | Management system, management method, and temporary storage document server |
US8836719B2 (en) | 2010-04-23 | 2014-09-16 | Ganz | Crafting system in a virtual environment |
US20120016761A1 (en) * | 2010-07-15 | 2012-01-19 | Enyama, Inc. | Techniques For Provisioning Content |
US8564621B2 (en) * | 2010-08-11 | 2013-10-22 | International Business Machines Corporation | Replicating changes between corresponding objects |
US20120038667A1 (en) * | 2010-08-11 | 2012-02-16 | International Business Machines Corporation | Replicating Changes Between Corresponding Objects |
US20150026587A1 (en) * | 2011-04-01 | 2015-01-22 | Arthur Austin Ollivierre | System and Method for Displaying Objects in a User Interface Based on a Visual Acuity of a Viewer |
US20120254779A1 (en) * | 2011-04-01 | 2012-10-04 | Arthur Austin Ollivierre | System and method for displaying objects in a user interface based on a visual acuity of a viewer |
US8881058B2 (en) * | 2011-04-01 | 2014-11-04 | Arthur Austin Ollivierre | System and method for displaying objects in a user interface based on a visual acuity of a viewer |
US20120295700A1 (en) * | 2011-05-17 | 2012-11-22 | Paul Reiche | Conditional access to areas in a video game |
US9808721B2 (en) | 2011-05-17 | 2017-11-07 | Activision Publishing, Inc. | Conditional access to areas in a video game |
US10315119B2 (en) * | 2011-05-17 | 2019-06-11 | Activision Publishing, Inc. | Video game with concurrent processing of game-related physical objects |
US9180378B2 (en) * | 2011-05-17 | 2015-11-10 | Activision Publishing, Inc. | Conditional access to areas in a video game |
US20120295699A1 (en) * | 2011-05-17 | 2012-11-22 | Paul Reiche | Conditional access to areas in a video game |
US9381430B2 (en) * | 2011-05-17 | 2016-07-05 | Activision Publishing, Inc. | Interactive video game using game-related physical objects for conducting gameplay |
US20130296058A1 (en) * | 2012-05-01 | 2013-11-07 | Activision Publishing, Inc. | Server based interactive video game with toys |
US9649565B2 (en) * | 2012-05-01 | 2017-05-16 | Activision Publishing, Inc. | Server based interactive video game with toys |
US9937417B2 (en) * | 2012-10-10 | 2018-04-10 | Activision Publishing, Inc. | Interactive video game with different sized toys having different abilities within the video game |
US20140100029A1 (en) * | 2012-10-10 | 2014-04-10 | Paul Reiche | Interactive video game with different sized toys having different abilities within the video game |
US9387407B2 (en) | 2012-12-26 | 2016-07-12 | Disney Enterprises, Inc. | Managing objectives associated with a virtual space based on characters made accessible responsive to corresponding tokens being detected |
US9552434B2 (en) | 2012-12-26 | 2017-01-24 | Disney Enterprises, Inc. | Providing a common virtual item repository in a virtual space |
US8726172B1 (en) | 2012-12-26 | 2014-05-13 | Disney Enterprises, Inc. | Managing an environment of a virtual space based on characters made accessible responsive to corresponding tokens being detected |
US8986115B2 (en) * | 2012-12-26 | 2015-03-24 | Disney Enterprises, Inc. | Facilitating customization of a virtual space based on accessible virtual items |
US8972369B2 (en) | 2012-12-26 | 2015-03-03 | Disney Enterprises, Inc. | Providing a common virtual item repository in a virtual space |
US9327200B2 (en) | 2012-12-26 | 2016-05-03 | Disney Enterprises, Inc. | Managing a theme of a virtual space based on characters made accessible responsive to corresponding tokens being detected |
US20140181820A1 (en) * | 2012-12-26 | 2014-06-26 | Disney Enterprises, Inc. | Facilitating customization of a virtual space based on accessible virtual items |
US9667624B2 (en) | 2012-12-26 | 2017-05-30 | Disney Enterprises, Inc. | Managing an environment of a virtual space based on characters made accessible responsive to corresponding tokens being detected |
US9922185B2 (en) | 2012-12-26 | 2018-03-20 | Disney Enterprises, Inc. | Linking token detection at a single computing platform with a user identification to effectuate modifications in virtual space instances presented via multiple computing platforms |
US8910266B2 (en) | 2012-12-26 | 2014-12-09 | Disney Enterprises, Inc. | Linking token detection at a single computing platform with a user identification to unlock content in virtual space instances presented via multiple computing platforms |
US9457263B2 (en) | 2012-12-26 | 2016-10-04 | Disney Enterprises, Inc. | Unlocking virtual items in a virtual space responsive to physical token detection |
US9517404B2 (en) | 2012-12-26 | 2016-12-13 | Disney Enterprises, Inc. | Apparatus, system, and method for effectuating modifications to a virtual space responsive to token detection |
US8909920B2 (en) | 2012-12-26 | 2014-12-09 | Disney Enterprises, Inc. | Linking token detection at a single computing platform with a user identification to effectuate modifications in virtual space instances presented via multiple computing platforms |
US9704336B2 (en) | 2012-12-26 | 2017-07-11 | Disney Enterprises, Inc. | Managing a theme of a virtual space based on characters made accessible responsive to corresponding tokens being detected |
USD748199S1 (en) | 2013-01-15 | 2016-01-26 | Disney Enterprises, Inc. | Multi-sided power disk |
USD748200S1 (en) | 2013-01-15 | 2016-01-26 | Disney Enterprises, Inc. | Power disk reader |
US9424689B2 (en) * | 2013-03-05 | 2016-08-23 | Nintendo Co., Ltd. | System,method,apparatus and computer readable non-transitory storage medium storing information processing program for providing an augmented reality technique |
US20140253591A1 (en) * | 2013-03-05 | 2014-09-11 | Nintendo Co., Ltd. | Information processing system, information processing apparatus, information processing method, and computer-readable recording medium recording information processing program |
US9582147B2 (en) * | 2013-03-08 | 2017-02-28 | Nintendo Co., Ltd. | Inserting game objects through contactless communication |
US20140256430A1 (en) * | 2013-03-08 | 2014-09-11 | Nintendo Co., Ltd. | Information processing apparatus, information processing system, information processing method and recording medium |
US9675895B2 (en) * | 2013-03-13 | 2017-06-13 | Hasbro, Inc. | Three way multidirectional interactive toy |
US20140273717A1 (en) * | 2013-03-13 | 2014-09-18 | Hasbro, Inc. | Three way multidirectional interactive toy |
US9547871B2 (en) * | 2013-03-15 | 2017-01-17 | Activision Publishing, Inc. | System and method for purchasing physical toys and corresponding virtual toys |
US20150294508A1 (en) * | 2013-03-15 | 2015-10-15 | Disney Enterprises, Inc. | Effectuating modifications within an instance of a virtual space presented via multiple disparate client computing platforms responsive to detection of a token associated with a single client computing platform |
US8904298B2 (en) | 2013-03-15 | 2014-12-02 | Disney Enterprises, Inc. | Effectuating modifications within an instance of a virtual space presented via multiple disparate client computing platforms responsive to detection of a token associated with a single client computing platform |
US20140274313A1 (en) * | 2013-03-15 | 2014-09-18 | Karthik Bala | System and method for purchasing physical toys and corresponding virtual toys |
US9092114B2 (en) | 2013-03-15 | 2015-07-28 | Disney Enterprises, Inc. | Effectuating modifications within an instance of a virtual space presented via multiple disparate client computing platforms responsive to detection of a token associated with a single client computing platform |
US9776071B2 (en) | 2013-05-09 | 2017-10-03 | Mattel, Inc. | Resonant coils for use with games and toys |
US10994191B2 (en) | 2013-05-09 | 2021-05-04 | Mattel, Inc. | Resonant coils for use with games and toys |
US10413807B2 (en) | 2013-05-09 | 2019-09-17 | Mattel, Inc. | Resonant coils for use with games and toys |
CN104281787A (en) * | 2013-07-01 | 2015-01-14 | 迪士尼企业公司 | Implementing encrypted content in a game |
JP2015014995A (en) * | 2013-07-08 | 2015-01-22 | 桑原 雅人 | Display device, display method, program, and display system |
US20150009190A1 (en) * | 2013-07-08 | 2015-01-08 | Masato Kuwahara | Display device, storage medium, display method and display system |
US9817987B2 (en) * | 2013-12-23 | 2017-11-14 | Dropbox, Inc. | Restricting access to content |
US11372990B2 (en) | 2013-12-23 | 2022-06-28 | Dropbox, Inc. | Restricting access to content |
US20150178516A1 (en) * | 2013-12-23 | 2015-06-25 | Dropbox, Inc. | Restricting access to content |
US9898844B2 (en) * | 2013-12-31 | 2018-02-20 | Daqri, Llc | Augmented reality content adapted to changes in real world space geometry |
US20150187108A1 (en) * | 2013-12-31 | 2015-07-02 | Daqri, Llc | Augmented reality content adapted to changes in real world space geometry |
US10188939B2 (en) * | 2014-03-11 | 2019-01-29 | Microsoft Technology Licensing, Llc | Modular construction for interacting with software |
US20150258435A1 (en) * | 2014-03-11 | 2015-09-17 | Microsoft Corporation | Modular construction for interacting with software |
US10445437B2 (en) | 2014-03-11 | 2019-10-15 | Microsoft Technology Licensing, Llc | Generation of custom modular objects |
US10089253B2 (en) | 2014-03-11 | 2018-10-02 | Microsoft Technology Licensing, Llc | Data store for a modular assembly system |
US10159894B2 (en) | 2014-03-11 | 2018-12-25 | Microsoft Technology Licensing, Llc | Gaming system for modular toys |
US10150043B2 (en) | 2014-03-11 | 2018-12-11 | Microsoft Technology Licensing, Llc | Interactive smart beads |
US10537821B2 (en) * | 2014-06-30 | 2020-01-21 | Microsoft Technology Licensing, Llc | Interactive play sets |
US20150375134A1 (en) * | 2014-06-30 | 2015-12-31 | Microsoft Corporation | Interactive play sets |
US20160171770A1 (en) * | 2014-12-10 | 2016-06-16 | Sixense Entertainment, Inc. | System and Method for Assisting a User in Locating Physical Objects While the User is in a Virtual Reality Environment |
US9690367B2 (en) * | 2014-12-10 | 2017-06-27 | Sixense Entertainment, Inc. | System and method for assisting a user in locating physical objects while the user is in a virtual reality environment |
US10881970B2 (en) | 2015-04-08 | 2021-01-05 | Lego A/S | Game system |
KR102513852B1 (en) * | 2015-04-08 | 2023-03-27 | 레고 에이/에스 | game system |
US10799802B2 (en) | 2015-04-08 | 2020-10-13 | Lego A/S | Game system |
KR20170134644A (en) * | 2015-04-08 | 2017-12-06 | 레고 에이/에스 | Game system |
JP2018516109A (en) * | 2015-04-08 | 2018-06-21 | レゴ エー/エス | Game system |
US10157487B2 (en) * | 2015-07-30 | 2018-12-18 | International Business Machines Corporation | VR biometric integration |
US11517809B2 (en) * | 2017-06-12 | 2022-12-06 | Bandai Co., Ltd. | Game device, program product, game system and article for games |
US11224814B2 (en) * | 2018-01-30 | 2022-01-18 | Netease (Hangzhou) Network Co., Ltd. | Processing method and apparatus for virtual resource in game scene |
WO2019183485A1 (en) * | 2018-03-22 | 2019-09-26 | Infinite Kingdoms Llc | Connected avatar technology |
US11348298B2 (en) | 2018-03-22 | 2022-05-31 | Infinite Kingdoms Llc | Connected avatar technology |
US20200108311A1 (en) * | 2018-10-09 | 2020-04-09 | Ivan Mauricio Buritica Suarez | System and method for emulating an electronic game to play in a physical gaming environment |
US11534685B2 (en) * | 2018-10-09 | 2022-12-27 | Ivan Mauricio Buritica Suarez | System and method for emulating an electronic game to play in a physical gaming environment |
US10912995B2 (en) * | 2018-10-09 | 2021-02-09 | Ivan Mauricio Buritica Suarez | System and method for emulating an electronic game to play in a physical gaming environment |
US20230226448A1 (en) * | 2019-12-09 | 2023-07-20 | Jeremy Noah Anderson | Game player created object and video game play |
US12017147B2 (en) * | 2019-12-09 | 2024-06-25 | Activision Publishing, Inc. | Game player created object and video game play |
US11369877B2 (en) * | 2019-12-24 | 2022-06-28 | Nintendo Co., Ltd. | Computer-readable non-transitory storage medium having game program stored therein, game apparatus, game system, and game processing method for providing user interface showing change in parameter corresponding to an additional item |
US11724190B2 (en) | 2020-01-21 | 2023-08-15 | Mythical, Inc. | Systems and methods for modifying game assets of an online gaming platform based on tradeable items |
US11247127B1 (en) | 2020-01-21 | 2022-02-15 | Mythical, Inc. | Systems and methods for modifying game assets of an online gaming platform based on tradeable items |
US20210387093A1 (en) * | 2020-01-21 | 2021-12-16 | Mythical, Inc. | Systems and methods for adjusting previously-made modifications of game assets based on changes in the ownership of tradeable items |
US11173401B1 (en) * | 2020-01-21 | 2021-11-16 | Mythical, Inc. | Systems and methods for adjusting previously-made modifications of game assets based on changes in the ownership of tradeable items |
US11376508B1 (en) | 2020-04-03 | 2022-07-05 | Mythical, Inc. | Systems and methods for enabling exchange of tradeable items |
US11865458B2 (en) | 2020-04-03 | 2024-01-09 | Mythical, Inc. | Systems and methods for enabling exchange of tradeable items |
US11717754B2 (en) * | 2020-08-27 | 2023-08-08 | Bandai Co., Ltd. | Game support system, program, information communication terminal, and connection device |
CN113521756A (en) * | 2020-08-27 | 2021-10-22 | 万代股份有限公司 | Game support system, program, information communication terminal, and connection device |
KR102605262B1 (en) * | 2020-08-27 | 2023-11-24 | 가부시키가이샤 반다이 | Game supporting system, program, and information communication terminal |
CN113476857A (en) * | 2020-08-27 | 2021-10-08 | 万代股份有限公司 | Game fighting system, program, terminal and connection device |
KR20220027754A (en) * | 2020-08-27 | 2022-03-08 | 가부시키가이샤 반다이 | Game supporting system, program, information communication terminal, and connection device |
CN115272630A (en) * | 2022-09-29 | 2022-11-01 | 南方科技大学 | Data processing method, device, virtual reality glasses and storage medium |
Also Published As
Publication number | Publication date |
---|---|
US7843471B2 (en) | 2010-11-30 |
WO2007101785A1 (en) | 2007-09-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7843471B2 (en) | Persistent authenticating mechanism to map real world object presence into virtual world object awareness | |
CN113457150B (en) | Information prompting method and device, storage medium and electronic equipment | |
US8261199B2 (en) | Breakpoint identification and presentation in virtual worlds | |
Condon | Computational models of games. | |
US20120276997A1 (en) | Systems and methods of importing virtual objects using barcodes | |
US9898700B2 (en) | Method, apparatus, and artificial intelligence server for determining artificial intelligence behavior | |
US8909920B2 (en) | Linking token detection at a single computing platform with a user identification to effectuate modifications in virtual space instances presented via multiple computing platforms | |
US20120295702A1 (en) | Optional animation sequences for character usage in a video game | |
JP2010227182A (en) | On-line game server processing method, on-line game server device, on-line game server processing programs, online game client processing programs, and on-line game server processing program recording media | |
CN106964155A (en) | The placement of user profile in gamespace | |
CN105214310A (en) | A kind of game fight method of calibration, system and service end | |
US20150033154A1 (en) | Effectuating modifications within an instance of a virtual space presented via multiple disparate client computing platforms responsive to detection of a token associated with a single client computing platform | |
CN113813603B (en) | Game display control method and device, electronic equipment and storage medium | |
US8878873B2 (en) | Enhanced visibility of avatars satisfying a profile | |
US8932138B2 (en) | System and method for placing players of a multiplayer game in a virtual world | |
JP7542743B2 (en) | Information display method, device, electronic device, and computer program | |
JP2013240365A (en) | Game system | |
WO2016130379A2 (en) | System and method for providing an enhanced marketing, sale, or order fulfillment experience related to action figures or action figure accessories having corresponding virtual counterparts | |
CN108025214A (en) | Game device and program | |
CN109499069A (en) | Operating result method of calibration, device, storage medium and electronic device | |
US20060123351A1 (en) | System and method for communicating objects status within a virtual environment using translucency | |
CN107638688A (en) | For strengthening the equipment, system and method for game experiencing | |
CN110801629A (en) | Method, device, terminal and medium for displaying virtual object life value prompt graph | |
JP2015003038A (en) | Server apparatus, control method thereof, program, and game system | |
CN114404971B (en) | Virtual object interaction method, device, equipment, medium and product |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:DOAN, CHRISTOPHER H.;HIRSAVE, PRAVEEN P.;SAENZ, III, LEONEL;AND OTHERS;REEL/FRAME:017495/0074;SIGNING DATES FROM 20060221 TO 20060222 Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:DOAN, CHRISTOPHER H.;HIRSAVE, PRAVEEN P.;SAENZ, III, LEONEL;AND OTHERS;SIGNING DATES FROM 20060221 TO 20060222;REEL/FRAME:017495/0074 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
AS | Assignment |
Owner name: ACTIVISION PUBLISHING, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:029900/0285 Effective date: 20121231 |
|
AS | Assignment |
Owner name: BANK OF AMERICA, N.A., WASHINGTON Free format text: SECURITY AGREEMENT;ASSIGNOR:ACTIVISION BLIZZARD, INC.;REEL/FRAME:031435/0138 Effective date: 20131011 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: ACTIVISION ENTERTAINMENT HOLDINGS, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:040381/0487 Effective date: 20161014 Owner name: ACTIVISION BLIZZARD INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:040381/0487 Effective date: 20161014 Owner name: BLIZZARD ENTERTAINMENT, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:040381/0487 Effective date: 20161014 Owner name: ACTIVISION PUBLISHING, INC., CALIFORNIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:040381/0487 Effective date: 20161014 Owner name: ACTIVISION ENTERTAINMENT HOLDINGS, INC., CALIFORNI Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:BANK OF AMERICA, N.A.;REEL/FRAME:040381/0487 Effective date: 20161014 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552) Year of fee payment: 8 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |