EP1159668A2 - Systemes et procedes de divertissement interactif - Google Patents

Systemes et procedes de divertissement interactif

Info

Publication number
EP1159668A2
EP1159668A2 EP00905848A EP00905848A EP1159668A2 EP 1159668 A2 EP1159668 A2 EP 1159668A2 EP 00905848 A EP00905848 A EP 00905848A EP 00905848 A EP00905848 A EP 00905848A EP 1159668 A2 EP1159668 A2 EP 1159668A2
Authority
EP
European Patent Office
Prior art keywords
page
physical object
recited
pages
audiovisual program
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP00905848A
Other languages
German (de)
English (en)
Inventor
Christopher H. Schmidt
Adam C. Jordon
Philippe P. Piernot
William A. Cesarotti
Despina Papadopoulos
Robin G. Petravic
David W. Laituri
Jenny Dana Wirtschafter
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Interlego AG
Lego AS
Original Assignee
Zowie Intertainment Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zowie Intertainment Inc filed Critical Zowie Intertainment Inc
Publication of EP1159668A2 publication Critical patent/EP1159668A2/fr
Withdrawn legal-status Critical Current

Links

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/214Input arrangements for video game devices characterised by their sensors, purposes or types for locating contacts on a surface, e.g. floor mats or touch pads
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/40Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
    • A63F13/42Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
    • A63F13/424Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving acoustic input signals, e.g. by using the results of pitch or rhythm extraction or voice recognition
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/40Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
    • A63F13/42Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
    • A63F13/428Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving motion or position input signals, e.g. signals representing the rotation of an input controller or a player's arm motions sensed by accelerometers or gyroscopes
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/50Controlling the output signals based on the game progress
    • A63F13/54Controlling the output signals based on the game progress involving acoustic signals, e.g. for simulating revolutions per minute [RPM] dependent engine sounds in a driving game or reverberation against a virtual wall
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/60Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor
    • A63F13/65Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor automatically by game devices or servers from real world data, e.g. measurement in live racing competition
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F3/00Board games; Raffle games
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F1/00Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/213Input arrangements for video game devices characterised by their sensors, purposes or types comprising photodetecting means, e.g. cameras, photodiodes or infrared cells
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/90Constructional details or arrangements of video game devices not provided for in groups A63F13/20 or A63F13/25, e.g. housing, wiring, connections or cabinets
    • A63F13/98Accessories, i.e. detachable arrangements optional for the use of the video game device, e.g. grip supports of game controllers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures

Definitions

  • This invention relates generally to computer interactions with a physical system and more particularly to interactions with a physical book.
  • the present invention is also applicable to other types of entertainment systems.
  • a conventional book is a bundle of text that provides the reader with a story having a specific set of characters and events. As one reads a book, the reader may imagine how the characters look and how they react to different events.
  • the conventional book may also include pictures that provide the reader with visual depictions of characters and events from the story.
  • a reader typically reads a conventional book from start to finish, wherein events of the story unfold sequentially in a predetermined ordered that is laid out by the text description. In other words, there is no deviation from the story line; the book's story or text remains fixed. In sum, conventional books are not interactive and do not allow the reader any input as to how the story is presented.
  • some books have been designed that provide limited interactive capabilities.
  • some children's books include sound files that may be selected and played with the push of a button that is built into the book. These sounds typically correspond to the story. For example, barnyard animal noises may be provided with a story about a farm.
  • the present invention provides a system and method for interfacing with a book, or any other type of physical entertainment device, in a complex manner.
  • the entertainment device includes physical objects that may be placed on or associated with a physical base, such as a book or a game board or platform.
  • a user may move the physical objects over the base and/or add and remove physical objects to and from the base.
  • a corresponding audiovisual program is executed to display or play one or more audiovisual segments that add to what is presented within the physical entertainment device.
  • an interactive entertainment system in one embodiment, includes a base that represents an entertainment device, a physical sensor object that is separate from the base, a detector arranged to detect the presence, location and identity of the sensor object relative to the base, and an interactive audiovisual program configured to output audiovisual segments related to the entertainment device represented by the base. Detection of the physical sensor object causes the interactive audiovisual program to output an audiovisual segment that is based at least in part upon the detected position of the sensor object relative to the base and that includes information that is not visible on the base but which is deemed detectable by the sensor object.
  • the sensor object is not itself displayed in the audiovisual segments.
  • the entertainment device is either a game, a toy, a book or a story.
  • the sensor object has a physical form that represents its function.
  • the audiovisual segment output by the audiovisual program includes an enhanced view of a portion of the base located at the detected position of the sensor object, and/or includes an audio segment that is related to a portion of the base located at the detected position of the sensor object.
  • the entertainment system includes a base that represents an entertainment device, a physical character object that represents a character within the interactive entertainment system, a detector arranged to detect the presence, location and identity of the character object relative to the base, and a gesture recognizer arranged to recognize gesture movements of the physical character object based upon a detected state of the character object.
  • the entertainment system further includes an interactive audiovisual program configured to output audiovisual segments related to the entertainment device represented by the base. The recognition of a selected gesture by the gesture recognizer causes the interactive audiovisual program to output an audiovisual segment that is based at least in part upon the detected gesture.
  • the interactive entertainment system includes a base that represents an entertainment device in the form of a game, a toy, or a book and a plurality of physical objects.
  • the physical objects include environmental objects, character objects and sensor objects.
  • the entertainment system also includes a detector arranged to detect the presence, location and identity of the physical objects relative to the base and an interactive audiovisual program configured to output audiovisual segments related to the entertainment device represented by the base based at least in part upon the detected position of the physical objects relative to the base.
  • Detection of a selected sensor object causes the interactive audiovisual program to output an audiovisual segment that includes information that is not visible on the base but which is deemed detectable by the sensor object, detection of a character object causes the interactive audiovisual program to include the detected character in an outputted audiovisual segment, and detection of an environmental object causes the interactive audiovisual program to include the detected environmental object in an outputted audiovisual segment.
  • a method of interfacing with a book system having a plurality of regions is disclosed.
  • the book system is scanned to extract data.
  • the extracted data including a position and an identifier of a physical object that is part of the book system.
  • a region that is nearest to the physical object is identified based on the extracted data.
  • An audiovisual program is run based at least in part on the determinations of the identifier of the physical object and the position of the physical object in relation to the identified region.
  • a computer system for interacting with a book system includes a data input device arranged to receive data that is extracted from the book system.
  • the extracted data includes at least a position and an identification of a selected physical object that is part of the book system.
  • the computer system also includes a data interpreter arranged to identify a region that is nearest to the selected physical object based on the extracted data and a display device configured to outputting an interactive visual image based at least in part on the determinations of the position and identifier of the selected physical object and the nearest region and associated identifier.
  • a computer readable medium containing program instructions for interfacing with a book system having a plurality of regions is disclosed.
  • the computer readable medium includes computer code for scanning the book system to extract data.
  • the extracted data includes a position and an identifier of a physical object that is part of the book system.
  • the computer readable medium also includes computer code for identifying a region that is nearest to the physical object based on the extracted data, computer code for running an audiovisual program based at least in part on the determinations of the identifier of the physical object and the position of the physical object in relation to the identified region, and a computer readable medium for storing the computer readable codes.
  • a book system in an apparatus aspect, includes a plurality of pages with each page having one or more regions, a physical object movable over the pages, and a position sensing device that may be coupled with a computer system.
  • the position sensing device is configurable by the computer system to detect a position of the pages, an identity of the physical object, and a position of the physical object.
  • the computer system is also programmed with instructions to configure the position sensing device to detect the positions of the pages and physical object and the identity of the physical object.
  • the computer system is also programmed with instructions to determine which region is nearest to the physical object and to generate an audiovisual program based on the pages' positions, the physical object's identity and position, and the nearest region.
  • the position sensing device is in the form of an electromagnetic sensing mechanism.
  • a method of interfacing with a book system having a plurality of pages, a first physical object, and a second physical object is disclosed.
  • Each physical object may be placed near one of a plurality of pages.
  • the book system is scanned to extract data, wherein the extracted data including at least a first position and a first identifier of the first selected physical object if the first physical object is placed near one of the pages and/or a second position and a second identifier of the second physical object if the second physical object is placed near one of the pages.
  • a first region that is nearest to the first selected physical object is identified based on the extracted data if the first physical object is placed near one of the pages.
  • a second region that is nearest to the second physical objects is identified based on the extracted data if the second physical object is placed near one of the pages.
  • An audiovisual program is run based at least in part on the first position and first identifier of the first physical object and the identifier of the first region if the first physical object is placed near one of the pages, and an audiovisual program is run based on at least in part on the second position and second identifier of the second physical object and the identifier of the second region if the second physical object is placed near one of the pages.
  • a computer readable medium containing program instructions for interfacing with a book system having a plurality of pages, a first physical object, and a second physical object.
  • Each physical object may be placed near one of a plurality of pages
  • the computer readable medium includes (i) computer code for scanning the book system to extract data with the extracted data including at least a first position and a first identifier of the first selected physical object if the first physical object is placed near one of the pages and including a second position and a second identifier of the second physical object if the second physical object is placed near one of the pages, (ii) computer code for identifying a first region that is nearest to the first selected physical object based on the extracted data if the first physical object is placed near one of the pages, (iii) computer code for identifying a second region that is nearest to the second physical objects based on the extracted data if the second physical object is placed near one of the pages, (iv) computer code for running an audiovisual program based at least in part on the
  • Fig. 1 is a top view of a book system in accordance with one embodiment of the present invention.
  • Figure 2A is a diagrammatic representation of a first example of a page detection mechanism that utilizes electromagnetic detection in accordance with one embodiment of the present invention.
  • Figure 2B is a diagrammatic representation of a second example of a page detection mechanism that utilizes electromagnetic detection in accordance with one embodiment of the present invention.
  • Figure 2C is a diagrammatic representation of a third example of a page detection mechanism that utilizes electromagnetic detection in accordance with one embodiment of the present invention.
  • Figure 2D is a diagrammatic representation of a first example of a page detection mechanism that utilizes capacitance value detection in accordance with one embodiment of the present invention.
  • Figure 2E is a diagrammatic representation of a second example of a page detection mechanism that utilizes capacitance value detection in accordance with one embodiment of the present invention.
  • Figure 2F is a diagrammatic representation of a page detection mechanism that utilizes resistance value detection in accordance with one embodiment of the present invention.
  • Figure 3 is digital photograph of a book system having a plurality of physical objects in accordance with one embodiment of the present invention.
  • Figure 4A is a digital photograph of a book system with a decoding device positioned over a page of the book in accordance with one embodiment of the present invention.
  • Figure 4B is a screen shot generated by an audiovisual program that is based on the position of the decoding device of Figure 4A in accordance with one embodiment of the present invention.
  • Figure 4C is a digital photograph of a book system with a second decoding device positioned over a page of the book in accordance with one embodiment of the present invention.
  • Figure 5 A is a digital photograph of a book system with a magnifying device positioned over a page of the book in accordance with one embodiment of the present invention.
  • Figure 5B is a screen shot generated by an audiovisual program that is based on the position of the magnifying device of Figure 5A in accordance with one embodiment of the present invention.
  • Figure 6 A is a digital photograph of a book system with a listening device positioned over a page of the book in accordance with one embodiment of the present invention.
  • Figure 6B is a screen shot generated by an audiovisual program that is based on the position of the listening device of Figure 6A in accordance with one embodiment of the present invention.
  • Figure 7A is a digital photograph of a book system with a selection device positioned over a page of the book in accordance with one embodiment of the present invention.
  • Figure 7B is a screen shot generated by an audiovisual program that is based on the position of the selection device of Figure 7A in accordance with one embodiment of the present invention.
  • Figure 8A is a digital photograph of a book system with an image card positioned over a page of the book in accordance with one embodiment of the present invention.
  • Figure 8B is a screen shot generated by an audiovisual program that is based on the position of the image card of Figure 8 A in accordance with one embodiment of the present invention.
  • Figure 9 is a flowchart illustrating a process of interfacing with a book system, such as the book system in Figure 1, in accordance with one embodiment of the present invention.
  • Figure 10 is a flow chart illustrating the operation of Figure 9 of interrupting the extracted data in accordance with one embodiment of the present invention.
  • Figure 11 is a flow chart illustrating the operation of Figure 9 of executing the audiovisual program in accordance with one embodiment of the present invention.
  • Figure 12 is a digital photograph of a game system with an environmental object, a character object, and a sensor object positioned over a base in accordance with one embodiment of the present invention.
  • Figure 13A is a screen shot illustrating a display output from an audiovisual program based on the character object, the environmental objects, and the sensing object of Figure 12 in accordance with one embodiment of the present invention.
  • Figure 13B is a screen shot illustrating a display output from an audiovisual program based on a gesture via character object of Figure 12 in accordance with one embodiment of the present invention.
  • Figure 14 is a flowchart illustrating the process of analyzing a gesture of the character object of Figure 12 in accordance with one embodiment of the present invention.
  • one of the embodiments of the current invention includes a method and apparatus for interfacing with a book in a complex manner.
  • the page positions are detected and a corresponding audiovisual program is executed on a computer system. That is, visual images are displayed and/or sounds are played.
  • the audiovisual program is related to the text of the physical book and adds to the book reading experience.
  • the book system may also include tools that may be used with the physical book and allow the audiovisual program to be manipulated or changed.
  • tools e.g., a magnifying tool
  • a particular portion of a particular page e.g., a picture
  • a corresponding audiovisual program e.g., an enlarged view of a portion of the picture
  • the book system includes mechanisms that detect pages and/or tool movements such that events or characters are inserted within an audiovisual representation of the story, in addition to the story represented by text and images on the pages of the physical book.
  • FIG. 1 shows an interactive book system 100 in accordance with one embodiment of the present invention.
  • the detection system includes a physical book unit 103, a computer system 110, and an I/O board 108 for interfacing between the physical book unit 103 and computer system 110.
  • the physical book unit 103 includes a printed circuit (PC) board 101 having an antenna 102, a physical book having a plurality of pages 104, and one or more tools 106. Some of the tools 106 may also include an interactive device 107, such as a button.
  • the book unit 103 may also include a fixed interactive device 109, such as a switch or button, (i.e., the device has a fixed resting position).
  • the computer system 110 includes, among other things, a driver 114, application program 112, display 116, and speakers 118.
  • a driver 114 application program 112
  • display 116 display 116
  • speakers 118 speakers 118.
  • the computer system 110 is shown as being a separate component from the physical book unit 103, of course, the computer system 110 may be integrated with the physical book unit 103. Additionally, it should be understood that other type of interfaces may be used, such as a television system or set top box. Also, the I/O board 108 may be integrated with the physical book unit 103.
  • the interactive book system 100 may include any suitable mechanisms for detecting the presence, positions, and identities of the various tools (e.g., 106) and pages (e.g., 104) associated with the book unit 103.
  • the detection mechanism may detect how many pages are turned to the right side of the book (e.g., page 104a) and how many pages are turned to the left side of the book (e.g., page 104b).
  • the detection mechanism may also be capable of identifying which tools have been placed on the pages of the book unit and on which page and which portion of the page a tool has been placed.
  • the tools and the pages will include a detectable marker.
  • the markers When the markers are placed or moved within the book unit 103 (e.g., by placing a tool on a page or by turning a page), the positions of the markers may be sensed by the detection mechanism. Preferably, the markers also identify the tool or page. That is, the detection mechanism is able to distinguish between the different markers of the different tools and pages.
  • the book unit includes an electromagnetic detection system to detect the tools and pages of the book and to input the detected data to a computer.
  • Any suitable electromagnetic detection system may be implemented. Of course, it should be recognized that other detection systems may be used, such as optical sensors or electronic sensors.
  • the book unit 103 includes a number of resonator circuits (not shown) that are detectable by the antenna 102 when the resonators are activated by an excitation signal having a particular resonator frequency.
  • the resonator circuits may be arranged in any suitable manner to facilitate detection of the tools and. pages of the book unit 103.
  • each tool 106 and page 104 includes a simple resonator circuit (e.g., a coil in series with a capacitor) that resonates at a particular resonating frequency.
  • An excitation signal having a frequency that activates one or more resonator circuit(s) is transmitted on antenna 102.
  • a resonating signal is induced on antenna 102 by the resonator(s) when the resonator(s) resonate at the frequency of the excitation signal.
  • the excitation signal on antenna 102 is affected by the activated resonator(s) and their corresponding positions relative to the antenna.
  • the alteration of the excitation signal may be analyzed to determine which resonator(s) are resonating, and their corresponding positions.
  • the excitation signal is stopped such that the resonating signal may be independently analyzed. That is, when the antenna's excitation signal is stopped, the activated resonator(s) continue to produce a "ringing" signal that continues to be induced on the antenna 102 and, thus, may be clearly detected on the antenna 102 without interference by the excitation signal.
  • the antenna 102 may be any suitable form for sensing positions of various resonator circuits.
  • the antenna is in the form of a loop that is embedded within the PC board 101.
  • the antenna 102 is placed such that resonator circuits and their positions may be sensed by the antenna 102.
  • a resonating signal induced within the antenna 102 depends on the activated resonator(s) position with respect to the antenna 102.
  • a resonator's position, as well as the associated tool or page position may be determined relative to the antenna 102 by analyzing the resonating signal induced within the antenna 102.
  • the book system may include any number and kind of tools.
  • Tools are defined as objects that may be used to interact with portions of the book unit 103.
  • the tool may represent a visual or auditory sensor. That is, as the tool is moved over a selected position of the book unit, additional images and/or sounds related to the selected position are revealed through the audiovisual program. For example, a tool may be used to magnify certain graphic portions of a page, which magnified view is generated by the audiovisual program. By way of another example, a tool may be used as a listening device to reveal sounds related to portions of the book unit.
  • the book system may also include interactive devices, which are defined as objects that have more than one state. For example, an interactive device may be in the form of a button, switch, or knob.
  • the tools and interactive devices may be fixed or movable within the book unit 103.
  • An interactive devices may be associated with a particular tool or may be an independent object within the book unit 103.
  • Each tool, page, and interactive device may be configured with a resonator that resonates at a distinct frequency.
  • the pair may have a single resonator.
  • positions of the tools and/or interactive devices may be separately detected by transmitting excitation signals having different frequencies through antenna 102.
  • a particular excitation signal with a selected frequency activates a corresponding tool's or interactive device's resonator, which produces a distinct resonating signal in antenna 102.
  • each object may be identified by its particular resonating frequency.
  • the resonator circuits of the tools may resonate at a same resonator frequency, but the resonators are positioned at varying " z" positions within the tool itself. Each tool is uniquely identified by the associated z position of the resonator.
  • This embodiment, as well as several other embodiments, for arranging and identifying resonators having a same frequency are described in U.S. patent application number 09/144,951, entitled “Detecting Physical Object States Using ElectroMagnetic Sensors" by Marcos R. Vescovi, et al., filed on 1 September 1998, which application is herein incorporated by reference in its entirety.
  • any suitable mechanisms may be implemented to convert the resonating or detected signals into signals that are appropriate for input into a computer system.
  • the detected signal is sent through the I/O board 108 through driver 1 14 to application program 112.
  • the I/O board 108 is arranged to convert the detected signals to digital signals (e.g., a binary pulse wave).
  • the application software 112 analyzes the detected signals to determine positions and identifications of the various objects of the book unit 103, as well as to generate a corresponding audiovisual program.
  • the driver 114 is arranged to allow communication between the application software 112 and the I/O board 108.
  • positional data is determined based on the detected signal. and the positional data is processed and used to generate or trigger an audiovisual program. Examples of methods for generating audiovisual programs based on positional data are described further in U.S. Patent Application No. 09/018,023 filed February, 2 1998 entitled “Computer Method and Apparatus for Interacting with a Physical System” by Piernot, et al, which application is herein incorporated by reference in its entirety.
  • the detected signal includes positional data regarding the resonator circuits of one or more physical object(s) of the book unit 103, such as the tools, interactive devices, and/or pages. That is, positions of one or more physical object(s) may be ascertained by analyzing the detected signal.
  • the positional data includes six degrees of position states: x, y, and z position, rotational angle, tilt, and yaw.
  • an electromagnetic sensing system that senses six degrees of positional data is described in International PCT Application Number PCT/GB96/02896 published on 29 May 1997 by Dames, et al, which is herein incorporated by reference in its entirety.
  • the positioned data may include any subset of these six degrees of position states.
  • each physical object is in the form of a page of the book
  • the positional information may include information about whether or not the page has been turned, whether or not the page is currently being turned, and/or an exact position (e.g., an angle relative to the book cover) of the page.
  • the following mechanisms may be applied, with minor modifications, to detecting tool positions and/or interactive device positions.
  • FIG. 2A is a diagrammatic representation of a first example of a page detection mechanism that utilizes electromagnetic detection in accordance with one embodiment of the present invention.
  • a plurality of pages 104 are positioned over an antenna 102 that is embedded in PC board 101.
  • the antenna 102 may stand alone without the PC board 101.
  • Each page 104 has an embedded resonator 202 having a particular resonating frequency.
  • Each resonator 202 may be positioned on any suitable location of the associated page such that page position may be accurately determined (e.g. which pages are being viewed within the physical book).
  • each resonator 202 is positioned over a different portion of the antenna 102 (see axis 204 for each resonator), depending on whether or not the page is turned.
  • page 104a when resonator 202a of page 104a is positioned over a left portion of the antenna, page 104a is open and has not been turned. In contrast, when resonator 202a of page 104a is positioned on the right portion of the antenna, page 104a has been turned such that a new page is displayed (i.e., page 104b). Additionally, when resonator 202a of page 104a is positioned between the left and right portions of the antenna, page 104a is being turned (e.g., moving from the left side to the right side of the book, or visa versa).
  • each resonator 202 resonates at a particular frequency of an excitation signal that is transmitted through antenna 102.
  • resonator 202a of page 104a may be activated by transmitting at a first frequency
  • resonator 202b of page 104b may be activated by transmitting at a second different frequency.
  • the excitation signal is deactivated, the resonating signal of one of the resonators 202 may be detected by antenna 102.
  • each resonator 202 is positioned over a different portion of antenna 102. As shown by each axis 204, each resonator 202 is positioned over a different portion of antenna 102.
  • FIG. 2B is a diagrammatic representation of a second example of a page detection mechanism that utilizes electromagnetic detection in accordance with one embodiment of the present invention.
  • the pages 104 are positioned over a base 206 having an embedded resonator 210.
  • the resonator 210 and base 206 are positioned over antenna 102.
  • Each page includes an associated coil 208 (e.g. without a capacitor or resonator).
  • the resonating signal of resonator 210 depends on how many coils 208 are positioned within a certain distance of the resonator 210.
  • Each coil 208 may be positioned adjacent to resonator 210 (i.e., the page is turned to the left side of the book), positioned near resonator 210 (i.e., as the page is being turned), or positioned far from resonator 210 (i.e., the page is turned to the right side).
  • the resonator 210 is located within the left portion of the book. As shown, pages 104a through 104c are flipped to the left side of the book and, thus. coils 208a through 208c are located in proximity to resonator 210. In contrast, pages 104a through 104c are flipped to the left side of the book and, thus. coils 208a through 208c are located in proximity to resonator 210. In contrast, pages
  • the coils may be replaced with strips of conductive material.
  • Each strip of conductive material from each pages may alter the detected signal in a quantifiable manner.
  • the detected signal may be analyzed to determine how many strips of conductive material are affecting the signal and thereby to determine how many pages have been turned.
  • FIG. 2C is a diagrammatic representation of a third example of a page detection mechanism that utilizes electromagnetic detection in accordance with one embodiment of the present invention. As shown, each page has an associated coil
  • the coils 211 are serially coupled together, along with a resonating capacitor
  • each coil either subtracts or adds to the amplitude of the resonating signal. For example, as each page that is turned to the right side of the book (as shown, pages 104d through 104f), the coils of the turned pages (coils 21 Id through 21 If) may subtract from the detected signal's amplitude. In this example, for each page that is not turned and remains on the left of the book (as shown, pages 104a through 104c), each associated coil (coil 211a through 211c) continues to contribute to the detected signal's amplitude. Thus, the detected signal may be analyzed to determine how many coils are contributing to the signal and thereby determine how many pages have been turned or not turned.
  • each page may contribute to a capacitance value between two points.
  • strips of conductive material may be configured on each page such that the capacitance value between two points is increased or decreased as pages of the book are turned.
  • Figure 2D is a diagrammatic representation of a first example of a page detection mechanism that utilizes capacitance value detection in accordance with one embodiment of the present invention. As shown, a first point 214a and a second point
  • each page 104 has an associated conductive strip 216 that may be positioned over the two points 214.
  • the points 214 are located on the right portion of the book.
  • the associated conductive strip 216 of that page contributes to the capacitance value as measured between the two points 214.
  • conductive strips 216d, 216e, and 216f of pages 104d, 104e, and 104f contribute to the capacitance value since these pages are positioned over the two points 214.
  • the conductive strips 216 may be positioned in any suitable position that affects the capacitance value between the two points 214.
  • the strips 216 stretch between the two points 214 when the corresponding pages are positioned over the two points 214.
  • the strips 216 may also overlap the two points 214 or cover an area that is larger than the area between the two points 214. Alternatively, the strips 216 may cover an area that is smaller than the area between the two points 214.
  • FIG. 2E is a diagrammatic representation of a second example of a page detection mechanism that utilizes capacitance value detection in accordance with one embodiment of the present invention.
  • each page 104 has an associated conductive strip 218, wherein the strips may be interdigitated with each other. That is, the strips 218 may be positioned over different areas between the points 220a and 220b.
  • each strip 218 affects the measured capacitance value between points 220a and 220b. Since the capacitance value depends on how many pages are positioned over the points 220, it may determined how many pages are turned to one side of the book (i.e., the side that includes the two measurement points 220).
  • FIG. 2F is a diagrammatic representation of a page detection mechanism that utilizes resistance value detection in accordance with one embodiment of the present invention.
  • each page has an associated conductive strip or wire 224 that shorts one of a plurality of contact pairs (e.g., 222c).
  • the contact pairs 222 are positioned along the right edge of the book. As each page is turned to the right side of the book, the associated conductive strip of the turned page touches between a pair of contacts 222 and creates a short circuit. For example, when page 104f is turned to the right side of the book, the conductive strip 224f of page 104f shorts contact pair 222f. Likewise, the conductive strip 224c shorts contact pair 222c when page 104c is turned to the right.
  • Page turning may be determined by measuring the resistance value between each pair of contact pairs 222. If the resistance value is relatively large, the corresponding page has not been turned. If the resistance value is zero, the corresponding page has been turned.
  • the conductive strips 224 may be arranged on each page in any suitable manner so as to short two contact points 222. As shown in Figure 2F, the conductive strips 224 are formed such that they rest between a pair of contacts 222 when the corresponding page is positioned directly over the contact pair 222. Alternatively, the conductive strips or wires may include plugs that are inserted into a pair of conductive holes as the page is turned. This embodiment provides a mechanism for ensuring that a conductive path is formed between the two point. Without the plugs, the conductive strip may not align correctly with the two contacts if, for example, the page is crumpled.
  • the book system may include a magnetic sensing device, and each page may include a magnetic tab that may be sensed by the magnetic sensing device at certain page positions. For example, each page may have a magnetic tab positioned in a different location.
  • the magnetic sensing device generates an signal that is affected by magnetic tabs that are nearby. Thus, by analyzing the signal of the magnetic sensing device, it may be determined whether or not a page is positioned near the magnetic sensing device.
  • photodetectors may be placed within a base of the book.
  • the photodetectors may be placed down the outside edge of one side of the base.
  • Each page is configured into a suitable shape such that a corresponding photodetector is covered or uncovered by the page, depending on whether the page is turned.
  • the pages may have tabs (i.e., in place of each conductive strips 224) that each cover or uncover a corresponding photodetector (i.e., in place of each contact pair 222).
  • FIG. 3 is digital photograph of a book system 103 having a plurality of physical objects in accordance with one embodiment of the present invention. As shown, a book 104 rests on a base 101.
  • the book system 103 also includes a number of moveable physical objects or tools (e.g., 106a through 106e), as well as a fixed physical object 106f. The tools may be placed within particular areas of certain pages of the physical book.
  • the tools may be used to accomplish tasks that are integrated into the story line, for example.
  • the tools are used to obtain clues regarding the identity of a spy. Accordingly, a number of tools are described in reference to Figures 4A through 8B that help the user to determine the identity of a spy.
  • different tools may be used with books having different story content.
  • a set of translation tools may be provided with a Spanish language textbook.
  • the user may move decoding device 106b over particular text portions 702a of a particular page 104b.
  • a decoded message 706 is displayed within audio-visual display window 704.
  • the user is given instructions on what to do next. Specifically, the user is instructed to go to a next page and use a audio tape reading tool to obtain audio clues.
  • a sound decoding tool 106a is moved over one of a plurality of tape strip representations (e.g., 708a through 708j).
  • an audio sample is played to indicate clues for the user as to the identity of the spy. Sounds may be played when the user moves from left to right along a tape strip 708 , from right to left, or any combination therein.
  • an audio segment is played in a forward direction as the tool 106a is moved from left to right, while the same audio segment is played in reverse as the tool 106a is moved from right to left.
  • the audiovisual program may include techniques for recognizing gestures, such as the tape reading gesture of the sound decoding tool 106a. Any suitable techniques may be implemented for recognizing this gesture, as well as other gestures. For example, when the sound decoding tool 106a is moved within a tape strip 708, an index to a sound file (e.g., a wave file) is calculated based on the position of the tool 106a along the strip 708. That is, if the tool 106a is positioned on the far left side of the strip 708, the index references the beginning of the sound file. As the tool 106a moves to the middle of the strip 708, the index also moves (and plays sounds) from the middle of the sound file to the end of the sound file.
  • a sound file e.g., a wave file
  • a magnifying device 106c may be placed over a particular portion of an image 710a of page 104e.
  • a magnified view 710b is displayed in the audiovisual display 704.
  • the audiovisual display 704 includes the image that is actually shown on the physical book (i.e., 710a), but it also shows an additional magnified image 710b.
  • the user gains access to a wealth of audiovisual material, in addition to the text and images that are provided within the physical book.
  • a morphing tool (not shown) may be provided.
  • the morphing tool When the morphing tool is placed over a particular image within the physical book pages, the particular image is displayed by the audiovisual program and morphed into another image.
  • the morphing tool When user places the morphing tool over a photo of a person of age 12, an image of the young person may be displayed and morphed into an older person, such as the same person at age 50.
  • sound may also be played (e.g., through a computer's speakers) as the user moves a particular tool over a particular portion of a page.
  • a listening device 106d is placed on a particular window of a building on page 104f. Appropriate sounds that may be heard through a building's window are then played within the audiovisual program 704.
  • a voice signal 712 is represented and displayed within the audiovisual program 704.
  • a person's voice is played through the computer speakers, for example. In the illustrated embodiment, the voice may describe important clues or indicate where to find other clues that will then help the user to determine the identity of the spy.
  • a tool is also provided that allows the user to select certain image objects that are shown on a particular page of the book.
  • a selection tool 106c is used to select a particular person within a group of people as depicted within a photo on pages 104g and 104h.
  • a corresponding audiovisual display may be generated based on the selected person.
  • the audiovisual program may then indicate whether the selected person is the spy.
  • the audiovisual display 704 indicates that the selected person 714 is not the spy.
  • the tools may also be in the form of visual images that are provided separately from the pages of the book. These separate images may be placed on certain pages of the book.
  • a plurality of cards e.g., 106e and
  • 106g having different faces are provided.
  • the user selects a particular card or person 106g and places it within the center of page 104i.
  • this feature allows the user to add visual images to the story line.
  • the user selects a card having a person that they have identified as the spy.
  • An audiovisual program is then generated that indicates whether the user has correctly identified the spy.
  • the audiovisual display 704 indicates that the user has correctly identified the spy 716.
  • the interfacing software for the physical book may include any suitable operations that accomplish the goal of generating appropriate audiovisual displays and/or sounds that correspond to a user's interactions with a physical book system.
  • Figure 9 is a flowchart illustrating a process 300 of interfacing with a book system, such as the book system in Figure 1 , in accordance with one embodiment of the present invention.
  • a book system is provided in operation 302.
  • portions of the book system are scanned to extract data.
  • the extracted data includes information regarding the physical objects (e.g., tools, pages, and interactive devices) of the book system.
  • the extracted data includes positions and orientation for each physical object.
  • the extracted data may include identifiers associated with each physical object.
  • the data is extracted by initiating an excitation signal at a predetermined frequency on an antenna.
  • the excitation signal is then stopped, and it is determined whether a detected signal is present on the antenna from one or more resonators within the book system.
  • the detected signal may include positional information related to the responding resonators of the book system.
  • the predetermined frequency may correspond to an identity of a responding resonator.
  • a succession of excitation signals having different predetermined frequencies may be initiated on the antenna to detect the presence, position, and identities of multiple physical objects having different resonating frequencies.
  • the extracted data is interpreted in operation 306.
  • the extracted data may be analyzed to determine the relative positions and identities of various physical objects within the book unit.
  • the extracted data may also be analyzed to determined the relative positions and identities of the detected physical objects and various regions of the book unit. Operation 306 will be described in further detail in reference to Figure 10.
  • an audiovisual program is run that is based on the extracted data and any previously extracted and recorded data in operation 308. For example, the relative position of regions of a particular page and the previous physical object positions may have been previously recorded and thereby accessed to generate a new audiovisual program. Operation 308 is explained in more detail below with reference to Figure 11.
  • the audiovisual program may include merely one visual frame or may include one frame of a sequence of frames from a video.
  • the operation of running an audiovisual program may initiate a continuous audiovisual sequence (e.g., a Quicktime movie) or merely continue at a particular point within the movie.
  • the audiovisual program may have only one frame.
  • the audiovisual program may include an interactive game, wherein the player is directed to perform certain tasks with the book and physical objects.
  • the extracted data information is recorded in operation 310.
  • the process 300 ends until it is initiated again (e.g., when the book is opened).
  • Figure 10 is a flow chart illustrating the operation 306 of Figure 9 of interpreting the extracted data in accordance with one embodiment of the present invention.
  • the book system may be designed such that physical objects may be placed on the pages of the book one at a time, or so that two or more physical objects may be placed simultaneously on the book pages.
  • the identifiers of any physical objects are determined in operation 404. For example, it may be determined that a physical object is in the form of a magnifying device.
  • any physical objects are associated with a region or " hot spot" on a particular page on operation 406.
  • certain areas of each page may be defined as interactive zones. That is, when a physical object is placed within a hot zone, a corresponding audiovisual program is generated. Alternatively, an entire page may be defined as an interactive zone.
  • the hot spots that are associated with a physical object are then identified in operation 408. For example, the page number and portion associated with the physical object may be identified to determine what type of interaction will take place with the particular physical object. By way of specific example, a portion of a particular page may be identified as being magnified (i.e., by the audiovisual program) when a magnifying tool is placed on it.
  • the states of any interactive devices are then determined in operation 410.
  • the interactive devices may be in any suitable form for indicating a different state of a physical object.
  • a physical object may include a button which may be pushed in or let out. It is then determined whether the button is pushed in operation 410. The process then returns to operation 308 of Figure 9 where an audiovisual program is executed that is based on the interpretation of the extracted data.
  • operations 402 through operation 410 may be executed in any order that is appropriate for the particular application.
  • operation 402 may be executed subsequent to operation 404.
  • other information from the extracted data may be interpreted or not interpreted depending on the particular requirements of the particular application.
  • operation 410 is not required.
  • an operation to determine the identification of the book itself may be included in order to aid in choosing which audiovisual program will be run in operation 308 of Figure 9.
  • operation 308 of Figure 9 is executed.
  • Figure 11 is a flow chart illustrating the operation 308 of Figure 9 for executing the audiovisual program in accordance with one embodiment of the present invention.
  • a hot spot is defined as a particular portion on a particular page or an entire page.
  • the hot spot regions may be defined as regions where moveable objects may be placed or where fixed interactive devices, such as a fixed push button, are located. If there are no physical objects associated with a hot spot region, the process returns to operation 310 of Figure 9, and the extracted data is recorded.
  • a physical object is selected in operation 504.
  • a physical object may be selected in any suitable manner. For example, a physical object which was placed upon the page first may be selected first.
  • a corresponding audiovisual program is then executed in operation 506. The audiovisual program is based on the selected physical object, the physical object identifier, a state of any associated interactive device, and the associated hot spot identifier. In other words, the audiovisual program depends on which physical object is selected and where such physical object is positioned in relation to hot spot regions.
  • the audiovisual program may also be based on whether or not a page is currently being turned, regardless of whether or not a physical object is present. In one embodiment, if it is determined that a page is being turned, an interstitial audiovisual segment is generated.
  • the interstitial audiovisual segment may include a video clip that shows a character leaving the jungle via an airplane, flying on the airplane from the jungle to the desert, and landing within the desert in the airplane.
  • FIGS 12, 13 A, and 13B include digital photographs of an alternative embodiment of the present invention.
  • a game system 1200 includes a plurality of physical objects (e.g., 1206a, 1204a, and 1202a) that may be positioned over a playing surface or base 1201.
  • the base 1201 may be any shape and/or layout that is suitable for a particular entertainment device.
  • the base 1201 is a board game that includes flat portions and a raised portion (e.g., a bridge 1203).
  • the surface pattern on the base 1201 may represent any suitable playing surface.
  • the base 1201 represents a backyard area.
  • the game system 1200 may include physical objects in the form of any of the described tools of the present invention. Additionally, the game system may include any other suitable types of physical objects. In the illustrated embodiment, the physical objects may be categorized into three general types: character objects, environmental objects, and sensor objects.
  • a character object is a physical object that represents a character within the game. For example, a character object may represent a person, animal, or any animated object. Generally, a character interacts with other game objects during the course of the game. Accordingly, the audiovisual program will include segments that show detected characters interacting with other items based in part on the position of detected character object.
  • An environmental object generally represents a secondary or background item or an environmental aspect of the game.
  • the audiovisual program may also be based on the detected position of an environmental object position, but the environmental object primarily influences the environment of the game, which is typically based more on the game's characters and their interactions with the each other and/or their surrounding environment.
  • an environmental object may be represented in the audiovisual segment(s) as an ancillary object in relation to a character object, which is typically the main focus.
  • an environmental object may affect conditions or parameters of the audiovisual segments.
  • the sensor object is generally a physical object that when detected, causes the audiovisual program to generate images and/or sounds within one or more audiovisual segment(s) that are not perceived within the physical game system (e.g., not displayed as part of the base pattern). For example, as described above, visual or auditory clues may be presented through the audiovisual program.
  • the audiovisual segment(s) may include an image or sound that is related to a pattern on the base.
  • the audiovisual segment displays an object that may have been detected by the type of sensing device that is represented by a utilized sensor object. For example, when a sensor object represents a magnifying glass, the audiovisual segment will include a magnified view of a portion of the base that is located under the sensor object.
  • the game system 1200 includes a character object 1202a in the form of a little girl that moves over the base 1201 of the game, which is in the form of the little girl's backyard.
  • the character object can take any suitable form such as a person, an animal, a mythical creature or any kind of animated object.
  • the game system 1200 also includes a sensor object 1204a in the form of a "magic crystal” that is used to reveal objects (i.e., via the audiovisual program) that are "buried" in the backyard 1201.
  • the game system 1200 also includes several environmental objects in the form of a gazebo 1210, a tree 1206a, a hiding place 1208a, a tree stump 1212, and a fountain 1214.
  • Some environmental object's movement may be constrained.
  • the tree stump 1212 is constrained in the x, y, and z positions, but may be rotated to select various game options (e.g., difficulty level). These game options may affect portions of one or more audiovisual segments or globally affect all audiovisual segments.
  • Some environmental objects may be movable and positioned anywhere on the base (e.g., the tree 1206a).
  • the audiovisual program disregards the environmental object's presence (e.g., the object is not represented within the audiovisual segment) unless it is positioned close to the character object 1202a.
  • Figure 13A is a screen shot illustrating a display output 1300 from an audiovisual program based on the character object 1202a, the environmental objects 1206a (the tree) and 1208a (the hiding place), and the sensing object 1204a (magic crystal) of Figure 12 in accordance with one embodiment of the present invention.
  • a character 1202b that corresponds to the physical character object 1202a is displayed.
  • the environmental objects that are positioned within a predetermined distance from the physical character 1202a are also displayed.
  • a tree 1206b and a hiding place 1208b are displayed relative to the character 1202b.
  • an enhanced view 1204b is displayed based on the position of the sensor object 1204a.
  • the sensor object is in the form of a "magic crystal” that is deemed to detect underground objects (as shown, a mole 1204b) that are located underneath the base 1201. In other words, it appears as if the user is using the "magic crystal” 1204a to detect objects beneath the base 1201 by moving the "magic crystal” 1204a across an area of the base.
  • an underground view which includes any underground objects, of the base area is simulated and displayed within the audiovisual segment.
  • object detection within the physical game base 1201 via the sensor object 1204a is simulated by the audiovisual program.
  • the present invention may also include a gesture recognizer for identifying gestures that are made using any of the physical objects.
  • a particular movement sequence by a physical object may be detected and used to generate an appropriate audiovisual segment.
  • the user may initiate a "jump rope" gesture by lifting the character object 1202a off of the base 1201.
  • the character is then represented within an audiovisual segment that includes a turning jump rope.
  • the displayed character's jumping movements correspond to the physical character's jumping gesture.
  • the displayed character's movements are synchronized with the physical character's movements.
  • the audiovisual segment may also include other features that are based on the character object's movement. For example, if the physical character (and displayed character) fails to synchronize their jumps with the displayed jump rope's turning, the displayed character is shown as tripping over the rope or getting hit in the head with the rope.
  • Figure 13B is a screen shot illustrating a display output 1301 from an audiovisual program based on a "jumping rope" gesture performed by character object 1202a of Figure 12 in accordance with one embodiment of the present invention.
  • a "jumping rope" gesture may be initiated by lifting the physical character 1202a off the base 1201.
  • a displayed character 1202c that corresponds to the physical character 1202a is shown as engaging in a jumping rope activity.
  • Two additional displayed characters 1302a and 1302b hold and turn a jump rope 1304, and the main character 1202c jumps or trips over the rope or the rope strikes the upper body of the character 1202c, depending on whether or not the character's jumps are synchronized with the rope's turning movements.
  • Figure 14 is a flowchart illustrating the process 1400 of analyzing a "jumping rope" gesture of the character object of Figure 12 in accordance with one embodiment of the present invention.
  • the physical character object's current position e.g., 1202a of Figure 12
  • An audiovisual program displays the character (e.g., 1202b of Figure 13A) based on the obtained position in operation 1404. It is then determined whether the physical character's (e.g., 1202a) position has changed in operation 1406. This determination step is repeated until the character changes position.
  • the character's movement indicates a "jumping rope" gesture in operation 1410.
  • Any suitable type of movement may indicate a gesture. For example, repeatedly moving the character up and down in the z direction to indicate a "jumping rope” gesture.
  • a new audiovisual segment may then be activated based on the detected gesture.
  • the character e.g., 1202c of Figure 13B
  • the character is then displayed as jumping rope (e.g., see Figure 13B) based on the position of the character.
  • the physical character's position is again repeatedly checked until a movement is detected in operation 1414.
  • the physical character e.g., 1202a
  • a new position e.g., changes its z position
  • the physical character's z position is compared to the displayed jump rope's z position to determine whether the displayed character has cleared the rope (or the rope has cleared the displayed character's head). This determination operation is repeated until the physical character becomes unsynchronized with the displayed jump rope's timing.
  • An appropriate audiovisual segment is then displayed for the physical character's mistiming in operation 1418. For example, the displayed character is shown tripping over the displayed rope.
  • the entertainment system may include any other tool types.
  • the system may include an x-ray tool that may be moved over any other physical object, such as a person or animal figure. As the user moves the x-ray tool over a portion of the figure, the audiovisual program will display an x-ray or inside view of the figure portion.
  • the x-ray tool may also be used to reveal objects that are deemed to be hidden under the game surface (e.g., buried objects).
  • the tool may have a stethoscope function, wherein a heart beat, or other sound, may be heard as the stethoscope moves over the figure.
  • the entertainment system may also include a flashlight tool that may also be used to reveal hidden objects within the physical game system.
  • a flashlight tool that may also be used to reveal hidden objects within the physical game system.
  • an audiovisual segment will include a lighted portion that may include other previously hidden objects of the surface or game base.
  • a tool may function as a metal detector.
  • the described interstitial feature may be applied to other types of physical object movements, besides page turning.
  • an appropriate interstitial audiovisual segment may be generated that shows a helicopter taking off from a first area, flying through the air, and landing at a second area.
  • the interstitial segment may also be based on the amount of time that the helicopter object is in the air. For example, if the helicopter leaves the base for a relatively long amount of time, the interstitial segment may show the helicopter travelling a great distance (e.g., from one country to another).
  • the gesture features may also be implemented within the described book system, as well as the embodiment illustrated in Figures 12 through 13.
  • the book system may be configured to recognize gestures of a particular tool.
  • a digging tool may be used to make a " scratching" movement across a selected page portion.
  • a corresponding audiovisual segment is then generated to reveal previously hidden objects. That is, hidden objects are uncovered via the audiovisual program in response to making a "scratching" gesture over the hidden object.
  • the entertainment systems of the present invention may include mechanisms for providing differently scaled audiovisual segments based on movements of a physical object. That is, as a physical object moves a distance within different regions of the game, different audiovisual segments may be displayed such that each segment represents this distance in different scales. In a specific example, when the physical object moves a distance within a first region of the game, the audiovisual segment represents that distance with a first scale (e.g., a zoomed out view). In contrast, when the physical object moves the same distance within a second region, the audiovisual segment represents that distance with a second scale (e.g., a zoomed in view).
  • a first scale e.g., a zoomed out view
  • a second scale e.g., a zoomed in view

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • User Interface Of Digital Computer (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

L'invention concerne un système de divertissement interactif comprenant une base (104) qui représente un dispositif de divertissement, un objet capteur physique (106) séparé de la base, un détecteur (102) décelant la présence, l'emplacement et l'identité de l'objet capteur par rapport à la base, et un programme audiovisuel interactif (112) conçu pour fournir des segments audiovisuels liés au dispositif de divertissement représenté par la base. La détection de l'objet capteur physique fait que le programme audiovisuel interactif produit un segment audiovisuel fondé au moins en partie sur la position décelée de l'objet capteur par rapport à la base et concernant une information non visible sur la base mais jugée détectable par l'objet capteur.
EP00905848A 1999-02-01 2000-01-31 Systemes et procedes de divertissement interactif Withdrawn EP1159668A2 (fr)

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US11838499P 1999-02-01 1999-02-01
US118384P 1999-02-01
US30664799A 1999-05-06 1999-05-06
US306647 1999-05-06
PCT/US2000/002342 WO2000045250A2 (fr) 1999-02-01 2000-01-31 Systemes et procedes de divertissement interactif

Publications (1)

Publication Number Publication Date
EP1159668A2 true EP1159668A2 (fr) 2001-12-05

Family

ID=26816293

Family Applications (1)

Application Number Title Priority Date Filing Date
EP00905848A Withdrawn EP1159668A2 (fr) 1999-02-01 2000-01-31 Systemes et procedes de divertissement interactif

Country Status (9)

Country Link
EP (1) EP1159668A2 (fr)
JP (1) JP2002535794A (fr)
KR (1) KR20020003364A (fr)
CN (1) CN1342275A (fr)
AU (1) AU2746600A (fr)
CA (1) CA2360940A1 (fr)
NO (1) NO20013633L (fr)
PL (1) PL349814A1 (fr)
WO (1) WO2000045250A2 (fr)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2009018988A2 (fr) * 2007-08-03 2009-02-12 Ident Technology Ag Jouet, en particulier une poupée ou un animal en peluche ou similaire
US8545283B2 (en) 2008-02-20 2013-10-01 Ident Technology Ag Interactive doll or stuffed animal
DE102010004813A1 (de) * 2010-01-15 2011-07-21 Schubert, Peter, 01159 Wiedergabegerät für ein Druckerzeugnis
JP2013538077A (ja) * 2010-08-04 2013-10-10 ピュア・イマジネーション・エルエルシー ゲームピースの静電容量検知と識別とを組み込んだ双方向ゲームボードシステム
TWI512547B (zh) * 2011-12-20 2015-12-11 Univ Nat Chiao Tung 互動式系統及互動式裝置
US10363486B2 (en) 2013-06-10 2019-07-30 Pixel Press Technology, LLC Smart video game board system and methods
JP6974173B2 (ja) * 2015-04-08 2021-12-01 レゴ エー/エス ゲームシステム
JP6979088B2 (ja) * 2017-09-05 2021-12-08 株式会社バンダイ 情報提供システムおよび情報提供プログラム
CN112675555B (zh) * 2018-01-08 2022-09-27 克兹二世怡人合资有限公司 具有电容式触摸交互性的儿童玩具
JP7311205B2 (ja) * 2020-03-18 2023-07-19 Necソリューションイノベータ株式会社 判定装置、判定方法、記録媒体

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2607400A1 (fr) * 1986-11-27 1988-06-03 Chacun Patrick Perfectionnement aux jeux de societe
CA2196048A1 (fr) * 1994-07-28 1996-02-08 Pinhas Gilboa Table de jeu avec ordinateur

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of WO0045250A2 *

Also Published As

Publication number Publication date
NO20013633L (no) 2001-09-12
CA2360940A1 (fr) 2000-08-03
AU2746600A (en) 2000-08-18
KR20020003364A (ko) 2002-01-12
WO2000045250A2 (fr) 2000-08-03
CN1342275A (zh) 2002-03-27
PL349814A1 (en) 2002-09-09
NO20013633D0 (no) 2001-07-24
JP2002535794A (ja) 2002-10-22
WO2000045250A3 (fr) 2000-12-28

Similar Documents

Publication Publication Date Title
KR100395719B1 (ko) 화상처리방법및전자장치
RU2673275C2 (ru) Способ воспроизведения информации, способ ввода/вывода информации, устройство воспроизведения информации, портативное устройство ввода/вывода информации и электронная игрушка, в которой использован точечный растр
EP2591466B1 (fr) Procédé et système permettant d'améliorer la lecture d'un livre
US7123742B2 (en) Print user interface system and its applications
US8057233B2 (en) Manipulable interactive devices
US6167353A (en) Computer method and apparatus for interacting with a physical system
US8376854B2 (en) Around device interaction for controlling an electronic device, for controlling a computer game and for user verification
US20110244937A1 (en) Game execution method, storage medium, and electronic instrument
EP1159668A2 (fr) Systemes et procedes de divertissement interactif
KR101263686B1 (ko) 증강 현실을 이용한 노래방 시스템 및 장치, 이의 노래방 서비스 방법
CN105723306B (zh) 改变标记在物体上的用户界面元素的状态的***和方法
JP2001212369A (ja) タイプ練習機能を有するアミューズメントシステム、タイプ練習システムおよびコンピュータ読み取り可能な記憶媒体
KR20150041249A (ko) 가상 체험북 시스템과 그 동작 방법
CN109448453A (zh) 基于图像识别追踪技术的点读问答方法与***
US20050153765A1 (en) Electronic memory game
JP2013123583A (ja) ゲーム装置
WO2015113440A1 (fr) Système et procédé permettant de changer l'état d'un élément d'interface utilisateur marqué sur des objets physiques
KR101267247B1 (ko) 증강 현실을 이용한 노래방 시스템 및 장치, 이의 노래방 서비스 방법
TWM330533U (en) Language learning board game
JP3815938B2 (ja) ゲーム装置および情報記憶媒体
JP7160707B2 (ja) 仮想キャラクタを含む画像を提供するための方法
CN105637465A (zh) 用物体操作计算机程序的***和方法
US20080032275A1 (en) Interactive system
CN1320418C (zh) 用于呈现交互内容的***
JP2021053466A (ja) ゲームプログラム、ゲームプログラムを実行する方法、および情報処理装置

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20010831

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

AX Request for extension of the european patent

Free format text: AL;LT;LV;MK;RO;SI

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: INTERLEGO AG

Owner name: LEGO A/S

17Q First examination report despatched

Effective date: 20011211

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20020622

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: INTERLEGO AG

Owner name: LEGO A/S