EP3449340A1 - Vorrichtung und verfahren zum teilen der immersion in eine virtuelle umgebung - Google Patents

Vorrichtung und verfahren zum teilen der immersion in eine virtuelle umgebung

Info

Publication number
EP3449340A1
EP3449340A1 EP17725317.6A EP17725317A EP3449340A1 EP 3449340 A1 EP3449340 A1 EP 3449340A1 EP 17725317 A EP17725317 A EP 17725317A EP 3449340 A1 EP3449340 A1 EP 3449340A1
Authority
EP
European Patent Office
Prior art keywords
target
immersive
display
source
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
EP17725317.6A
Other languages
English (en)
French (fr)
Inventor
Jean-Baptiste DE LA RIVIERE
Valentin LOGEAIS
Cédric KERVEGANT
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Immersion Corp
Original Assignee
Immersion Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from FR1653758A external-priority patent/FR3050834B1/fr
Priority claimed from FR1659160A external-priority patent/FR3056770B1/fr
Priority claimed from FR1659768A external-priority patent/FR3057430B1/fr
Application filed by Immersion Corp filed Critical Immersion Corp
Publication of EP3449340A1 publication Critical patent/EP3449340A1/de
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/017Head mounted
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/0138Head-up displays characterised by optical features comprising image capture systems, e.g. camera
    • GPHYSICS
    • G02OPTICS
    • G02BOPTICAL ELEMENTS, SYSTEMS OR APPARATUS
    • G02B27/00Optical systems or apparatus not provided for by any of the groups G02B1/00 - G02B26/00, G02B30/00
    • G02B27/01Head-up displays
    • G02B27/0101Head-up displays characterised by optical features
    • G02B2027/014Head-up displays characterised by optical features comprising information/image processing systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2219/00Indexing scheme for manipulating 3D models or images for computer graphics
    • G06T2219/024Multi-user, collaborative environment

Definitions

  • the present invention belongs to the field of virtual reality and the presentation of virtual environments and the sharing of virtual environments.
  • the invention relates to a device and a method for sharing the immersion in a virtual environment with users having means of immersion other than the means having generated the virtual environment.
  • one or more users of an immersion system in a virtual environment are immersed in the virtual environment by means of a material environment producing stimuli to give users sensations that are close, if not identical, if not supposed, of those that would be felt in a real environment that the virtual environment is supposed to reproduce.
  • the representation is not necessarily realistic but can also be distorted, attenuated or exaggerated in certain effects, to produce particular sensations such as increased perception of phenomena in the virtual world.
  • a scale effect a user can be immersed in a microscopic universe that is normally inaccessible to him.
  • a user can be immersed to observe a non-visible environment in ordinary conditions such as aerodynamic flow, or to observe a representation of a non-physical environment such as a database visualized in 3D.
  • the same virtual environment is shared among multiple users. Subsequently, it will be considered a visual representation of the virtual environment, two or three dimensions, and it will be more particularly the problems and solutions related to this visual representation in the context of the invention.
  • a known solution to allow at least two users to be immersed in a virtual environment is to duplicate the data used to build the visual representation of the virtual environment on two immersive systems, each system being assigned to a user, to obtain on each of the two systems rendering real time corresponding to the point of view of each user associated with a system.
  • the information relating to the position and or the interactions made by each of the users are synchronized to allow each of the virtual environments reproduced on each of the systems to be in an identical state.
  • This solution has the advantage of allowing each of the users sharing the immersion to have their own point of view on the shared virtual environment.
  • a user immersed by means of a first immersive system can look at the left side of an object and a second user immersed in the same virtual environment by means of a second immersive system can simultaneously look at the right side of the same object.
  • each of the immersive systems must have the data used to build the virtual environment, and its visual representation, and of all the applications, in particular software, making it possible to exploit these data, conditions which impose go through a step of duplication of data and have licenses, which can be expensive, on the software implemented on each of the systems.
  • Another disadvantage of this solution is to require on each immersive system digital processing systems with sufficient performance to generate in real time the images of the virtual environment displayed on the display means of the immersive system, even greater performance when the immersive system must operate a three-dimensional environment and or produce a three-dimensional representation.
  • Another disadvantage of this solution is that the comfort of sharing the experience of the virtual world very much depends on the quality of the network communication means set up. Because sharing is based on data synchronization between multiple immersive systems, the latency times introduced by the network communication means affect the speed at which data from a remote immersive system is updated in a given immersive system, causing potentially significant difference between the actual state of the virtual environment at a given instant of the immersive system remote and that retranscribed within the immersive system considered at the same time.
  • images produced by an immersive system for a user are filmed by a camera carried by the user, oriented to show the images in the user's field of view, and are transmitted to be observed on a screen by people more or less distant.
  • this method does not offer the quality expected by a remote person to observe a virtual environment because of the exploitation of the image filmed on a single screen and the fact that the filmed image generally includes defects and artifacts in because of the implementation by the immersive system of image display systems comprising several screens assembled to form the immersive environment.
  • people who observe the image remotely are passive and have no interaction with said images.
  • such a solution is not conceivable when the immersive system implements a virtual reality headset, the images being in this case displayed very close to the eyes of the user.
  • the skilled person knows that the images are composed to be viewed from a given location of the display system and that viewed from another location the images are seen distorted and with connection faults when it is implemented multiple screens.
  • the invention provides a solution to the various problems of the prior art in the field of immersion of an observer in a virtual environment generated for another system.
  • One advantage is to avoid increasing the costs of software licenses. Another advantage is to avoid duplication of data and data transmission constraints.
  • Another advantage is not to be limited in the choice of hardware architectures of each of the immersive systems implemented.
  • Another advantage is to preserve the confidentiality of the data of the virtual environment.
  • the invention relates to an immersion sharing device in a virtual environment, comprising:
  • a source system delivering operating data of a visual representation of the virtual environment, said operating data comprising display data relating to images of said visual representation and including data of the observation conditions in which the visual representation of the virtual environment has been generated;
  • At least one target immersive system comprising a digital processing system delivering images to at least one display system of said target immersive system.
  • the device comprises:
  • means for transforming the display data, representative of images of the visual representation of the virtual environment delivered by the source system associated with data of the observation conditions, for constructing images displayed in the system display (22) of the target immersive system (20), such as the set of images displayed by said display system of said target immersive system, is the result of processing to promote immersion on said target immersive system (20) in the display data during a transition from the viewing conditions transmitted by the source system (10) to actual viewing conditions in the display system (22) of said target immersive system.
  • promoting immersion includes compensating for any visual deformations resulting from the use of display data on different display systems and or under different observation conditions.
  • Encourage immersion may also include changing the position in the virtual environment of the target system of the content portion of the virtual environment visible through the system display system source to present it in a stabilized manner, mitigating the impact of frequent head movements of the user of the source system.
  • the sharing can be done in real time, the transmission of the data then being carried out continuously, or in deferred time.
  • the target immersive system by realizing by the target immersive system the transformation of the display data for the function part of the actual observation conditions, including the observation position data and the observation direction data, in the display system. of the target immersive system, it minimizes the impact of the latency between the display in the target system of the transformed images and the taking into account for the transformation of the display data of the actual observation point in the system of display of the target system.
  • the source system is an immersive system having a source display system displaying images of the display data of a visual representation of a virtual environment generated by said source system.
  • a system is then obtained in which the operating data of the source system is generated with the interaction of a user of the source immersive system, the immersion being able to be shared in real time or in deferred time with an observer of the target system. .
  • the transformation of the display data includes a reconstruction of a 3D virtual representation of a source display system of the physical environment of the source display system, a position of a point of view, observing the observer in the display system of the target system being forced, for the transformation of the display data, to the same position in said 3D virtual representation of the source display system as the observation position of the user in the source display system.
  • the source immersive system includes a real-time measurement system of the viewing direction, and / or the viewing position, in the display system of the source immersive system, of a user who would be immersed in the virtual environment of that system immersive source, and the at least one target immersive system comprises a system for measuring the observation direction, respectively and or the observation position, in the display system of said target immersive system, of an observer who would be immersed in the virtual environment represented on said target immersive system.
  • Partial processing of the display data may be performed on the digital processing means of the source immersive system, or on an ancillary computer connected to a network through which data is transmitted between the source system and at least one target immersive system.
  • said software transforming the display data generated by the source immersive system, representative of the images of the visual representation of the virtual environment of the source system, into display data corresponding to non-dimensional images independent of a structure of a viewing system to display the images, said display data corresponding to dimensionless images being transmitted to the at least one target immersive system.
  • the transformation calculations of the display data that do not depend on the target immersive system are advantageously performed on the source system or an independent annex computer.
  • Such a solution makes it possible, in a device comprising several target immersive systems, to transmit to the different target immersive systems display data in a neutral, dimensionless form, of which only the subsequent processing, specific to the target immersive system, is performed by each of the immersive systems. targets.
  • the dimensionless images correspond to images projected on an inner wall of a sphere, in the center of which sphere is placed the point of observation of the source system, to form the dimensionless images in a solid angle corresponding to the images displayed on the visualization system of the target immersive system and up to four Pi steradians.
  • symbolic representations are advantageously superimposed on the visual representation of the virtual environment represented on the target immersive system to provide the observer with guidance and selection of an observation direction.
  • symbolic representations comprise for example the graphical materialization of a sphere or a grid associated with the representation of horizontal surfaces, for example a floor, and or vertical, for example a wall.
  • Such symbolic representations which can only be materialized temporarily, for example by a command from the observer, or under certain conditions, for example a rapid evolution of the observation conditions of the source system, provide assistance to the observer to facilitate his choice of direction of observation and to identify or find his position in the virtual scene.
  • the display data processing software for constructing images displayed on the display system screen (s) of the at least one target immersive system may also be executed on the calculation means of said target immersive system.
  • the device includes a plurality of target immersive systems simultaneously connected to the source immersive system. It is thus possible with the benefits of the invention to allow many observers, who can be remote and dispersed, to follow the immersion experienced by a user of the source system.
  • a target immersive system may be located at any distance from the source system, the synchronization problems of the visual representation displayed in the target immersive system with the observational conditions of the observer finding precisely a solution in the system. 'invention.
  • the source system includes a digital or analog signal capturing equipment, carrying the display data, generated by the source system and transmitting as digital data said display data corresponding to the captured signal. , directly or indirectly, to at least one target immersive system.
  • the source immersive system includes software, executed on a digital processing system, capturing contents of the display data of the source system, after the display data has been computed, and transmitting under digital data form said contents of the display data captured, directly or indirectly, to the at least one target immersive system.
  • such contents advantageously correspond to a set of data sufficient to reconstruct the images generated by the source system.
  • Said data are, for example, the digital data sent to a display system of the source system, where appropriate before an analog conversion, or data subject to spatial and or temporal compression algorithms to reduce its size, or sets of attributes of each image: outlines, textures, colors ...
  • the contents are for example images, or video streams, in yet another form of stereo images, in another example form a stereo video stream.
  • a form of the contents is determined according to the calculation means of the source system and / or of the target immersive systems, and transmission means of data, to limit as much as possible the delays between the processing of the operating data by the source system and the target immersive system (s) taking into account the performance of the equipment implemented by the entire chain of data transmission and their treatments by the systems.
  • the device includes a data logging system configured to record operating data generated by the source system and to transmit said operating data to at least one immersive target system at a later time.
  • the source system includes an image processing software and a source display system
  • a target immersive system includes a computer graphics calculation software and a visual representation of a virtual environment in the form of images displayed on one or more screens of the display system of said target immersive system and also comprises a display data capture equipment and software corresponding to the images displayed on the display system said target immersive system, so that each of said target and source immersive systems can alternatively be a source system and a target system.
  • the device is in this embodiment particularly suitable for collaborative work between a user and an observer likely to reverse their roles.
  • the two target and source immersive systems that can invert their conditions can be different both in their Hardware configurations that software, the device according to the invention to overcome these points of compatibility constraints usually encountered in known systems.
  • This configuration is also compatible with the simultaneous connection of other target immersive systems.
  • the display system and the processing system of the target immersive system and or the source system are associated with interaction means configured to modify, via the display system and / or the data processing system.
  • immersive system considered, the content and or the behavior of the images displayed by the display system of the source system or the target immersive system considered.
  • each of the users of the source system and observers of the target immersive system has means of intervention to act on the behavior of the displays without destroying their coherence.
  • the interaction means includes a gel control for selectively enabling, and disabling, the freezing of all or part of the display data and / or data of the viewing conditions used to calculate the images to be displayed by the display system of the source system and / or the target immersive system.
  • the interaction means of the target immersive system includes a freeze control of a subset of the viewing conditions in the display system of the source system and or the target immersive system.
  • the interaction means of the at least one target immersive system are further configured to point and or annotate the images displayed by the display system of said target immersive system.
  • annotations are for example materialized in the form of incrustations in the images.
  • data characterizing the scores or annotations formed on images of a target immersive system are transmitted via the transmission means to at least one other connected immersive system. It is thus possible for this other system, the source system or another target system, receiving this data to process this data for display.
  • the interaction means comprise for example one or more sensitive surfaces to the touch or the presence of a hand or a finger, and or one or pointers.
  • the possible display system of the source system, and the display system of the at least one immersive target system each belong to one of the categories among implementing flat screens and or curved screens:
  • At least one target immersive system is located in a remote location of the source system, the physical separation of the two systems being such that the user of the source system and the observer of the target immersive system are not able to to communicate without using technical means of communication. It is in this configuration pushed back the spatial limits that allow a sharing of an immersion.
  • at least one target immersive system is located near the source system, the physical separation being such that a user of the source system and an observer of the target immersive system can communicate directly without a physical barrier; ie without the need to implement technical means. It is in this configuration obtained to share an immersion, for example in the context of a collaborative work, without suffering unavoidable artifacts when an observer is close to the user in the hope of feeling the effects.
  • Such a configuration is not merely hypothetical in that it allows, even in the same place, to simulate immersion, with simple hardware, for one or more observers, in the user's virtual environment. of the source system, on the one hand without disturbing said user and on the other hand under optimal conditions since it is impossible for an observer to place himself in the display system of the source system with the same observation conditions as those of the user.
  • the invention also relates to a method of sharing immersion in a virtual environment of an observer in a target immersive system, comprising the steps of:
  • operating data of a visual representation of the virtual environment comprising display data relating to images of said visual representation and including data of conditions. observation means associated with said display data;
  • the method comprises, for each image or set of images to be displayed, between the generation step and the display step, a transformation step, carried out at least partly on the target immersive system. , display data of the virtual environment generated in the generating step, the transforming step comprising determining conditions of effective observation by the operator in the display system of the target immersive system and transform the display data associated with the observation conditions of the step of generating the display data, into display data corresponding to the actual observation conditions by the observer.
  • each image point of a transformed image, at the transformation step to be displayed by the display system of the target immersive system is determined according to at least one observation point. effective in the display system of the target immersive system for maintaining, at a constant proportionality factor for the entire image at a given instant, the same relative direction of observation with respect to any other image point of said transformed image , observed from the actual observation point, that the relative direction between said image points in the display data of the visual representation of the virtual environment generated at the generation stage delivered by the source system and including the image data and the observation condition data associated with said image data.
  • the step of generating display data of the representation of the virtual environment is performed on a source system, which source system is: an image generation system for a virtual display; or a system for broadcasting a flow of images, real and or virtual, generated in real time or recorded; or an immersive system in which a user is immersed.
  • a source system is: an image generation system for a virtual display; or a system for broadcasting a flow of images, real and or virtual, generated in real time or recorded; or an immersive system in which a user is immersed.
  • the reconstitution of a 3D virtual representation of the source display system of the physical environment of the source display system is performed by the source immersive system or the ancillary computer connected to the network.
  • the transforming the display data comprises reconstituting a 3D virtual representation of the source display system, an observer's observation position being forced, for the transformation of the display data, to a same position in said representation virtual 3D display system source that the user's observation position in the source display system.
  • a virtual source display system is created whose position relative to the observer is modeled on the position of the user in the actual source display system.
  • the display system of the source immersive system virtual reconstruction which is displayed on the display system of the target immersive system so that the observation conditions of the observer of said virtual reconstruction of the source system in the target immersive system at least partially correspond to the user's observation conditions of the source immersive system in the source immersive system.
  • This transformation eliminates duplication of data in the virtual environment and can be applied for any form of target immersive system and any software running on the source immersive system.
  • the step of transforming the display data is performed completely on the target immersive system after a step of transmitting the viewing conditions and display data of the virtual environment, generated by the digital processing system of the source immersive system.
  • the step of transforming the display data is performed partly on the source immersive system, or on an ancillary computer connected to a network of the data transmission means, before a transmission step.
  • the display data partially transformed by the source immersive system, or the auxiliary computer are independent of the characteristics and operating data of a target system, thereby making it possible to factorize the processing step or even the reduce to simple processing for display on some target systems.
  • the transformation step includes a display data freezing step of representing the virtual environment or freezing observation conditions, generated in the generating step.
  • the data of the observation conditions in the display system of the target immersive system continue to be taken into account during the display data freezing step or the freezing of the display conditions. 'observation.
  • the observer has the possibility to freeze, at least in part, the representation of the environment on the target system, and momentarily obtain a stable image on which it can for example provide annotations
  • the method includes an initialization step in which the target immersive system initializes a register including data on conditions in which the display data is generated in the generating step.
  • the step of transforming the display data is performed partly on the source immersive system, or on an ancillary computer connected to a network of the data transmission means, before a transmission step.
  • Figure 1 a simplified example of an immersion sharing device in a visual environment with a source immersive system comprising a multi-screen type display system and a target immersive system type flat mono-screen;
  • Figure 2 An illustrated example of image transformation principles, generated in a source immersive system with a display system and observed in a target immersive system with a different display system, to maintain visual immersion conditions ;
  • Figure 3 an illustration of a device comprising several target immersive systems with display systems of different technologies reproducing the virtual environment of the same source immersive system
  • Figure 4 a simplified block diagram of the method according to the invention.
  • a visual immersive system generally comprises a set of hardware and software means for at least: generating images of a visual representation of a virtual environment; display images for at least one user; the displayed images being calculated to represent the virtual environment as it should be perceived by the user according to his observation conditions of the virtual environment.
  • Display system set of hardware means and, if necessary, software comprising means for displaying images, for example one or more display screens arranged between them in space to present images to a user or to an observer , where appropriate using conventional optical systems such as mirrors, semi-transparent plates, collimation optics, filters, etc.
  • the display system is associated, if necessary, with means for determining in real time. , in a frame linked to the screens, the position and or direction of observation, of a user or an observer looking at the images displayed by said screens.
  • Helmet a particular case of a display system in which the screens are fixed on a support carried on the head by the operator.
  • the observation position is fixed in the frame of reference related to the screens and the direction of observation is a function of the movements of the eyes.
  • This type of equipment can be associated with means for real-time monitoring of the orientation, or even the position, of the head in the physical environment, which will not of course change the position of the head of the operator by report to the display system but may be exploited by the calculation software of the representation of the virtual environment to modify the position and the direction of observation in a repository of the virtual environment.
  • Image any form of visual representation displayed on a screen.
  • An image can be fixed, animated resulting from a succession of images in time, resulting from a video stream ...
  • An image can be transmitted for display on a screen by an analog signal or by a set of digital data representing all or part (for example in the case of compressed signals) of an image, or both according to a location considered in the chain of transmission of the image between the generation of the image and its display.
  • Virtual environment an environment defining a set of numerical data defining the characteristics of this environment and necessary for the construction of a visual representation of this environment as it must be perceived.
  • the virtual environment can be the result of a 2D or 3D construction from a digital model.
  • the virtual environment can also be the result of virtualization of a real environment, for example by digitized images or video streams.
  • the virtual environment can also be a combination of a digital model and a real environment following a principle of augmented reality.
  • Immersion A situation in which representations of a virtual environment are presented to one or more people in such a way as to give those people the impression of being physically surrounded by the information contained in the virtual environment.
  • the display system may include a system of reconstructions of the effects of the relief, for example occulting glasses synchronized with the displayed images to present to the user of the immersive system images corresponding to each of his eyes. to produce a stereoscopic effect.
  • the rendered virtual environment can be two-dimensional, 2D, or three-dimensional, 3D.
  • the digital information used by the immersive system can correspond, and often corresponds to, data from a 3D immersive environment, in which case a software application of the immersive system is a synthetic image creation application that elaborates each instant one or more new computer images by digital processing of the 3D information as a function of a position of a point of observation, for example corresponding to an optical center of an objective of a virtual camera.
  • a software application of the immersive system is a synthetic image creation application that elaborates each instant one or more new computer images by digital processing of the 3D information as a function of a position of a point of observation, for example corresponding to an optical center of an objective of a virtual camera.
  • the digital information may also be data of a two-dimensional, 2D environment, such as films, possibly pre-recorded panoramas, which can be digitally processed by a software application to display images viewed from an observation point controlled by the user or controlled by more or less interactive logic.
  • a software application to display images viewed from an observation point controlled by the user or controlled by more or less interactive logic.
  • a single screen, computer or video monitor or a computer telephone can form the display portion of a display system of an immersive system.
  • the exploitation data comprises all the information of a representation of the virtual environment seen in a viewing system by an observer, a situation that can be real or virtual.
  • the operating data includes the data of the observation conditions and the display data.
  • the display data notably comprise information relating to the images generated by a source system 10.
  • the display data comprise attributes of each of the image points of the images generated by the source system 12, these attributes comprising in particular, in addition to luminance and color characteristics, the position of the image point in an image, for example an image. image to be displayed in said display system.
  • the display data is for example information contained in video frames.
  • the data of the observation conditions include data relating to the position from which and to the direction in which a user or an observer looks in an image display system, or with which images have been generated for example by a device for shooting a real environment then virtualized.
  • FIG. 1 illustrates, as an exemplary embodiment, a device 100 comprising two immersive systems, a source immersive system 10 and a target immersive system 20, connected by data transmission means.
  • the source immersive system 10 of FIG. 1 is a known system which comprises, in the illustrated example, a digital processing system 11, comprising a permanent or temporary database 11a for storing digital information defining an environment. virtual, and a display system 12 images generated by the digital processing system.
  • the digital processing system 11 is in practice a computer, of power adapted to real-time image calculations, on which computer graphics calculation software is executed.
  • the database 11 has, in practice a memory or a file storage system, contains a set of data defining the virtual environment and necessary for calculating a visual representation of said virtual environment.
  • the database also includes data relating to the source immersive system itself, in particular parameters relating to the display system 12, including dimensions, positions and screen shapes (12a, 12b, 12c, 12d), on which are displayed the images to form a representation of the virtual environment in which a user 90 is immersed.
  • the calculating means 11 also receive variable values, generally depending on a user 90 in the immersion condition. and or an operator of the immersive system, which are necessary for software for the calculation of synthetic images.
  • the variables here comprise an observation position and an observation direction of the user 90 in a repository of the display system 12, position and direction of observation on which the images representing the virtual environment to be displayed on these depend. screens for presenting, viewed by the user, a correct representation of said virtual environment from a position in said virtual environment where the user is supposed to be.
  • the position and direction of observation to be known for the calculation of each new image displayed are, in a manner known to those skilled in the art, measured or estimated in real time or presupposed.
  • variables are for example orders sent to the digital processing system 11, for example by the user 90 via control members, to act directly or indirectly on the virtual environment.
  • the variables are cyclically processed by the software application that synthesizes the virtual environment to present in real time, that is to say with a latency and a cooling period adapted to the immersion. desired, a point of view in the virtual environment corresponding to a point of view of the user 90 in the display system of the physical immersive system.
  • the display means 12 of the source immersive system 10 of FIG. 1 comprise a plurality of screens 12a, 12b, 12c, 12d forming a multi-screen system with four screens, in the example illustrated, frequently called CAVE in the field.
  • virtual reality determining a volume in which the user 90 can stand and move.
  • an arrangement of two vertical L-shaped screens promotes vertical immersion.
  • an arrangement with three, four or five screens with angles more open than 90 ° favors the immersion of several users simultaneously.
  • Some screens can also be curved, cylindrical or spherical, and avoid or limit the number of juxtapositions of screens forming angles.
  • the screens can also be worn by the user by means of equipment such as viewing headsets, which can be more or less transparent depending on the result sought for the perception of the surrounding physical environment.
  • the source immersive system 10 belongs, in the context of the invention, to a device 100 also comprising at least one target immersive system 20.
  • the target immersive system 20 comprises at least one digital processing system 21 and a display system 22.
  • the target immersive system may be similar, or even identical, in structure and means to or different from the source immersive system 10.
  • the display system 22 of the target immersive system 20 may be identical to the display system 12 of the source immersive system 10 or on the contrary be different, for example by a number of screens implemented, and / or by a form of the screens, and or by the size of the screens, and or by an arrangement of the screen or screens, and or by the fact that the screen or screens are worn by the user or not, and or by the capacity or not of to render images for stereoscopic observation.
  • the display system 22 of the target immersive system has only one screen 22a while the display system 12 of the source immersive system has four.
  • the screen 22a is for example a projection screen, or a conventional video screen, television or computer screen, or even a display screen. computer phone.
  • the digital processing system 21 of the target immersive device 20 mainly comprises a memory, or local database, in which are stored characteristics of the source immersive system and the target immersive system used for the transformation of display data corresponding to the images, and a digital computer on which a display data processing software is executed, the principles of which are explained later in the description.
  • source and target will be used to designate the elements respectively of the source immersive system 10 and the target immersive system 20.
  • source immersive system 10 may be designated “Source system” and the target immersive system designated “target system”.
  • the device 100 also comprises data transmission means 30 between the source system 10 and the target system 20.
  • Such data transmission means 30 consist of any means known in the field of data transmission between computers, which in the case of the invention will be selected according to the data rate which should be ensured, the distance between the immersive systems of the device, the number of connected systems and the resources available in communication means.
  • the data transmission is carried out by a communication network 31 such as the Internet network and the implementation of the interfaces and communication protocols adapted to this network.
  • Data transmission can also implement optical links, radio links, and wired links of different types.
  • the source system 10 when the source system 10 generates the display data of a virtual environment whose images are displayed on the source display system 12, virtual environment in which the User 90 is immersed, and said source system transmits to the target system 20 data relating to the display data generated by the source digital processing system 11.
  • the device 100 may comprise any number of target systems, all the target systems being independent of each other, which may be different in their structures, and operate in a similar way with the data received from the source system.
  • the image calculation software executed on the source digital processing system 11 generates a representation of the virtual environment in which the user 90 of the source system is immersed and produces display data of said representation of the virtual environment for each of the different screens 12a, 12b, 12c, 12d of the source display system 12.
  • This display data is generated according to the parameters that can modify the visual perception of the virtual environment by the user 90, in particular a position in the space of a point from which the user 90 sees each screens 12a, 12b, 12c, 12d of the source display system 12 and an observation direction in which the user is looking, the observation direction corresponding to an optical axis of the means observation, in practice eyes.
  • This position and this direction which are associated with it are designated globally thereafter "point of observation”.
  • the observation point may be unique and for example correspond to the user 90 at an intermediate position between his two eyes.
  • the observation point can also be double, when a perception of the relief is sought by a binocular vision, and then the position of each of the two eyes of the user 90 determines a point of observation.
  • the digital processing system associated with the display system necessarily has data relating to the characteristics and the data. arrangement of the different screens of the display system.
  • the calculated images are displayed on the corresponding screens.
  • the operating data generated by the source system is transmitted, via the data transmission means 30, to the target system.
  • said operating data can be transmitted in any form of data according to a protocol understood by the target system and which allows the reconstruction of the display data and the associated observation conditions.
  • the operating data generated by the source system can not directly produce images displayed correctly on the target display system 22, in particular because of differences in the structure of the target and source display systems, and observation, assigned respectively to the user 90 and the observer 91, in each of the source and target systems that are not correlated.
  • target display 22 is maintained at all times identical to the observation point of the user 90 in a repository of the source display system 12, and assuming that display data generated by the source system is displayed directly under the form of images on system screens target display, they would give an observer 91 of the target system a vision of the virtual environment distorted or unusable.
  • each image through the operating data received from the source system by the target system is transformed. by a display data processing software according to the observation conditions in the source system of the target system.
  • the image processing software is preferably operated, at least in part, by the target digital processing system 21.
  • part of the image processing can also be performed by the source system, or by any system having the ability to communicate with the source and target systems to receive and transmit the necessary data.
  • the transformation of the display data consists mainly of virtually reconstructing the physical environment of the source display system 12 and forcing the observation point of the observer 91 to the same position in the virtual source environment as the point of view. observation of the user 90 in the source display system 12.
  • a virtual 3D representation of the source display system 12 ' is created in the target system 20, as a function of the parameters of the source display system 12, on virtual screens 12' a, 12'd, 12'c of which virtual representation of the source display system are displayed the images of the received display data corresponding to each of the screens 12a, 12d, 13c real source system, see Figure 2. It is thus formed in the target system a 3D virtual environment representing at any time, at the latency period, the source display system 12 with its visual representation of the virtual environment of the source system.
  • the target system 10 then digitally processes the virtual 3D representation of the source display system 12 'to construct a visual representation thereof on the target display system 22 based on the observation conditions of the observer 91 of the target system, in which visual representation on the target display system, the actual, measured or presupposed position and orientation of observation in the target display system are also the position and orientation of observation in the 3D virtual representation of the display system source and which correspond to the actual and measured positions or presuppositions of observation of the user in the real source visualization system, as illustrated in FIG. 2, where a center 40 of the projections corresponds both to the observation point in the 3D virtual representation of the source system and at the actual observation point in the target system e 22.
  • the digital processing corresponds to 3D projections of the images of the 3D virtual representation of the source display system 12 'on the screen (s) of the target display system 22, said projections having as their center of projection the position and observation orientation in the 3D virtual representation of the observer 91.
  • the target system 20 thus presents the observer 91 with a virtual source system in that it reconstructs at any moment the behavior, for the displayed images of the virtual environment, of the source system 10 and that it ensures that, whatever the position and orientation of the observer 91, said observer has a position and an orientation in the virtual 3D representation of the source display system 12 ', identical to that of the user 90 in the source system 10, although that the source and target systems may have different configurations and that the user 90 and the observer 91 have observation conditions that, in practice, evolve independently in their respective immersive systems. For example, for calculating a coherent representation of the virtual environment on the target immersive system, the observation points of the user 90 of the source system and the observer 91 of the target system are superimposed in the 3D virtual representation.
  • each image point 121 of the actual 12 or virtual source display system 12 'in a given direction 42 from the observation point 40 corresponds to an image point 221 of the target display system 22 in the same direction and the image point 221 in the target system then resumes the characteristics (color, brightness ...) of the image point 121 of the source system.
  • This condition is of course valid only when the two image points visually superimposed in the two immersive systems exist.
  • an image point of the source system has no projection equivalent in the target system, it can not be displayed in said target system. Conversely, if an image point of the target system has no projection equivalent in the source system, the image point in said target system will correspond to a "blank" zone in the target system display.
  • the images are displayed by the target system 20 on the screen or screens of the display system of said target system.
  • the position and orientation of the observer 91 of the target system with respect to the target display system 22 are parameters taken into account to effect the transformation of the display data. This consideration is necessary so that the representation of the virtual environment is coherent with the point of view that generated it on the source system and that it is reproduced on the target system the desired immersive effect.
  • the superposition condition of the observation positions is necessary while the superposition condition of the observation directions is optional, and the latter condition is justified only when 'he is sought that the visual representation seen by the observer 91 in the target system 20 is that of the user 90 in the source system, including direction.
  • the observer 91 in the target system has, within the solid angle of the space in which an image is displayed, the possibility of viewing the representation of the virtual environment in a direction different from that in which the user 90 looks at this virtual environment in the source system, but which then corresponds to the visual representation that said user would have in the source system if he looked in the same direction as the observer in the target system.
  • the position of the observation point and the observation orientation of the observer 91 are determined, for example measured in real time and transmitted to the target digital processing system 21, to be taken into account in the data transformation calculations. display. Where appropriate, the position and or direction of observation are simply estimated or presupposed. For example, when several observers are looking at the screen (s) in the target system's display system, average values of the observing conditions of the different observers can be calculated or assumed, being understood as a compromise producing necessarily distorted images displayed by the target system that may be perceived by some observers.
  • a target system 20 which allows an observer 91 to be immersed in a virtual environment created by a source system 10, which may be remote.
  • This result is obtained without transmission of data from the database 11 implemented by the source system, including with a target system 20 whose configuration, in particular of the display system, is different. that of the source system 10. It allows, among other things, to implement in the target system 20 a digital processing system 21 of performance lower than that necessary for the source system 10 to generate the virtual environment. It does not require the implementation of expensive software for calculating computer generated images. It does not require the compatibility of the target and source system display systems, nor does the observer 91 be required to maintain imposed viewing conditions in the target display system 22 to view the images as the user 90 perceives them in the source display system 12.
  • the embodiment of the device 100 described is cited by way of example.
  • the hardware architectures of the source system and the target immersive system (s) may be different from those schematized in the description for performing the same functions.
  • the transformation of the transmitted display data can be carried out by any transformation algorithm that can make the necessary projections, the method described above being only a simplified theoretical illustration of the transformation by spherical projections.
  • the transformation of the display data is advantageously carried out by the target system which knows the definition of its display system 22 and which takes locally, in real time, with a minimum latency, the other parameters which are specific to it, such as the conditions of the display. observation of the observer.
  • the target system must also have the definition of the conditions under which the source system generates the display data, parameters of the display system in particular, for example transmitted at the beginning of connection with a set of data invariable from the data sources. device operating data of the source and target systems, as well as the user's viewing position and orientation 90 in the source display system 12.
  • the transformation of the display data can be carried out in part by the source system, or by another system, provided that the target system transmits the operating data required by the system. processing the display data to produce the images to be displayed on the display system of said target system.
  • the transformations applied to the display data requiring the use of variables specific to the target system 20, in particular the position and direction of observation of the user 91 of said target system are carried out on the system digital processing 21 of said target system, which makes it possible to synchronize the images displayed by the target display system 22 with said variables specific to the target system.
  • the source system 10, or another ancillary calculator performs a first transformation of the operating data into dimensionless operating data, independent of the structure of a display system to display the images, for example assigning the characteristics of each image point associated with a direction of said determined image point in spherical coordinates in an axis system attached to the observation point in the display system 12 of said source system.
  • the target system 20 uses the received non-dimensional operating data to produce display data adapted to its display system 22 by assigning to each point of a screen in a given direction observation conditions in the display system. target display 22 the characteristics assigned to the image point corresponding to the same direction in the dimensionless display data.
  • the target immersive system enables the observer (91) to correctly display a representation of a virtual environment substantially simultaneously with the generation of said representation by another system. immersive.
  • a source system delivering the representation of the virtual environment is not necessarily associated with display means.
  • a source system is for example a record of the display data produced on a system for generating synthetic images, which may be an immersive system or not.
  • a source system is for example a system delivering images of a real world, recorded or real-time, virtualized by said system for transferring display data.
  • Such a source system may also be a system combining the capabilities of producing synthetic images combined with images of a real world according to an augmented reality logic.
  • each eye sees a different image corresponding to different viewing conditions which can lead to transformations of the display data for each of the images. viewing conditions, at least when both source and target systems are configured to display images of the virtual environment with a relief effect.
  • the display data comprise virtual or virtualized images, representations of the virtual environment seen in an observation window of the source system, equivalent to the display system of the source immersive system of the embodiment described. supra.
  • the display data received from the source system by the target immersive system is interpreted to reconstruct, in the display system of the target immersive system, a representation of the environment viewed in the source observation window.
  • the observer is presented to the observer in the immersive system a stable representation of the representation of the environment seen by the source system by ensuring the coherence of the images displayed in the display system of said target immersive system with those of the window of source observation by which this environment is or has been generated and or seen.
  • the target digital processing system 21 has display data synthesis capabilities adapted to its own display system, which can advantageously be exploited to facilitate the understanding and / or reinforcement of the observer of the target system.
  • the target immersive system comprises, where appropriate, all or some of the following characteristics, which can be combined with one another in the target immersive system provided that they are compatible with each other: zones of the display system of the target immersive system, not covering the images transmitted by the source system in the display data, display a display background generated by the calculation means of said target immersive system, in particular for improve the visibility or readability of the displayed representation.
  • the display background has one or more of: a uniform color, gray, or black; zones of different colors, gray or black, whose directions of separations and or transitions correspond to privileged directions of the environment; one or more textures.
  • a texture is superimposed on the image of the observation window. Such a texture increases, if necessary, the quality of the perception of the representation of the environment by the user of the target immersive system by producing a representation corresponding to a customary perception of the environment.
  • the calculation means of the target immersive system are configured to embed one or more interface elements in the display of the target display system, represented as objects with one, two or three dimensions.
  • the user of the target immersive system has a set of functions accessible to him through said interface elements, for example image freezing functions, recording functions, annotation functions. .
  • a position and / or dimensions of the images of the source display data represented in the display system of the target immersive system are a function of a position from which, and or to a direction to which, in the display system of said target immersive system, the observer immersed in said target immersive system looks at said display system of the target immersive system. In this way the movements of said observer are taken into account to correct the displayed images and maintain the vision with an exact perspective.
  • At least a portion of the images of the source display data are maintained displayed in the target display system when said images are theoretically out of a display area of the target display system, said at least a portion of the display image of the viewing window being shown held against an edge of said system display area target display in a direction in the repository of said target display system to which an observer of the target immersive system should move a direction to which it is looking or should move the orientation in the virtual environment of the target display system for closer to the observation direction corresponding to the display data received from the source system.
  • the user of the target immersive system does not totally lose the vision of the environment although its representation is theoretically out of his field of vision, allowing him to intuitively find the direction towards which the gaze must look to find his current position the representation of the virtual environment.
  • a symbol representation is embedded near an edge or attached to an edge of said system; display area in one direction, in the repository of the target display system to which the observer of the target immersive system should move a direction to which he is viewing or should move the orientation in the virtual environment of the display system target to approach the observation direction associated with the display data transmitted by the source system.
  • the target immersive system comprises interaction capabilities by a user of said target immersive system, by real or virtual commands accessible to said user, with the images displayed by the target display system.
  • the interaction capabilities include superimposing annotations, texts and or drawings to the displayed images.
  • the immersive system comprises a recorder of the images displayed by the target visualization system during an immersion session.
  • the invention also relates to a method 200, FIG. 4, for sharing an immersive environment between a source system 10 and at least one system immersive target 20, a device 100, limiting the data exchanged both because of their confidential appearance and limiting the hardware and software requirements of the target system.
  • a method 200, FIG. 4 for sharing an immersive environment between a source system 10 and at least one system immersive target 20, a device 100, limiting the data exchanged both because of their confidential appearance and limiting the hardware and software requirements of the target system.
  • the method there is exchanged between the source system and one or more target systems a set of information or data allowing each target system receiving the source system of the operating data to reconstitute and update in real time a visual representation.
  • each of the source system and the target immersive system or systems that will intervene in the immersion sharing integrates means of connection to transmission means 30 for connecting said source system and immersive systems. targets.
  • each source system and target immersive systems is associated a set of configuration data.
  • the configuration data includes all the information necessary for the physical or virtual description of the source system or the target immersive system, and in particular for its reconstruction at scale 1.
  • the configuration data of an immersive system comprise for example:
  • the configuration data are established for each system, immersive or not, of the device, manually, automatically or semi-automatically depending on the capabilities of the system.
  • some of these data may vary over time, for example in the case of a mobile immersive system whose orientation and or position of the screens vary over time during a period of time. immersion.
  • the configuration data is transmitted to the target system (s) in a manner similar to the operating data as discussed for the following steps.
  • the configuration data can be incorporated into the operating data.
  • a first initialization step 220 the device 100, combining the source systems 10 and targets 20 by means of which an immersion must be provided by the target immersive systems, or even shared with the source system, is initialized.
  • each of the target immersive systems 20 of the device initializes a register of the connected source system, said register advantageously comprising all the configuration data of each of said connected systems, and advantageously all of the configuration data at the less than the source system 10 for the target systems 20.
  • the configuration data, as well as initialization values of variables, initialized by an immersive or non-given system may be limited to the data strictly necessary for the data transmissions and calculations that will have to be performed by said system.
  • all the configuration data of the source system must be accessible by a target immersive system so that said source system can transmit the display data to said target system, the accessibility to these data can be realized in a embodiment of a device 100 for example via a registration in a database local to the digital processing system of said target immersive system.
  • a data server 32a connected to the network 31 of the data transmission means 30, and accessible by each of the source systems 10 and target immersive systems 20, comprises a database in which are stored configuration data of each of the immersive systems or not of said device, at least for those which are necessary during the initialization step, and each of the immersive systems during this initialization step receives from the server the data of configuration, at least for those which are necessary to him considering the transformations which it will have to realize on the data of posting.
  • the database may be updated during the implementation of the method depending on the connection, disconnection or change of role of at least one immersive system. network.
  • the operating data sources include:
  • the data of the observation conditions comprise the position from which and the direction in which, each of the position and direction can be measured or presupposed, the user looks in the source display system 12 at a given time, or in which images have been generated for example by a device for shooting a virtualized real environment.
  • This data is taken into account in calculating the representation of the virtual environment for the attention of a user of the source system immersed in said environment and involved in the construction of the displayed images.
  • the positions and or the direction are presupposed, for example by identification of a position and or ideal orientation to use the display system, or average if several users can be accommodated, or imposed if the immersive system. encourages the user (s) to approach a specific position and orientation.
  • the display data, calculated for each display cycle as a function of a refresh rate of said display data are captured, once calculated, by a software application operated in the digital processing system 11 of the source system.
  • the display data is captured during their transmission to the source display means 12, for example by a connection to the video link 13 of a video equipment 14 between the digital processing system. source 11 and the source display means 12.
  • the captured display data is captured for only one of the viewing conditions, for example the display data corresponding to a single eye, or stereo, depending on the capabilities of the target system to render or not stereoscopic images or according to options selected by an operator of the target system. It is thus possible to reduce the data flow and the bandwidth requirements of the data transmission medium 30.
  • the display data is captured for each of the observation conditions and associated with the corresponding observation conditions in the data. operating.
  • the video equipment 14 thus makes the capture of the display data during their transmission to the source display system 12, for example by reading the signals corresponding to the images on the video link 13, converts said images into a format adapted to their transmissions and transmits them via the data transmission means 30 of the device.
  • the source operating data are sent, directly or indirectly, by the source system to the users of the information contained in said source operating data.
  • the users are for example each of the target immersive systems which are connected to said source system or an associated computer 32b connected to the network 31 which receives the data for retransmitting them to the target systems, possibly after performing a processing of the operating data.
  • a second transformation step 240 the source operating data is transformed according to the configuration data of a target immersive system, and as a function of variables generated by said target immersive system 24, especially the observation data corresponding to the actual viewing conditions in the display system 22 of said target immersive system.
  • the virtual 3D reconstruction of the source display system without an image displayed on display means of the virtual source system, can be cached by the digital processing system of the target immersive system, and thus be reused for the following iterations.
  • the data of the observation conditions of the virtual environment of the source system are the same as the data of the observation conditions of the observer 91 in the displayed virtual environment in the virtual source display system on the target system, the visual representation of said virtual environment being reconstructed for this purpose.
  • This condition is forced so that the images generated by the source system 10 can be seen without deformation by the observer 91 in the target display system 22.
  • the display data initially calculated by the source system therefore do not need to be recalculated for new observation conditions in the target system, the observation conditions, in position and in direction, being virtually superimposed in the virtual environment of the source system and in the virtual environment of the target system.
  • the set of data of the observation conditions can be duplicated to correspond to each of the points of observation.
  • observation or the observation conditions for the second point being limited to relative position data of the second point with respect to the first point, for example distance and direction of the second point with respect to the first point. point, limiting the volume of the observation condition data to be transmitted.
  • the observation data may also correspond to a fictitious observation point, for example a cyclopic eye, and the observation conditions for each of the points being limited to relative position data of each of the real observation points with respect to the observation point. fictitious point.
  • all or part of the operating data transmitted by the source system is frozen either by the source system before transmission or by the target immersive system upon receipt.
  • These may be display data, or observation condition data, or a subset of the observation condition data.
  • the frozen data is no longer issued by the source system, or is issued with constant values, or is temporarily no longer taken into account by the target system and the latest farm data received and processed by the system target are frozen.
  • the gel is controlled by an instruction of one of the source or target systems, for example on an order from the user 90, or from the observer 91, or from an operator of the source system or of the target system.
  • the order can also affect the source system which is then frozen and in this case the target system undergoes the freezing of the concerned data, or affects only the target system which in this case does not necessarily imply the freezing of the source system but only a stop of taking into account the evolutions of the concerned data emitted by the source system.
  • this data freezing step 241 When this data freezing step 241 is activated, the possibility of updating the images on the target immersive system by processing the operating data fixed in said operating freeze mode is maintained, said target immersive system in this case, it is possible to recalculate in real time the display data for the images to be displayed by the target display system 22 as a function of the modification of the observation conditions in said target display system in order to maintain a correct representation of the virtual environment frozen.
  • the display data in the target immersive system is then interpreted, when the corresponding data is frozen, as if the virtual environment were totally static and the user of the system source was no longer moving, resulting in a kind of 3D capture of the state of the source system.
  • the freezing step can also concern all the data relating to the observation conditions, or only a subset of these data. For example, during a freezing step, an operator, a user or an observer may choose whether to freeze the position variables of the observation data and whether to freeze the orientation variables of the observation data. .
  • the data displayed by the virtual source display system no longer varies, and the image points of said virtual source display system on which the scores and or annotations are made are constant.
  • a step 243 for recording the source operating data is activated.
  • said prerecorded source operating data is used by the processing system as source operating data, thereby allowing the target system, which could be said source system, to replay a prerecorded immersion. .
  • a third display step 250 the set of display data resulting from the transformation is transmitted to the target display system 22 whose operating data have been taken into account during the transformation of the images generated by the display. source system.
  • the process is looped back after the third step 250 on the second step 240.
  • the cycle with which the process is looped back to display a new representation of the virtual environment on the display system of the target system a cycle whose time can be constrained by the computational performance of the digital processing systems of each of the source systems and targets, as well as by the performance of the data transmission means by which the data is exchanged between said source and target systems, is repeated each time the target immersive system has acquired operating data corresponding to a representation of the virtual environment on the source system.
  • the calculation is advantageously repeated as quickly as the digital processing systems of the source and target systems allow, at least to provide an observer 91 a perceived immersion coherent with its generation by the source system, for example almost simultaneous with that of the user 90 of the source system when said source system is an immersive system.
  • the notion of observation conditions includes the notions of observation position and or direction of observation. These position and direction parameters can be derived from measurements or analyzes. For example, the direction of observation can be deduced from a position measurement of the two eyes of a user or an observer.
  • Presupposed or predetermined observation conditions can also be used for one or the other of the source system and a target immersive system, or for both, the implementation of the method for adapting the displays by the taking into account the differences between the two systems.
  • the device may comprise a source system and any number of target immersive systems, as shown diagrammatically in FIG. 3 of a device 100 comprising three target immersive systems 20 of different types, that can simultaneously reproduce a same virtual environment generated by the source system in the target systems.
  • the invention makes it possible to perform immersion sharing in numerous immersive system configurations, in particular display systems of said immersive systems, FIG. 3 illustrating possible configurations in a nonlimiting manner.
  • the source system and a target immersive system both include a multi-sided display system.
  • the immersive systems in this case have systems for measuring the position of the observation point and the observation direction of the user or the observer.
  • the source system includes a multi-sided display system and a target system includes a headset type display system.
  • the helmet-type visualization systems use screens that are physically very close to the eyes of the user of the helmet but which are provided, in order to remain usable, with optical systems that provide a collimation that places a virtual image, here considered at home. sense of geometric optics, of the surface of the screen at a sufficient distance from the eye. It is this virtual image which is then considered as the image displayed in the visualization system.
  • the observer using the headset is placed in the virtual environment containing the reconstituted source system at exactly the same position and orientation as those of the source system user in the source system.
  • the source orientation around the axis defined by the center of the two eyes is not taken into account to give a little more freedom to the observer.
  • the observer in the target system sees exactly what the user of the source system sees.
  • the observer using the headset is placed in the virtual environment containing the source system reconstructed at exactly the same position as that of the source system user in the source system.
  • the observer in the target immersive system can look in other directions than the user of the source system.
  • the source system includes a headphone type display system.
  • the configuration information then includes a description of the inverse geometric transformation of the geometric transformation associated with the pixels of the images intended for each of the two eyes as well as the field of vision associated with the helmet.
  • the target immersive system receives the display data directly captured by the capture system and applies to them the geometric transformation described in the configuration information to find flattened images for the left eye and the right eye without the optical deformations. often required for display in this type of immersive helmet type system.
  • the source immersive system is of helmet type and the target immersive system is also helmet type.
  • the transformation performed by the target system then consists in placing around the observer in the target system, in a virtual environment, a moving window in which the contents coming from the source system are displayed, the size of the window depending on the characteristics the source system to at least match the field of view rendered by said source system.
  • the window is displayed all the time in the field of view of the observer, thus imposing that its position and its orientation in the virtual environment exactly correspond to the position and orientation of the user of the system. source in the source system.
  • the position of the observer in the target system is constrained in the virtual environment.
  • the observer can orient his gaze in any direction, even if he only perceives the content visualized by the user of the source system when the viewing directions of the one and the other are sufficiently close.
  • the source system does not necessarily include a source display system, however, the operating data includes all the data of the observation conditions associated with the display data of said operating data.
  • the symbolic representations are advantageously superimposed on the visual representation of the virtual environment represented on the target immersive system to provide the observer with assistance with the orientation and the choice of an observation direction.
  • symbolic representations comprise for example the graphical materialization of a sphere or a grid associated with the representation of horizontal surfaces, for example a floor, and or vertical, for example a wall.
  • Such symbolic representations which can only be materialized temporarily, for example by a command from the observer, or under certain conditions, for example a rapid evolution of the observation conditions of the source system, provide assistance to the observer to facilitate his choice of direction of observation and to find his position in the virtual scene.
  • Such graphical elements integrated into the viewer's field of view will show him in which direction to look for the image viewed by the user of the source system, ie how to look in the same direction.
  • the display data of the old images displayed in the virtual environment of said observer may take a long time to disappear, allowing then to superimpose in his field of vision the contents currently viewed according to the direction of gaze current user of said user and the contents previously viewed according to past viewing directions of said user. If said observer does not look exactly in the same direction as said user, he will be able to perceive a larger portion of the virtual environment viewed by said user.
  • the source system is headset type and the target system is a traditional flat screen, possibly stereoscopic.
  • transforming the display data coming from said headset as a function of the position data and or orientation coming from said headset makes it possible to position in the virtual environment of the target system the content portion of the virtual environment visible through the display system of the source system in a stabilized manner, mitigating the impact of frequent head movements of the user of the source system.
  • a processing allowing such a stabilization can be implemented whatever the source and target systems used, and is mainly meaningful if the source system has means for measuring the orientation of the observation point.
  • the virtual environment window of the target immersive system in which the image corresponding to the visible source virtual environment portion is displayed is positioned on a geometric shape comparable to a fixed sphere. and invisible centered on the head of the observer, its position on the surface of the sphere being a function of the viewing direction of the user of the source system, said function being linear or not.
  • the sphere can turn on itself to accompany the user's view of the source immersive system when the user looks in a direction initially located outside the field of the display system target.
  • the source immersive system may also be augmented reality viewing headphones, in which case the environment taken into account by the capture means of said source system is a combination of the virtual displays and the images. of the real scene, both superimposed.
  • the display means of the source immersive system are virtual, the main issue being to compute the contents to be displayed on a target immersive system by the digital processing means of the source immersive system without worrying about display information about the display system of the source immersive system.
  • the source immersive system can use as position and orientation of the calculation of the display data the position and orientation of the observer in the target immersive system, which reaches said source system via the network connection means.
  • the source system computes a dimensionless image consisting of a 360 ° projection of the virtual environment, resulting in an image or set of images that can be transformed by the target immersive system. taking into account the difference between the observation conditions used by the source system and the conditions of effective observation of the observer in the target immersive system at the time of display of the image.
  • the source system calculates a set of display data corresponding to the geometric configuration of the display system of the target immersive system.
  • the transformations performed by the target immersive system take into account the difference between the position used by the source system and the actual position of the observer in the target immersive system at the time of display of the image.
  • an immersive system comprising other forms of screens can be implemented in the device and in the process as both source system and target system.
  • the necessary transformations are ensured by taking into account the curvature characteristics of the surface on which is, in the case of the source system, or must be, in the case of the target system, displayed an image .
  • the invention may, in particular in this situation, adapt an image to a different curvature between a screen of the source system and a screen of the target system, since the processing system associated with the target system has all the geometric information associated with the source system. to build virtually exactly the same screen with the same curvatures.
  • the display system of a target immersive system comprises a single conventional flat screen, optionally having stereoscopic display capabilities.
  • Such an embodiment allows several viewers to live in immersion on the target system a user experience immersed in the source system.
  • the target user's position and viewing direction of the target system is presupposed to correspond to default position and orientation in the target system.
  • the display system of the target system is a mobile device or portable tablet, smartphone or laptop.
  • the display system and / or the target immersive system and / or source system processing system are associated with equipment, generally combining hardware and software, implemented. to interact with the visual representations on the display system, which will be grouped under the expression "means of interaction".
  • These Interaction means can be of any form as long as they make it possible to influence, via the display system and or the processing system, the content and or the behavior of the images.
  • These interaction means include, for example, non-exhaustively, terminals with direct or indirect tactile capabilities, a keyboard, a pointer such as a mouse or a tracking ball, digital pens, a 3D pointing device. , a gyro pointing device, gaze tracking systems ...
  • the freezing function of the display data on the target system can then be activated by these interaction means and the observer (s) on the target system can using the interaction means of said target system to point and or annotate, with text and or with drawings and or with symbols, to the surface of the frozen image, and then record these annotated images by a suitable interface.
  • Such possibilities of interactions, pointing and annotations are advantageously implemented on each of the source and target systems so that the user on the source system and the observer on the target immersive system can use the capabilities thereof.
  • the scores and annotations performed on a system, target or source are also transmitted by the communication means in the form of digital data to the other or to several other interconnected target immersive systems so that the scores and annotations, made by an observer or by the user, can be embedded in the visual representations displayed on each of the other immersive systems concerned.
  • these parameters can be estimated as functions of measurements or observations more or less precise, but sufficient to obtain a reconstruction of the image of the immersive environment acceptable to the observer.
  • the target system may be a computer phone whose screen forms the display system and a camera on the front of said computer phone provides, by a dedicated software application, monitoring the position of the head or eyes of the observer.
  • the parameters of the source and target observation conditions are necessary for the formulation of the transformations of the images of the source system into coherent images on the target system.
  • the values of the parameter are taken into account or failing this, are taken to presupposed values according to the display means and the observation conditions.
  • the display system of the target immersive system may be a screen associated with tactile capabilities as means of interaction.
  • the observer can then use these interaction means to modify the position and / or direction of observation parameters and thus dynamically determine a position and / or an observation direction according to which to observe the virtual representation of the source immersive environment.
  • the touch capabilities can be replaced by a mouse device.
  • the 2D or 3D display control by these interaction means is well known to those skilled in the art.
  • the interaction means of the target immersive system thus allow the observer to control a subset of the observation conditions for the target immersive system.
  • all the types of immersive systems described can be combined in a more or less complex device that does not restrict itself to a single pair of source system / target system.
  • the invention can also be applied to different types of virtual environments.
  • the virtual environment can be a three-dimensional environment.
  • the virtual environment can also be a two-dimensional environment represented by still or animated stereoscopic images or not.
  • the knowledge of properties of the camera can make it possible to restore the correct proportions for the observer on the target system.
  • it is possible to reproduce an environment generated on a source system in one or more target immersive systems without necessarily transmitting information other than that corresponding to the configuration and operating data generated. by the source system.
  • the environment perceived by a user of the source system is reproduced in a consistent manner with an observer of the target system which thus benefits from the immersion experience in the system. virtual environment of the source system.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Optics & Photonics (AREA)
  • Processing Or Creating Images (AREA)
EP17725317.6A 2016-04-27 2017-04-27 Vorrichtung und verfahren zum teilen der immersion in eine virtuelle umgebung Pending EP3449340A1 (de)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
FR1653758A FR3050834B1 (fr) 2016-04-27 2016-04-27 Dispositif et procede de partage d'immersion dans un environnement virtuel
FR1659160A FR3056770B1 (fr) 2016-09-28 2016-09-28 Dispositif et procede de partage d'immersion dans un environnement virtuel
FR1659768A FR3057430B1 (fr) 2016-10-10 2016-10-10 Dispositif d'immersion dans une representation d'un environnement resultant d'un ensemble d'images
PCT/FR2017/051004 WO2017187095A1 (fr) 2016-04-27 2017-04-27 Dispositif et procede de partage d'immersion dans un environnement virtuel

Publications (1)

Publication Number Publication Date
EP3449340A1 true EP3449340A1 (de) 2019-03-06

Family

ID=58765863

Family Applications (1)

Application Number Title Priority Date Filing Date
EP17725317.6A Pending EP3449340A1 (de) 2016-04-27 2017-04-27 Vorrichtung und verfahren zum teilen der immersion in eine virtuelle umgebung

Country Status (5)

Country Link
US (1) US11727645B2 (de)
EP (1) EP3449340A1 (de)
CA (1) CA3022298A1 (de)
SG (1) SG11201810432YA (de)
WO (1) WO2017187095A1 (de)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102336997B1 (ko) * 2017-08-16 2021-12-08 삼성전자 주식회사 서버, 디스플레이장치 및 그 제어방법
CA3139465A1 (en) 2019-06-20 2020-12-24 Barrie A. Loberg Voice communication system within a mixed-reality environment
FR3102592B1 (fr) * 2019-10-29 2022-02-11 Immersion Procédé d’interaction avec un utilisateur d’un système immersif et dispositif pour la mise en œuvre d’un tel procédé
WO2021190280A1 (en) * 2020-03-24 2021-09-30 Guangdong Oppo Mobile Telecommunications Corp., Ltd. System and method for augmented tele-cooperation

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103039072A (zh) * 2010-05-25 2013-04-10 维德约股份有限公司 用于使用多个摄影机和多个监视器的可缩放视频通信的***和方法
US20120162384A1 (en) * 2010-12-22 2012-06-28 Vesely Michael A Three-Dimensional Collaboration
US8934015B1 (en) * 2011-07-20 2015-01-13 Google Inc. Experience sharing
US9268406B2 (en) 2011-09-30 2016-02-23 Microsoft Technology Licensing, Llc Virtual spectator experience with a personal audio/visual apparatus
BR112014010230A8 (pt) * 2011-10-28 2017-06-20 Magic Leap Inc sistema e método para realidade virtual e aumentada
CN105188516B (zh) * 2013-03-11 2017-12-22 奇跃公司 用于增强和虚拟现实的***与方法
JP2015095802A (ja) 2013-11-13 2015-05-18 ソニー株式会社 表示制御装置、表示制御方法、およびプログラム
US20160005209A1 (en) * 2014-07-07 2016-01-07 Karlsruher Institut für Technologie Method and system for light transport path manipulation
US20160027218A1 (en) * 2014-07-25 2016-01-28 Tom Salter Multi-user gaze projection using head mounted display devices
US20160133230A1 (en) * 2014-11-11 2016-05-12 Bent Image Lab, Llc Real-time shared augmented reality experience
US9516255B2 (en) * 2015-01-21 2016-12-06 Microsoft Technology Licensing, Llc Communication system
WO2016164342A1 (en) * 2015-04-06 2016-10-13 Scope Technologies Us Inc. Methods and apparatus for augmented reality applications
US10380800B2 (en) * 2016-04-18 2019-08-13 Disney Enterprises, Inc. System and method for linking and interacting between augmented reality and virtual reality environments

Also Published As

Publication number Publication date
US20190139313A1 (en) 2019-05-09
SG11201810432YA (en) 2018-12-28
US11727645B2 (en) 2023-08-15
CA3022298A1 (fr) 2017-11-02
WO2017187095A1 (fr) 2017-11-02

Similar Documents

Publication Publication Date Title
US11838518B2 (en) Reprojecting holographic video to enhance streaming bandwidth/quality
US10527846B2 (en) Image processing for head mounted display devices
US8928659B2 (en) Telepresence systems with viewer perspective adjustment
US12026833B2 (en) Few-shot synthesis of talking heads
CN105264478A (zh) 全息锚定和动态定位
WO2017187095A1 (fr) Dispositif et procede de partage d'immersion dans un environnement virtuel
TWI813098B (zh) 用於新穎視圖合成之神經混合
CN107810634A (zh) 用于立体增强现实的显示器
CN115413353A (zh) 扩展现实记录仪
JP7426413B2 (ja) ブレンドモード3次元ディスプレイシステムおよび方法
CN111699460A (zh) 多视图虚拟现实用户界面
FR3056770A1 (fr) Dispositif et procede de partage d'immersion dans un environnement virtuel
BE1022580A9 (fr) Méthode d'obtention de vidéos immersives avec parallaxe interactive et méthode de visualisation de vidéos immersives avec parallaxe interactive
FR3066304A1 (fr) Procede de compositon d'une image d'un utilisateur immerge dans une scene virtuelle, dispositif, equipement terminal, systeme de realite virtuelle et programme d'ordinateur associes
FR3050834A1 (fr) Dispositif et procede de partage d'immersion dans un environnement virtuel
Barabas Holographic television: measuring visual performance with holographic and other 3D television technologies
FR3057430A1 (fr) Dispositif d'immersion dans une representation d'un environnement resultant d'un ensemble d'images
Nocent et al. 3d displays and tracking devices for your browser: A plugin-free approach relying on web standards
US20220232201A1 (en) Image generation system and method
Hristov Research of Modern Technologies and Approaches for the Development of a Web-Based Information System for Visualization of Three-Dimensional Models...
CN116848507A (zh) 应用程序投屏

Legal Events

Date Code Title Description
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: UNKNOWN

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20181126

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

DAV Request for validation of the european patent (deleted)
DAX Request for extension of the european patent (deleted)