CN207898959U - A kind of simulation laboratory based on virtual reality mixing of stabilization - Google Patents
A kind of simulation laboratory based on virtual reality mixing of stabilization Download PDFInfo
- Publication number
- CN207898959U CN207898959U CN201721845153.1U CN201721845153U CN207898959U CN 207898959 U CN207898959 U CN 207898959U CN 201721845153 U CN201721845153 U CN 201721845153U CN 207898959 U CN207898959 U CN 207898959U
- Authority
- CN
- China
- Prior art keywords
- scene
- host
- virtual
- experiencer
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
- 238000004088 simulation Methods 0.000 title claims abstract description 16
- 230000006641 stabilisation Effects 0.000 title claims abstract description 12
- 238000011105 stabilization Methods 0.000 title claims abstract description 12
- 238000002156 mixing Methods 0.000 title claims description 10
- 230000015572 biosynthetic process Effects 0.000 claims abstract description 39
- 238000003786 synthesis reaction Methods 0.000 claims abstract description 39
- 238000009877 rendering Methods 0.000 claims abstract description 33
- 230000005540 biological transmission Effects 0.000 claims description 12
- 239000004973 liquid crystal related substance Substances 0.000 claims description 3
- 239000002131 composite material Substances 0.000 claims description 2
- 238000007654 immersion Methods 0.000 abstract description 6
- 230000002452 interceptive effect Effects 0.000 abstract description 4
- 208000002173 dizziness Diseases 0.000 abstract description 3
- 238000012549 training Methods 0.000 description 10
- 238000005452 bending Methods 0.000 description 3
- 238000000034 method Methods 0.000 description 3
- 241001269238 Data Species 0.000 description 2
- 230000009471 action Effects 0.000 description 2
- 238000010276 construction Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 239000011521 glass Substances 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 210000000697 sensory organ Anatomy 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Landscapes
- Processing Or Creating Images (AREA)
Abstract
The utility model provides a kind of simulation laboratory of stabilization mixed based on virtual reality, is related to virtual emulation practice teaching room field, including:Aobvious equipment;Handle;Shoot camera;Locator is fixed on shooting camera;Scene of game renders host, for virtual scene to be split into flat image, and the data of the flat image is transferred to rendering synthesis host;Router;Synthesis host is rendered, mixes scene for the data of the flat image and the image data of the experiencer of shooting camera shooting to be synthesized three-dimension virtual reality by composite software, and three-dimensional is shown that the data of scene are transferred to viewing screen;Show equipment;Scene of game renders host and renders is mounted on NDI softwares on synthesis host.The problem of the utility model solves existing bore hole 3D immersions interactive system and evaluator can not be assessed experiencer suitable for practice teaching field, and the picture seen of traditional simulation laboratory easy tos produce dizziness.
Description
Technical field
It is mixed based on virtual reality that the utility model is related to virtual emulation practice teaching room field more particularly to a kind of stabilizations
The simulation laboratory of conjunction.
Background technology
In practice teaching field, rail traffic vehicle braking teaching is for being giveed training to student, and student is to rail for culture
The understanding of road vehicular traffic brake unit design feature is grasped, to efficiently, scientifically be learnt;Existing rail traffic
The most of teaching using modes such as the PPT of book contents combination teacher of teaching, learning efficiency be not high;Another aspect training base
Because railroad track itself takes up a large area, Meteorological is high, scene set is fixed, many real training places can not put into construction, i.e.,
Make the corresponding training base of construction, cannot achieve during real training and replace scene or interaction according to demand, so as to cause real training
Effect is poor.
In order to solve this problem, a kind of bore hole 3D immersions interaction system based on simulated teaching Training Room may be used
System, specifically includes:Host is rendered, for connecting control engine and display device, and combination rendering engine realizes the rendering of picture;
Rendering engine, for using the software realizations such as ventuz fusion with it is synchronous after, pass through connect display device present bore hole 3D pictures;
Display device, for by showing the picture of rendering after being connect with rendering host;Router, for being connected by wireless network card
Render host and control engine;Engine is controlled, real-time, interactive control is realized for connecting router.But in actual application
In, although bore hole 3D immersions exchange method has certain benefit, still, experiencer can only stand could experience in specific position
Scene therein can not be experienced for onlooker to the sense organ of this immersion.
It in practice teaching field, needs to assess the training of experiencer, if using bore hole 3D immersions
Interactive system, the evaluator in specific position that do not stand will be unable to assess the training of experiencer;Therefore, for body
The assessment situation for the person of testing need the picture transmission for seeing experiencer in VR glasses to evaluator it can be seen that screen on, this
Sample evaluator carries out VR experiencers more accurate it can be seen that the real-time pictures that the action of VR experiencers and experiencer see
Assessment.
We mix virtually and reality scene by using synthesis host mixing scene of game rendering host is rendered, but
It is that, if rendering the data transmission between host and rendering synthesis host only with video frequency collection card for scene of game, can deposit
In the disadvantage that connection is more, transmission is unstable.
Utility model content
The purpose of this utility model is that:It can not be suitable for teaching in fact to solve existing bore hole 3D immersions interactive system
Evaluator assesses experiencer in instruction field, preferably should can not simultaneously solve in simulation type laboratory in teaching field
If certainly scene of game renders the data transmission between host and rendering synthesis host only with video frequency collection card, there can be circuit
The problem that connection is more, transmission is unstable, the utility model provide a kind of simulation laboratory of stabilization mixed based on virtual reality.
The technical solution of the utility model is as follows:
On the one hand, the utility model provides a kind of simulation laboratory of stabilization mixed based on virtual reality, including:Head is aobvious
Equipment watches virtual scene for experiencer, and the virtual scene is transferred to scene of game and renders host;
Handle, for being interacted between experiencer and virtual scene;
Camera is shot, for shoot the experiencer for showing equipment with head and being located in green background, and by the experiencer of shooting
Image data be transferred to rendering synthesis host;
Locator is fixed on shooting camera, the location information for tracking shooting camera, and location information is sent to
Scene of game renders host;
Scene of game renders host, for virtual scene to be split into flat image, and by the data of the flat image
It is transferred to rendering synthesis host;
Router renders host and rendering synthesis host for connecting scene of game;
Synthesis host is rendered, the image data of the experiencer by the data of the flat image and shooting camera shooting is used for
Three-dimension virtual reality is synthesized by composite software and mixes scene, and three-dimensional is shown that the data of scene are transferred to display and set
It is standby;
Equipment is shown, for showing virtual display scene;
The scene of game renders host and renders synthesis installs NDI softwares between host.
Specifically, it includes foreground, background, channel that the scene of game, which renders the flat picture picture that host generates,.
Preferably, data and the shooting for rendering synthesis host and synthesizing the flat image using VENTUZ composite softwares
The image data of the experiencer of camera shooting.
Specifically, the display equipment is LED display or liquid crystal display.
Specifically, it is connected by HDMI interface or SDI interfaces between the rendering synthesis host and display equipment.
After adopting the above scheme, the beneficial effects of the utility model are as follows:
(1) after the simulation laboratory of the utility model being used in practice-training teaching field, evaluator comments experiencer
Estimate no longer is that can only see the independent picture of virtual scene either experiencer, but experiencer is put into entire virtual scene
In, more really experience to evaluator, evaluator it can be seen that VR experiencers action and virtual scene, to VR experiencers
Have and more asks accurate assessment.
(2) image of the experiencer taken due to shooting camera is flat two-dimensional images, and virtual scene is three-dimensional
Scene, so, during scene of game rendering host is transmitted virtual scene, need first to tear three-dimensional virtual scene open
It is divided into the two dimensional image of plane, the two dimensional image of the flat two-dimensional images and experiencer synthesize three commonly through synthesis host is rendered
The scene of dimension and experiencer, scene of game render host by virtual scene split into containing have powerful connections, foreground and channel it is multiple
Two dimensional image, during these two-dimensional image datas are transferred to rendering synthesis host, if using capture card to these
Two dimensional image carries out ability transmission, then needs multiple connections, multiple connections can to transmit it is unstable, therefore, this reality
Router is connected between host and rendering synthesis host so that the two wirelessly connects with novel rendered in scene of game
It connects, is then transmitted again by NDI transmitting softwares, this addresses the problem transmit unstable problem.This of the utility model changes
Although small into putting, the proposition of problem is simultaneously readily apparent that.
Description of the drawings
Fig. 1 is the structural schematic diagram of the utility model
Fig. 2 is the data transmission scheme that virtual camera positions are determined in the utility model embodiment;
Specific implementation mode
The following will be combined with the drawings in the embodiments of the present invention, the technical solution in the present embodiment is carried out it is clear, complete
Site preparation describes, it is clear that described embodiment is only a part of the embodiment of the utility model, rather than whole implementation
Example.Based on the embodiments of the present invention, those of ordinary skill in the art are obtained without making creative work
The every other embodiment obtained, shall fall within the protection scope of the present invention.
Embodiment 1
As shown in Figure 1, a kind of simulation laboratory based on virtual reality mixing of stabilization, including:
Aobvious equipment watches virtual scene for experiencer, and the virtual scene is transferred to scene of game and renders master
Machine;It is exactly the usual VR glasses that the head, which shows equipment,.
Handle, for being interacted between experiencer and virtual scene;In general, the aobvious equipment of head and handle, which are into, is arranged
It sets;In the utility model, head used shows equipment and handle is the VIVE series produced using HTC companies.
Camera is shot, for shoot the experiencer for showing equipment with head and being located in green background, and by the experiencer of shooting
Image data be transferred to rendering synthesis host;Experiencer stands before green background, this is in order to facilitate in image processing process
By experiencer it is quick, accurately deduct from background.
Locator is fixed on shooting camera, the location information for tracking shooting camera, and location information is sent to
Scene of game renders host.
Scene of game renders host, and for virtual scene to be split into flat image, scene of game renders what host generated
It is flat to draw as including foreground, background, channel, and the data of the flat image are transferred to rendering synthesis host;Scene of game
It renders the data transmission that host and head are shown between equipment and uses USB interface and HDMI interface.
Router renders host and rendering synthesis host for connecting scene of game.
Synthesis host is rendered, the image data of the experiencer by the data of the flat image and shooting camera shooting is used for
Three-dimension virtual reality is synthesized by composite software and mixes scene, and three-dimensional is shown that the data of scene are transferred to display and set
It is standby.
Equipment is shown, for showing that virtual display scene, display equipment are LED display or liquid crystal display.It renders
It synthesizes and is connected by HDMI interface or SDI interfaces between host and display equipment.
The scene of game renders host and renders synthesis installs NDI softwares between host.
Embodiment 2
As shown in Figure 1, a kind of emulation mode of the simulation laboratory based on virtual reality mixing of stabilization, including walk as follows
Suddenly:
S1:Locator is fixed on shooting camera, locator is used to obtain the location information and shooting angle of shooting camera
Spend information;
S2:Experiencer is located in front of green background, and then experiencer shows equipment by head and watches virtual scene, by handle with
It is interacted between virtual scene, in virtual scene, establishes virtual shooting camera, specific angle is obtained by virtually shooting camera
The virtual scene of degree;
S3:Host is rendered by scene of game, the virtual scene of special angle is split into flat image, and these are put down
Face image transmitting gives rendering synthesis host;The angle for choosing virtual scene changes according to the variation of shooting camera position;
S4:While carrying out S3, experiencer's image is shot by shooting camera, shooting camera gives experiencer's image transmitting
Render synthesis host;
S5:Render the experiencer that scene of game is rendered the flat image and shooting camera shooting of host transmission by synthesis host
Image synthesizes three-dimension virtual reality and mixes scene, and gives the image transmitting of three-dimension virtual reality mixing scene to display equipment;
S6:Three-dimension virtual reality is mixed scene and shown by display equipment.
Embodiment 3
On the basis of embodiment 2, in the S3, determine the virtual scene of special angle the specific steps are:
S31:Virtual camera and virtual fixture are created in virtual scene, and virtual fixture is determined with the S1
Position device establishes set membership;Specifically, the data of locator are father, the data of virtual fixture are son, and the data of locator become
Virtual fixture is driven when change;
S32:A virtual small ball one, the position of this small ball are created in virtual scene according to the position of locator
For determining location information of the shooting camera in virtual environment;
S33:Virtual camera is moved at virtual small ball one;
S34:An object for there are corner angle is placed in front of shooting camera, while drawing a red point an edges and corners, is moved
Dynamic shooting camera makes red point just at the center position of camera;
S35:A virtual small ball two is created in corresponding virtual scene at red point using handle, virtual small ball two
To shoot the center of camera shooting picture;
S36:The rotation angle information of virtual camera is determined by virtual small ball one and virtual small ball two;
S37:A line is created at the center of virtual camera, while the father and son for establishing virtual camera and shooting between camera is closed
System, at this point, virtual camera is just overlapped with the position of real camera.
Operation principle:Shooting camera shoots the green background after experiencer and experiencer, then by the picture of shooting
Face sends rendering to and renders synthesis host, is to render synthesis host for the image segmentation between experiencer and green background
Middle progress, it should be understood that the segmentation this point for image is inevitable conventional steps, therefore the utility model the step of
Do not illustrate in description.Scene of game renders the virtual scene that host obtains special angle, the virtual scene of this special angle
It is to be shot by virtual camera, therefore, it is necessary to obtain the shooting angle of the position of virtual camera and camera lens, how obtains virtualphase
The position of machine and the shooting angle of camera lens subsequently illustrate.
The image for the experiencer that shooting camera takes is flat two-dimensional images, and virtual scene is three-dimensional scene, institute
During scene of game rendering host is transmitted virtual scene, to need first to split into three-dimensional virtual scene flat
The two dimensional image of the two dimensional image in face, the flat two-dimensional images and the experiencer field three-dimensional commonly through synthesis host synthesis is rendered
Scape and experiencer, scene of game render host by virtual scene split into containing have powerful connections, multiple X-Y schemes of foreground and channel
Then these two dimensional images are transferred to rendering synthesis host, render synthesis host by ventuz softwares by scene of game by picture
Experiencer's image synthesis three-dimension virtual reality of the flat image and shooting camera shooting that render host transmission mixes scene, then
Three-dimension virtual reality is mixed into scene by showing that equipment is shown, evaluator's direct viewing shows equipment, so that it may with simultaneously
See the virtual environment residing for experiencer and experiencer, the behavior of experiencer is preferably assessed.In practice,
The simulation laboratory of the utility model or method can not only be used for assessing, and can also be used in other amusements and come to other people
Viewing, to reach better entertainment effect.
It is bending since the head of VR aobvious equipment of experiencer's wearing shows minute surface, the virtual scene that experiencer sees is practical
On be bending, if directly by this bending virtual scene give evaluator watch, can allow people generate dizziness, therefore, this
Utility model is the virtual scene of special angle to be chosen by virtual camera, and be not the virtual scene that experiencer sees, is solved
The problem of evaluator that determined will produce dizziness, and the virtual scene for choosing special angle needs to determine by virtual camera, under
Face will illustrate the principle of the specific position of description virtual camera and the determination of rotation angle:As shown in Fig. 2, positioning
Device can position the position location of shooting camera, and virtual fixture and virtual camera and virtual hand are established in virtual scene
Handle, virtual fixture, virtual camera and virtual handle should establish father respectively with locator, handle and the locator in the display world
Subrelation.The location information of virtual fixture can be obtained directly by locator, and the location information of virtual fixture is obtained
The location information that virtual shooting camera can be directly acquired in of the invention, creates virtual small ball one and virtual small ball two,
When virtual small ball one and virtual small ball two point-blank when, the shooting angle of virtual camera and shooting camera overlaps,
To which virtual camera can form the state being completely superposed with the position of shooting camera.In the utility model, the position of virtual camera
It sets according to the position of actual shooting camera and the variation of angle and changes, evaluator is facilitated to watch render scenes from all angles
Scene is mixed with the virtual reality of experiencer's composition, more accurately the operation of experiencer more accurately can be commented
Estimate.
By these two-dimensional image datas be transferred to rendering synthesis host during, if using capture card to these two
Tie up image and carry out ability transmission, then need multiple connections, multiple connections can to transmit it is unstable, therefore, this practicality
Novel rendered in scene of game connects router between host and rendering synthesis host so that the two wirelessly connects
It connects, is then transmitted again by NDI transmitting softwares, this addresses the problem transmit unstable problem.This of the utility model changes
Although small into putting, the proposition of problem is simultaneously readily apparent that.
It is obvious to a person skilled in the art that the present invention is not limited to the details of the above exemplary embodiments, and
And without departing substantially from the spirit or essential attributes of the utility model, it can realize that this practicality is new in other specific forms
Type.Therefore, in all respects, the present embodiments are to be considered as illustrative and not restrictive, this practicality is new
The range of type is indicated by the appended claims rather than the foregoing description, it is intended that containing in the equivalent requirements of the claims will be fallen
All changes in justice and range are embraced therein.Any reference numeral in claim should not be considered as limitation
Involved claim.
In addition, it should be understood that although this specification is described in terms of embodiments, but not each embodiment is only wrapped
Containing an independent technical solution, this description of the specification is merely for the sake of clarity, and those skilled in the art should
It considers the specification as a whole, the technical solutions in the various embodiments may also be suitably combined, forms those skilled in the art
The other embodiment being appreciated that.
Claims (4)
1. a kind of simulation laboratory based on virtual reality mixing of stabilization, which is characterized in that including:
Aobvious equipment watches virtual scene for experiencer, and the virtual scene is transferred to scene of game and renders host;
Handle, for being interacted between experiencer and virtual scene;
Camera is shot, for shoot the experiencer for showing equipment with head and being located in green background, and by the figure of the experiencer of shooting
As data are transferred to rendering synthesis host;
Locator is fixed on shooting camera, the location information for tracking shooting camera, and location information is sent to game
Scene rendering host;
Scene of game renders host, for virtual scene to be split into flat image, and by the data transmission of the flat image
Host is synthesized to rendering;
Router renders host and rendering synthesis host for connecting scene of game;
Synthesis host is rendered, for passing through the image data of the data of the flat image and the experiencer of shooting camera shooting
Composite software synthesizes three-dimension virtual reality and mixes scene, and three-dimensional is shown that the data of scene are transferred to display equipment;
Equipment is shown, for showing virtual display scene;
The scene of game renders host and renders synthesis installs NDI softwares between host.
2. a kind of simulation laboratory based on virtual reality mixing of stabilization according to claim 1, which is characterized in that institute
State the experiencer for rendering data and shooting camera shooting that synthesis host synthesizes the flat image using VENTUZ composite softwares
Image data.
3. a kind of simulation laboratory based on virtual reality mixing of stabilization according to claim 1, which is characterized in that institute
It is LED display or liquid crystal display to state display equipment.
4. a kind of simulation laboratory based on virtual reality mixing of stabilization according to claim 1, which is characterized in that institute
It states and is connected by HDMI interface or SDI interfaces between rendering synthesis host and display equipment.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201721845153.1U CN207898959U (en) | 2017-12-25 | 2017-12-25 | A kind of simulation laboratory based on virtual reality mixing of stabilization |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201721845153.1U CN207898959U (en) | 2017-12-25 | 2017-12-25 | A kind of simulation laboratory based on virtual reality mixing of stabilization |
Publications (1)
Publication Number | Publication Date |
---|---|
CN207898959U true CN207898959U (en) | 2018-09-25 |
Family
ID=63569088
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201721845153.1U Expired - Fee Related CN207898959U (en) | 2017-12-25 | 2017-12-25 | A kind of simulation laboratory based on virtual reality mixing of stabilization |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN207898959U (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107976811A (en) * | 2017-12-25 | 2018-05-01 | 河南新汉普影视技术有限公司 | A kind of simulation laboratory and its emulation mode based on virtual reality mixing |
CN109782903A (en) * | 2018-12-18 | 2019-05-21 | 北京牡丹电子集团有限责任公司数字电视技术中心 | A kind of garden display systems and method based on virtual reality technology |
CN113128716A (en) * | 2021-04-25 | 2021-07-16 | 中国科学院计算机网络信息中心 | Operation guidance interaction method and system |
-
2017
- 2017-12-25 CN CN201721845153.1U patent/CN207898959U/en not_active Expired - Fee Related
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107976811A (en) * | 2017-12-25 | 2018-05-01 | 河南新汉普影视技术有限公司 | A kind of simulation laboratory and its emulation mode based on virtual reality mixing |
CN107976811B (en) * | 2017-12-25 | 2023-12-29 | 河南诺控信息技术有限公司 | Virtual reality mixing-based method simulation laboratory simulation method of simulation method |
CN109782903A (en) * | 2018-12-18 | 2019-05-21 | 北京牡丹电子集团有限责任公司数字电视技术中心 | A kind of garden display systems and method based on virtual reality technology |
CN113128716A (en) * | 2021-04-25 | 2021-07-16 | 中国科学院计算机网络信息中心 | Operation guidance interaction method and system |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107976811A (en) | A kind of simulation laboratory and its emulation mode based on virtual reality mixing | |
CN110233841B (en) | Remote education data interaction system and method based on AR holographic glasses | |
CN107274469A (en) | The coordinative render method of Virtual reality | |
CN207898959U (en) | A kind of simulation laboratory based on virtual reality mixing of stabilization | |
CN208126674U (en) | A kind of architectural engineering tutoring system based on VR virtual reality technology | |
CN104484327A (en) | Project environment display method | |
CN202003534U (en) | Interactive electronic sand table | |
CN207601427U (en) | A kind of simulation laboratory based on virtual reality mixing | |
CN105429989A (en) | Simulative tourism method and system for virtual reality equipment | |
CN109636916A (en) | A kind of a wide range of virtual reality roaming system and method for dynamic calibration | |
CN107154197A (en) | Immersion flight simulator | |
CN108831231A (en) | A kind of fire-fighting simulation virtual platform | |
CN106340064A (en) | Mixed-reality sandbox device and method | |
CN107783306A (en) | A kind of bore hole 3D immersions exchange method and system based on simulated teaching Training Room | |
CN113011723A (en) | Remote equipment maintenance system based on augmented reality | |
CN106558017A (en) | Spherical display image processing method and system | |
CN208985412U (en) | Wisdom education physical environment system | |
WO2020177318A1 (en) | Virtual reality-based craft-beer saccharification operation system and method | |
CN108830944B (en) | Optical perspective three-dimensional near-to-eye display system and display method | |
TW201828258A (en) | Method and device for rendering scene | |
CN110290409A (en) | Data processing method, VR equipment and system | |
Zerman et al. | User behaviour analysis of volumetric video in augmented reality | |
CN106530408A (en) | Museum temporary exhibition planning and design system | |
CN109445598A (en) | A kind of augmented reality system and device of view-based access control model | |
CN111840999A (en) | Game education method based on three-dimensional object graph recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20180925 |