Disclosure of Invention
In order to solve the problems, the invention provides a method and a system for realizing a user-defined picture or video with a channel based on augmented reality display, wherein a user can select one or more suitable user-defined pictures or videos according to a preset three-dimensional scene model template, then the user-defined pictures or videos are input into a server database through a production end, and then the original data is processed by a manufacturing end, so that the one or more user-defined pictures or videos can be combined with a three-dimensional scene model on a handheld device and browsed in an augmented reality environment, thereby improving the immersion of the user in browsing the pictures or videos and improving the interestingness and the interactivity. In addition, the invention provides a method and a system for realizing the user-defined picture or video based on the augmented reality display band channel, which are composed of a production end, a manufacturing end, a server and a handheld device, and provides a flow framework for multi-user and multi-data distributed manufacturing, and the method and the system have the advantages of high conversion speed and simpler realization process.
In order to achieve the above object, in one aspect, the present invention provides an AR display system applied to images or videos, which includes a production end, a server, a storage unit, a camera, an AR processing unit, and a display terminal, where the production end is configured to upload original production data to the server; the original production data consists of user character information and an original user scene; the manufacturing end is used for processing the original production data and then synthesizing the original production data into scene manufacturing data; the camera is used for acquiring a user image; the server is used for acquiring scene making data, combining the scene making data with a transparent model part in a three-dimensional scene model preset by the server and matching user images, the scene making data, the three-dimensional scene model, audio and user character information; the storage unit is used for storing the matched user pictures, scene making data, three-dimensional scene models and audios in the server; the AR processing unit is used for identifying the user image, combining the user image with the three-dimensional scene model, the scene making data and the audio in the storage unit and finishing displaying on the display terminal.
The original scene of the user is a scene image of the user.
The original scene of the user is a user scene video.
The scene making data are a model chartlet image and a recognizable image.
The scene making data is a model map video.
The recognizable image is obtained by presetting a recognizable image template and a model chartlet image by a server and manufacturing the recognizable image by a manufacturing end.
The manufacturing end processes the original scene of the user through Adobe Photoshop; the AR processing unit is a Vuforia AR unit.
And the model chartlet image is synthesized into an identifiable image through a production end.
And the manufacturing end submits the recognizable image to the AR processing unit to form recognizable data.
And the production end acquires the processed recognizable image through the server.
On the other hand, the invention discloses an AR display method applied to images or videos, which comprises the following implementation processes: the production end uploads the original production data of user information, images and videos to the server, wherein the user information comprises: user name, gender, selected three-dimensional scene, mobile phone number, commemorative day, remark information. The user information is user character information, the user information forms a unique character string on the server, and the image and the video refer to video and images provided by the user and are collected through the camera equipment of the user. The production end acquires the original production data of the image and the video through the server, and submits the recognizable image to the Vufaria AR unit to form the recognizable data through Adobe Photoshop processing, model mapping image synthesis and the recognizable image. The model chartlet image, the recognizable image and the recognizable data are transmitted back to the server through the production end. And the production end acquires the processed recognizable image through the server. The camera carries out data matching and identification on the user information through the server, specifically, the camera obtains the image information and then matches with the user information character string on the server, and the matching refers to searching the corresponding data information on the server after obtaining the identifiable image information. And storing the user information character strings and the processed recognizable image, the map image, the three-dimensional scene model, the audio and the video into a storage unit in a one-to-one correspondence manner through a server. After the image data is identified by the camera through the Vuforia AR unit, the mapping image, the three-dimensional scene model, the audio and the video data are extracted from the storage unit, and the interactive functions of display, clicking and swiping a fixed or unfixed area and the like are completed on the display terminal. The identification refers to identifying the above identifiable images by using a Vuforia AR unit, then acquiring continuous real-time pictures through a camera, and outputting and jointly displaying three-dimensional scene models, audio, model map pictures and model map video data which are correspondingly matched with the identifiable images on a display.
The method comprises the following concrete steps:
step 1: a user inputs original manufacturing material data including one or more original picture data and/or original video data to a server through a production end, and the original manufacturing material data is applied to user information data of user identity information verification; the method comprises the following specific steps:
1) inputting user information data to a server through a production end, and creating a data packet of the user in a database;
2) outputting a two-dimensional scene image template in a database to a production end by a server, selecting favorite one or more user-defined original picture data and/or original video data by a user through previewing and selecting the template, inputting the user-defined original picture data, an additional piece of original picture data or two-dimensional scene image data into a user data packet, and designating the user-defined original picture data, the additional piece of original picture data or the two-dimensional scene image data for a channel mapping model as a marked image for augmented reality identification;
step 2: the making end is connected with the server, original picture data and/or original video data in the user data packet are output from the database, the data are processed through the making end to complete making, and then the data are output into the user data packet; inputting the marked image used for augmented reality identification in the data packet into an AR tool packet and outputting AR tool packet data of the image into the user data packet on a server, checking the file type, format, quantity and specification of the processed data packet at a manufacturing end, and finishing when the result is correct; when the result is wrong, returning the wrong result to the manufacturing end;
the specific steps are divided into three cases,
case a is where only the original video data and the tagged image for augmented reality recognition are included within the data packet, the original video data is input to a video compression program including but not limited to such as: QuickTime, After Effects, Final Cut, etc. process video programs. The mp4 file is output to the user data package by the server. Then, the marked image used for augmented reality recognition in the data packet is input into an AR tool packet and the AR tool packet data of the image is output to the user data packet on the server, and the AR tool packet is provided by an AR engine. Checking the file type, format, quantity and specification of the processed data packet at a manufacturing end, and finishing when the result is correct; and when the result is wrong, returning the wrong result to the manufacturing end.
Case b is where only raw picture data is included in the data packet, the raw picture data is input to an image processing program including, but not limited to, such as: adobe Photoshop, Affinity Photo, and other image processing programs. And (3) carrying out channel separation on the original picture used for the channel mapping model, removing the channel image of the non-display area, and outputting the channel image of the display area into the png file and then outputting the png file into the user data packet through the server. Then, the marked image for augmented reality in the data packet is input into an AR tool packet, and the AR tool packet data of the image is output to the user data packet on the server. Checking the file type, format, quantity and specification of the processed data packet at a manufacturing end, and finishing when the result is correct; and when the result is wrong, returning the wrong result to the manufacturing end.
Case c is where only the raw picture data and the two-dimensional scene image data are included in the data packet, and the raw picture data and the two-dimensional scene image data are input to an image processing program, which includes but is not limited to such as: image processing programs such as Adobe Photoshop, Affinity Photo, and the like; and inputting a jpeg file into an AR toolkit and outputting AR toolkit data of the image to the user data packet on the server, and outputting the channel image of the display area into a png file and then outputting the png file to the user data packet through the server. Checking the file type, format, quantity and specification of the processed data packet at a manufacturing end, and finishing when the result is correct; and when the result is wrong, returning the wrong result to the manufacturing end.
And step 3: inputting user information data on the handheld device, connecting the handheld device with the server through the communication unit to verify the user information, and ending when the result is an error; when the result is correct, the user data packet is output from the server: storing the three-dimensional scene model, the mapping model with the channel, the picture data with the channel, the compressed video data, the AR tool package data, the recognizable image data and the audio data into a storage unit;
and 4, step 4: the operation unit is used for calling the camera through the AR engine unit to acquire the real world continuous image, the recognizable image is placed in the range of the camera for acquiring the real world continuous image, the AR engine unit anchors the display position according to the spatial relation of the recognizable image in the real world, and the data packet content is output through the three-dimensional engine unit and displayed on the display unit.
The method comprises the following specific steps: and outputting the three-dimensional scene model in the data packet through a three-dimensional engine unit, mapping the picture data with the channel and/or the compressed video data on the mapping model with the channel, and playing audio data by using a loudspeaker of the device.
And 5: and controlling the playing, pausing, skipping and stopping of the data output by the three-dimensional engine by using the interaction unit. The method comprises the following specific steps: inputting a playing command to the interaction unit, and starting outputting a three-dimensional scene model, picture data with a channel and/or compressed video data by a three-dimensional engine, wherein the picture data with the channel and/or the compressed video data are subjected to mapping and audio data on a mapping model with the channel; and inputting a pause command to the interaction unit, pausing and stilling the three-dimensional scene model, the picture data with the channel and/or the compressed video data, and pasting the picture and the audio data on the map model with the channel by the three-dimensional engine. When a plurality of original picture data with channels and/or compressed video data are stored in the data packet, a skip command is input to the interaction unit, the three-dimensional engine continues to output a three-dimensional scene model and audio data and replaces the picture data with channels and/or the compressed video data with the mapping model with channels to map; inputting a stop command to the interaction unit, and stopping outputting the content by the three-dimensional engine; and moving the recognizable image away from the range of the real world continuous image acquired by the camera, and stopping the three-dimensional engine from inputting data.
The invention has the beneficial effects that: the method and the system for realizing the three-dimensional scene model with the channel in the augmented reality environment have the advantages of high conversion speed, simplified user operation steps, simple integral realization method, improvement of the fusion degree between original data and the three-dimensional scene model and the immersion feeling during browsing, and improvement of the augmented reality effect.
A user transmits raw material data to the database through the production end connecting server, then the data making is completed by the making end, and then the three-dimensional digital content generated by the processed raw data is browsed in the augmented reality environment by using the handheld device. For a further part of the description of the invention:
(1) the invention provides a flow framework for providing distributed production which is more convenient for multiple users, multiple data and high concurrency. Aiming at multiple users, the invention inputs user information data to a server through a production end and creates a data packet of the user in a database; the implementation steps of the system are simplified, the efficiency is higher, and the cost is reduced. Aiming at multiple data, the invention provides an implementation method for customizing the picture once or for multiple times or customizing the picture or the video based on the augmented reality display band transparent channel. Any picture or video can be displayed in a combined mode through a preset three-dimensional model scene, and one or more pictures and/or videos can be displayed in the same three-dimensional model scene.
(2) The invention not only provides a method and a system for realizing a user-defined picture with a transparent channel based on an augmented reality display, but also is suitable for videos, and provides detailed steps of the realization process.
(3) The invention provides a picture and/or video display method for a handheld device, wherein one or more user-defined original picture data and/or original video data are removed and compressed through a transparent channel, and are combined with a preset three-dimensional scene model to be applied to an augmented reality environment, so that the fusion degree between the original data and the three-dimensional scene model and the immersion feeling during browsing are improved.
(4) The invention further provides a perfecting method in use of the system, which is based on network communication and comprises a production end, a manufacturing end, a server and a handheld device system which are formed by a three-dimensional engine, an AR tool kit and a device basic unit.
(5) The method combines the existing picture or video with the three-dimensional model through processing and shows the picture or video in the augmented reality environment, and does not need to generate additional pictures or videos in real time.
(6) The invention realizes a method for displaying one or more pictures or video data with transparent channels in a three-dimensional scene on a handheld device through a camera, a display unit, an interaction unit, a storage unit, an operation unit, a three-dimensional engine unit and an AR engine unit; the system can realize end-to-end and end-to-device standard processes through a production end, a manufacturing end, a server and a handheld device.
Interpretation of terms:
the real world: real world refers to images taken from reality, such as the physical real world situation using electronic photo capture technology, e.g., video recording.
Augmented reality: the technology is a technology for calculating the position and the angle of a camera image in real time and adding a corresponding image, and aims to sleeve a virtual world on a screen in a real environment and interact with the virtual world.
Production end and manufacturing end: the computer is provided with a micro program with a network access function and is responsible for transmitting data to the server or acquiring data from the server.
Video compression program: including but not limited to such things as: video editing programs such as QuickTime, After Effects, Final Cut, etc. An image processing program: including but not limited to such things as: adobe Photoshop, Affinity Photo, and other image editing programs.
AR toolkit: including but not limited to augmented reality developer toolkits such as VuforiaAR, easylar, etc.
A three-dimensional engine: including but not limited to three-dimensional programs as applied to computers such as Untiy3D, Ureal Engine, etc. which are widely used in computers, particularly hand-held devices.
Three-dimensional scene model: the data packet composed of digital resources in a three-dimensional engine and a certain real world scene logical relationship comprises: three-dimensional models, maps, animations, special effects, audio, etc.
Detailed Description
Example 1
Step 1: a user inputs original manufacturing material data including one or more original picture data and/or original video data to a server through a production end, and the original manufacturing material data is applied to user information data of user identity information verification; the method comprises the following specific steps:
1) inputting user information data to a server through a production end, and creating a data packet of the user in a database; specifically, the user name, the gender, the selected three-dimensional scene, the mobile phone number, the commemorative day and the remark information are selected, and the user character information forms a unique user character string on the server;
2) outputting a two-dimensional scene image template in a database to a production end by a server, selecting a plurality of favorite original video data by a user through previewing and selecting the template, inputting the favorite original video data into a user data packet, and designating the original video data used for the mapping model with the channel as a marked image used for augmented reality identification;
step 2: the production end is connected with the server, original video data in a user data packet is output from the database, the original video data is input into a QuickTime video compression program, and after the original video data is compressed, the mp4 file is output into the user data packet through the server. Then, the marked image used for augmented reality recognition in the data packet is input into an AR tool packet and the AR tool packet data of the image is output to the user data packet on the server, and the AR tool packet is provided by an AR engine. Checking the file type, format, quantity and specification of the processed data packet at a manufacturing end, and finishing when the result is correct; and when the result is wrong, returning the wrong result to the manufacturing end.
And step 3: inputting user information data on the handheld device, connecting the handheld device with the server through the communication unit to verify the user information, and ending when the result is an error; when the result is correct, the user data packet is output from the server: storing the three-dimensional scene model, the mapping model with the channel, the compressed video data, the AR tool package data, the recognizable image data and the audio data into a storage unit;
and 4, step 4: the operation unit is used for calling the camera through the AR engine unit to acquire the real world continuous image, the recognizable image is placed in the range of the camera for acquiring the real world continuous image, the AR engine unit anchors the display position according to the spatial relation of the recognizable image in the real world, and the data packet content is output through the three-dimensional engine unit and displayed on the display unit.
The method comprises the following specific steps: and outputting the three-dimensional scene model in the data packet through a three-dimensional engine unit, mapping the compressed video data on the mapping model with the channel, and playing the audio data by using a loudspeaker of the device.
And 5: and controlling the playing, pausing, skipping and stopping of the data output by the three-dimensional engine by using the interaction unit. The method comprises the following specific steps: inputting a playing command to the interaction unit, and starting outputting a three-dimensional scene model, a compressed video data mapping on a mapping model with a channel and audio data by a three-dimensional engine; and inputting a pause command to the interaction unit, pausing and stopping the three-dimensional scene model by the three-dimensional engine, pasting the compressed video data on the pasting model with the channel, and audio data. When a plurality of compressed video data are stored in the data packet, a skip command is input to the interaction unit, the three-dimensional engine continues to output the three-dimensional scene model and the audio data and replaces the compressed video data to be pasted on the map model with the channel; inputting a stop command to the interaction unit, and stopping outputting the content by the three-dimensional engine; and moving the recognizable image away from the range of the real world continuous image acquired by the camera, and stopping the three-dimensional engine from inputting data.
Example 2
Step 1: a user inputs original manufacturing material data including one or more original picture data and/or original video data to a server through a production end, and the original manufacturing material data is applied to user information data of user identity information verification; the method comprises the following specific steps:
1) inputting user information data to a server through a production end, and creating a data packet of the user in a database; specifically, the user name, the gender, the selected three-dimensional scene, the mobile phone number, the commemorative day and the remark information are selected, and the user character information forms a unique user character string on the server;
2) outputting a two-dimensional scene image template in a database to a production end by a server, selecting a plurality of favorite original image data by a user through previewing and selecting the template, inputting the favorite original image data into a user data packet, and designating the original image data used for the mapping model with the channel as a marked image used for augmented reality identification;
step 2: and the manufacturing end is connected with the server, outputs original picture data in the user data packet from the database, inputs the original picture data into the Affinity Photo image processing program, performs channel separation on the original picture with the channel mapping model, removes the channel image in the non-display area, and outputs the channel image in the display area into the png file and then outputs the png file into the user data packet through the server. Then, the marked image for augmented reality in the data packet is input into an AR tool packet, and the AR tool packet data of the image is output to the user data packet on the server. Checking the file type, format, quantity and specification of the processed data packet at a manufacturing end, and finishing when the result is correct; and when the result is wrong, returning the wrong result to the manufacturing end.
And step 3: inputting user information data on the handheld device, connecting the handheld device with the server through the communication unit to verify the user information, and ending when the result is an error; when the result is correct, the user data packet is output from the server: storing the three-dimensional scene model, the mapping model with the channel, the picture data with the channel, the AR tool package data, the recognizable image data and the audio data into a storage unit;
and 4, step 4: the operation unit is used for calling the camera through the AR engine unit to acquire the real world continuous image, the recognizable image is placed in the range of the camera for acquiring the real world continuous image, the AR engine unit anchors the display position according to the spatial relation of the recognizable image in the real world, and the data packet content is output through the three-dimensional engine unit and displayed on the display unit.
The method comprises the following specific steps: and outputting the three-dimensional scene model in the data packet through a three-dimensional engine unit, mapping the picture data with the channel on the mapping model with the channel, and playing audio data by using a loudspeaker of the device.
And 5: and controlling the playing, pausing, skipping and stopping of the data output by the three-dimensional engine by using the interaction unit. The method comprises the following specific steps: inputting a playing command to the interaction unit, and starting outputting a three-dimensional scene model, a chartlet on the chartlet model with the channel of the picture data with the channel and audio data by the three-dimensional engine; and inputting a pause command to the interaction unit, pausing and stopping the three-dimensional scene model by the three-dimensional engine, pasting the picture on the picture model with the channel and audio data. When a plurality of original image data with channels are stored in the data packet, a skip command is input to the interaction unit, and the three-dimensional engine continues to output the three-dimensional scene model and the audio data and replaces the image data with channels to be pasted on the image model with channels; inputting a stop command to the interaction unit, and stopping outputting the content by the three-dimensional engine; and moving the recognizable image away from the range of the real world continuous image acquired by the camera, and stopping the three-dimensional engine from inputting data.
Example 3
Step 1: a user inputs original manufacturing material data including one or more original picture data and/or original video data to a server through a production end, and the original manufacturing material data is applied to user information data of user identity information verification; the method comprises the following specific steps:
1) inputting user information data to a server through a production end, and creating a data packet of the user in a database; specifically, the user name, the gender, the selected three-dimensional scene, the mobile phone number, the commemorative day and the remark information are selected, and the user character information forms a unique user character string on the server;
2) the server outputs a two-dimensional scene image template in the database to a production end, a user selects favorite original picture data and two-dimensional scene image data to input into a user data packet through previewing and selecting the template, and the original picture data used for the mapping model with the channel is designated as a marked image used for augmented reality identification;
step 2: the manufacturing end is connected with the server, original picture data and two-dimensional scene image data in a user data packet are output from the database, the original picture data and the two-dimensional scene image data are input into an Adobe Photoshop image processing program, the original picture with a channel mapping model is subjected to channel separation, a non-display area channel image is removed, a display area channel image and a two-dimensional scene image are merged and output to be a jpeg file, the AR tool packet data of the image is input into the AR tool packet, the AR tool packet data of the image is output into the user data packet on the server, and the display area channel image is output to be a png file and then is output into the user data packet through the server. Checking the file type, format, quantity and specification of the processed data packet at a manufacturing end, and finishing when the result is correct; and when the result is wrong, returning the wrong result to the manufacturing end.
And step 3: inputting user information data on the handheld device, connecting the handheld device with the server through the communication unit to verify the user information, and ending when the result is an error; when the result is correct, the user data packet is output from the server: storing the three-dimensional scene model, the mapping model with the channel, the picture data with the channel, the AR tool package data, the recognizable image data and the audio data into a storage unit;
and 4, step 4: the operation unit is used for calling the camera through the AR engine unit to acquire the real world continuous image, the recognizable image is placed in the range of the camera for acquiring the real world continuous image, the AR engine unit anchors the display position according to the spatial relation of the recognizable image in the real world, and the data packet content is output through the three-dimensional engine unit and displayed on the display unit.
The method comprises the following specific steps: and outputting the three-dimensional scene model in the data packet through a three-dimensional engine unit, mapping the picture data with the channel and the mapping model with the channel, and playing audio data by using a loudspeaker of the device.
And 5: and controlling the playing, pausing, skipping and stopping of the data output by the three-dimensional engine by using the interaction unit. The method comprises the following specific steps: inputting a playing command to the interaction unit, and starting outputting a three-dimensional scene model, a chartlet on the chartlet model with the channel of the picture data with the channel and audio data by the three-dimensional engine; and inputting a pause command to the interaction unit, pausing and stopping the three-dimensional scene model by the three-dimensional engine, pasting the picture on the picture model with the channel and audio data. When a plurality of original image data with channels are stored in the data packet, a skip command is input to the interaction unit, and the three-dimensional engine continues to output the three-dimensional scene model and the audio data and replaces the image data with channels to be pasted on the image model with channels; inputting a stop command to the interaction unit, and stopping outputting the content by the three-dimensional engine; and moving the recognizable image away from the range of the real world continuous image acquired by the camera, and stopping the three-dimensional engine from inputting data.