CN109089097A - A kind of object of focus choosing method based on VR image procossing - Google Patents
A kind of object of focus choosing method based on VR image procossing Download PDFInfo
- Publication number
- CN109089097A CN109089097A CN201810989621.5A CN201810989621A CN109089097A CN 109089097 A CN109089097 A CN 109089097A CN 201810989621 A CN201810989621 A CN 201810989621A CN 109089097 A CN109089097 A CN 109089097A
- Authority
- CN
- China
- Prior art keywords
- focus
- image
- merging
- priority
- candidate regions
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2411—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Human Computer Interaction (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
Abstract
This application discloses a kind of object of focus choosing methods based on VR image procossing, include the steps of determining that object of focus;High definition transmission object of focus simultaneously synchronizes non-high definition transmission background image.The application has reached reduction transmitted data on network amount, avoids transmission delay and meets the technical effect that client's viewing video is HD video.
Description
Technical field
This application involves VR image transmission technology field more particularly to a kind of object of focus selections based on VR image procossing
Method.
Background technique
Now to higher and higher to the requirement of resolution ratio in image transmitting process, especially in VR field of image transmission, because
Binocular is needed to export for VR, so the demand for just having double sized image to export in transmission.According to current feasible scheme,
If transmitting terminal then can be with the compressed video flowing of real-time Transmission using h264 decoding in receiving end using the compression of X264 coding
Data, and transmitted data on network amount can be effectively reduced.But still there is transmission high definition, extra-high clear video council in network transmission field
Cause delay aggravate the case where and can not instantaneity observation a possibility that, the transmission of video pressure of especially this 4K or 8K is then
It is bigger.
In addition, the more high then clarity of corresponding field of video encoding compression ratio is lower, it is unable to satisfy and can passes immediately
It is defeated but can allow client watch HD video a possibility that, be also unable to satisfy VR binocular output panoramic high-definition figure demand.
Summary of the invention
The application's is designed to provide a kind of object of focus choosing method based on VR image procossing, can reach reduction
Transmitted data on network amount avoids transmission delay and meets the technical effect that client's viewing video is HD video.
In order to achieve the above objectives, the application provides a kind of object of focus choosing method based on VR image procossing, including such as
Lower step: object of focus is determined;High definition transmission object of focus simultaneously synchronizes non-high definition transmission background image.
Preferably, the method for object of focus is determined specifically: receive the video that depth camera is captured or shot;Obtain view
The image information and deep image information of each frame image in frequency;It is all right in each frame image to be listed using R-CNN algorithm
As;Priority ranking, the high conduct object of focus of priority are carried out to each object.
Preferably, the minimum distance between camera carries out priority ranking according to each object, and minimum distance is smaller
Priority it is higher.
Preferably, type according to each object carries out priority ranking, and type includes at least personage, animal, object, people
The priority of object is higher than the priority of animal, and the priority of animal is higher than the priority of object.
Preferably, the method for all objects in image is listed using R-CNN algorithm specifically: generate every image more
A candidate regions;To each candidate regions, feature is extracted using depth network;The feature extracted from each candidate regions is sent into each
The SVM classifier of class is classified;After confirming candidate regions classification, further verifying is done by position refine;Pass through depth information
Candidate regions are analyzed, are provided one by one with the object as focus possibility.
Preferably, the method for candidate regions is generated specifically: divide the image into multiple zonules;Check each zonule,
Merge the neighboring community domain for meeting merging type, repeats to be merged into a regional location until whole image;Output is from being divided into
Merge the region occurred in complete process as candidate regions.
Preferably, merging type includes: the first kind: zonule similar in color;Second class: zonule similar in texture;
Third class: the small zonule of the gross area after merging;4th class: after merging, the gross area big cell of proportion in its BBOX
Domain.
Preferably, merging and meeting the method for the zonule of merging type is the first kind, the second class, third class and the 4th class four
Kind merging type is calculated simultaneously.
Preferably, small-region merging carries out in multiple color spaces, and it is empty that multiple color spaces include at least RGB color
Between, hsv color space and Lab color space.
Preferably, high definition transmission object of focus and the method for synchronizing non-high definition transmission background image are as follows: calculate each focus
The boundary of object;Calculate the average depth value, first edge profile and second edge profile of all object of focus;According to the first side
Edge profile and second edge profile fusion treatment object of focus and background image, and complete to transmit.
The application realize have the beneficial effect that: can reach reduce transmitted data on network amount, avoid transmission delay and
Meet client and watches the technical effect that video is HD video.
Detailed description of the invention
In order to illustrate the technical solutions in the embodiments of the present application or in the prior art more clearly, to embodiment or will show below
There is attached drawing needed in technical description to be briefly described, it should be apparent that, the accompanying drawings in the following description is only this
The some embodiments recorded in application can also be obtained according to these attached drawings other for those of ordinary skill in the art
Attached drawing.
Fig. 1 is a kind of flow chart of embodiment of object of focus choosing method based on VR image procossing;
Fig. 2 is a kind of flow chart of embodiment of method for determining object of focus;
Fig. 3 is a kind of flow chart of embodiment of method that all objects in image are listed using R-CNN algorithm;
Fig. 4 is a kind of flow chart of embodiment of method for generating candidate regions;
Fig. 5 is that high definition transmits object of focus and synchronizes a kind of process of embodiment of method that non-high definition transmits background image
Figure.
Specific embodiment
With reference to the attached drawing in the embodiment of the present invention, technical solution in the embodiment of the present invention carries out clear, complete
Ground description, it is clear that described embodiments are some of the embodiments of the present invention, instead of all the embodiments.Based on the present invention
In embodiment, those skilled in the art's every other embodiment obtained without making creative work, all
Belong to the scope of protection of the invention.
As shown in Figure 1, the present invention provides a kind of object of focus choosing method based on VR image procossing, including walk as follows
It is rapid:
S110: object of focus is determined.
Further, as shown in Fig. 2, the method for determining object of focus specifically:
S210: the video that depth camera is captured or shot is received.
Specifically, staff using depth camera capture or shoot video, and by the video be uploaded to computer into
Row processing.Wherein, the video for capturing or shooting can be 4K video, but be not limited only to 4K video, can also be the view lower than 4K
Frequency or the video higher than 4K, the present embodiment are preferably 4K video, high resolution.
S220: the color image information and deep image information of each frame image in video are obtained.
Specifically, color image information is two dimensional image, deep image information is that each pixel is corresponding on two dimensional image
The distance between camera and real-world object point obtain deep image information using equipment such as Kinect or RealScene.
S230: all objects in each frame image are listed using R-CNN algorithm.
Specifically, R-CNN algorithm is existing deep learning algorithm, participated in when the object that can be identified is according to preparatory training
Object type determines, such as the training of human of property, cat, dog, four class of chicken, then in computing object list, can only provide above-mentioned
Four class objects.
Further, as shown in figure 3, the method for listing all objects in image using R-CNN algorithm specifically:
S310: every image is generated into multiple candidate regions.Specifically, candidate regions are 1K~2K.
Further, as shown in figure 4, the method for generating candidate regions specifically:
S410: multiple zonules are divided the image into.
S420: checking each zonule, merges the neighboring community domain for meeting merging type, repeats to merge until whole image
At a regional location.
Specifically, merging type has following four kinds:
The first kind: block of cells similar in color (color histogram).Specifically, the range of the first kind is color space model
It encloses.
Second class: zonule similar in texture (histogram of gradients).Specifically, two classes, which are that gradient is all similar, participates in meter
It calculates.
Third class: the small zonule of the gross area after merging.Guarantee that the scale of union operation is more uniform, avoids an area Ge great
Other zonules " are eaten up " successively in domain.
Specifically, merging calculating according to 1K~2K candidate regions.Equipped with region a-b-c-d-e-f-g-h.Preferably
Merging mode is: ab-cd-ef-gh- > abcd-efgh- > abcdefgh.
4th class: after merging, the gross area big cell of proportion in its BBOX (BoundingBox, image size)
Domain.
Preferably, in the method for the generation candidate regions of the present embodiment, the first kind, the second class, third class and the 4th class are needed
It to be calculated simultaneously.The regular shape for merging rear region is to be merged automatically according to algorithm and calculated.Provincial characteristics after merging
Can be directly from subregion feature calculation, fast speed.(including color histogram, texture are straight for provincial characteristics after merging
Fang Tu, area and position) can be directly from the provincial characteristics calculating of subregion, fast speed.
S430: it exports from the region for merging and occurring in complete process is divided into as candidate regions.
Specifically, referring to all cells including being partitioned into from the region for merging and occurring in complete process is divided into
Domain, zonule merge in all combined region and merged later region.For example, being equipped with region a-b-c-d-
e-f-g-h.Preferred merging mode is: ab-cd-ef-gh- > abcd-efgh- > abcdefgh, the then candidate regions exported have area
Domain a, region b, region c, region d, region e, region f, region g, region h, region ab, region cd, region ef, region gh, area
Domain abcd, region efgh, region abcdefgh.
Further, to avoid not omitting candidate regions, aforesaid operations carry out (RGB color simultaneously in multiple color spaces
Space, hsv color space, Lab color space etc.).In a color space, the first kind, the second class, third class and are used
Four classes merge again after carrying out various combination.The region after the repeat region in all areas after removal merging is as candidate
Area's output.
Continuing with referring to Fig. 3, S320: to each candidate regions, extracting feature using depth network.
Specifically, depth network is existing feature point extraction technology, its characteristic value can be calculated to a region,
It is carried out on tri- color spaces of RGB, HSV, Lab respectively when calculating.
Further, first preprocessed data is needed before extracting feature.
Specifically, candidate regions are normalized into same size, this is preferably dimensioned to be 227 × 227.
S330: the SVM classifier that the feature extracted from each candidate regions is sent into every one kind is classified.
Specifically, class is manually demarcated.For example, one class of setting is behaved, the image of this feature of people will be manually calibrated
It is sent into depth network to be trained, the object that training result is demarcated is also that people is this kind of, after training, if scheming when use
As the similar feature of upper appearance then determines that, for people's this kind, the object that do not demarcate will not be then correctly validated.
S340: after confirmation candidate regions classification, further verifying is done by position refine.
Specifically, carrying out refine using candidate frame of the linear ridge regression device to the candidate regions of every one kind.Depth network pool5
4096 dimensional features of layer export as the zooming and panning in the direction xy, and training sample is determined as in the candidate frame of this class and true value weight
Folded area is greater than 0.6 candidate frame.
S350: analyzing candidate regions by depth information, is provided one by one with the object as focus possibility.
Specifically, requiring to calculate possible object before candidate regions merge and after merging, to pair with same possibility
The merging candidate regions of elephant are recalculated, and all objects having as focus possibility are listed.
Continuing with referring to fig. 2, after determining all objects in a width figure, priority row S240: is carried out to each object
Sequence, the high conduct object of focus of priority.
Further, as one embodiment, the minimum distance between each object and camera is recorded, by each object
Classification and ordination is carried out according to distance value from the near to the distant, the object nearest from camera becomes the highest priority of object of focus.
Specifically, using preparatory trained depth void seat model calculate all objects in each image and camera it
Between distance.Deep learning model is that will mark focus pair by being manually labeled to the object of focus in each image
The data of elephant input in existing deep learning model for training.For example, in the first artificial mark: the first object distance is taken the photograph
As 10 meters of head, all animals from 10 meters of camera over, then the first personage is object of focus.In the second artificial mark:
The first object is apart from 10 meters of camera, and unknown form biology is apart from 5 meters of camera, then unknown form biology is object of focus.?
During third manually marks: the first object is apart from 10 meters of camera, and the first bear and the second bear are respectively apart from 6 meters and 8 meters of camera, and
One personage is object of focus.
As another embodiment, classification row can also be carried out according to the type of personage, animal, object for each object
Sequence.Preferably, the priority of personage is higher than the priority of animal, and the priority of animal is higher than the priority of object.Priority is high
Conduct object of focus.
Preferably, when determining object of focus, first judge the type of object, then judge the distance of object distance camera.Tool
Body, in one image, no matter the distance of object distance camera is how many, and personage, unknown form biology and animal are simultaneously
When appearance, personage is higher than unknown form biology as the priority of object of focus, and unknown form biology is as the excellent of object of focus
First grade is higher than animal.
Continuing with referring to Fig. 1, S120: high definition transmission object of focus simultaneously synchronizes non-high definition transmission background image.
Further, as shown in figure 5, high definition transmits object of focus and synchronizes the method for non-high definition transmission background image such as
Under:
S510: the boundary of each object of focus is calculated.
Specifically, calculate the boundary of the object of focus by the depth information of each object of focus, depth information value for
It is to have an apparent excessive boundary for some object, depth information value can be obtained by depth network trained in advance.
For example, when people's object distance is from camera 10m, from being exactly between camera 9.8m to 10.2m around personage
The depth information of this people, if it exceeds this threshold value is exactly the scenes such as the mountain forest building of other objects or distant place.
S520: the average depth value, first edge profile and second edge profile of all object of focus are calculated.
Specifically, an average depth value is calculated according to the depth information of all object of focus, in this depth value model
Enclose it is interior each pixel is carried out calculating threshold value using existing depth network be greater than the boundary of 20cm range as object
Boundary, and first edge profile and second edge profile are calculated, first edge profile is the edge wheel of approximate object of focus
Exterior feature, second edge profile are greater than first edge profile and first edge profile are enclosed in second edge profile.
S530: it according to first edge profile and second edge profile fusion treatment object of focus and background image, and completes
Transmission.
Specifically, video and object of focus synchronous transfer that depth camera is captured or shot give VR client.First
Original image is rendered into a non-high definition figure as background image in VR client, then corresponds to focus pair originally on background image
The object of focus figure that one layer of high definition is rendered on the region of elephant, is gradually melted to second edge profile according to first edge profile
It closes, integration region is formed, with comfortable visual impression when image is shown after guarantee processing.
The application realize have the beneficial effect that: can reach reduce transmitted data on network amount, avoid transmission delay and
Meet client and watches the technical effect that video is HD video.
Although the preferred embodiment of the application has been described, it is created once a person skilled in the art knows basic
Property concept, then additional changes and modifications may be made to these embodiments.So it includes excellent that the following claims are intended to be interpreted as
It selects embodiment and falls into all change and modification of the application range.Obviously, those skilled in the art can be to the application
Various modification and variations are carried out without departing from spirit and scope.If in this way, these modifications and variations of the application
Belong within the scope of the claim of this application and its equivalent technologies, then the application is also intended to encompass these modification and variations and exists
It is interior.
Claims (10)
1. a kind of object of focus choosing method based on VR image procossing, which comprises the steps of:
Determine object of focus;
High definition transmission object of focus simultaneously synchronizes non-high definition transmission background image.
2. object of focus choosing method according to claim 1, which is characterized in that the method for determining object of focus is specific
Are as follows:
Receive the video that depth camera is captured or shot;
Obtain the image information and deep image information of each frame image in video;
All objects in each frame image are listed using R-CNN algorithm;
Priority ranking, the high conduct object of focus of priority are carried out to each object.
3. object of focus choosing method according to claim 2, which is characterized in that according to each object between camera
Minimum distance carry out priority ranking, the smaller priority of minimum distance is higher.
4. object of focus choosing method according to claim 2, which is characterized in that type according to each object carries out excellent
First grade sequence, type include at least personage, animal, object, and the priority of personage is higher than the priority of animal, the priority of animal
Higher than the priority of object.
5. object of focus choosing method according to claim 2, which is characterized in that listed in image using R-CNN algorithm
All objects method specifically:
Every image is generated into multiple candidate regions;
To each candidate regions, feature is extracted using depth network;
The SVM classifier that the feature extracted from each candidate regions is sent into every one kind is classified;
After confirming candidate regions classification, further verifying is done by position refine;
Candidate regions are analyzed by depth information, are provided one by one with the object as focus possibility.
6. object of focus choosing method according to claim 5, which is characterized in that the method for generating candidate regions specifically:
Divide the image into multiple zonules;
It checks each zonule, merges the neighboring community domain for meeting merging type, repeat to be merged into an area until whole image
Domain position;
It exports from the region for merging and occurring in complete process is divided into as candidate regions.
7. object of focus choosing method according to claim 6, which is characterized in that the merging type includes:
The first kind: zonule similar in color;
Second class: zonule similar in texture;
Third class: the small zonule of the gross area after merging;
4th class: after merging, the gross area big zonule of proportion in its BBOX.
8. object of focus choosing method according to claim 7, which is characterized in that merge the zonule for meeting merging type
Method be the first kind, the second class, four kinds of merging type of third class and the 4th class simultaneously calculated.
9. object of focus choosing method according to claim 6, which is characterized in that small-region merging is in multiple color spaces
Middle progress, multiple color spaces include at least RGB color, hsv color space and Lab color space.
10. object of focus choosing method according to claim 6, which is characterized in that high definition transmits object of focus and synchronization
The method of non-high definition transmission background image are as follows:
Calculate the boundary of each object of focus;
Calculate the average depth value, first edge profile and second edge profile of all object of focus;
According to first edge profile and second edge profile fusion treatment object of focus and background image, and complete to transmit.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810989621.5A CN109089097A (en) | 2018-08-28 | 2018-08-28 | A kind of object of focus choosing method based on VR image procossing |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810989621.5A CN109089097A (en) | 2018-08-28 | 2018-08-28 | A kind of object of focus choosing method based on VR image procossing |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109089097A true CN109089097A (en) | 2018-12-25 |
Family
ID=64794931
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810989621.5A Pending CN109089097A (en) | 2018-08-28 | 2018-08-28 | A kind of object of focus choosing method based on VR image procossing |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109089097A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021073268A1 (en) * | 2019-10-15 | 2021-04-22 | 北京市商汤科技开发有限公司 | Augmented reality data presentation method and apparatus, electronic device, and storage medium |
CN113297876A (en) * | 2020-02-21 | 2021-08-24 | 佛山市云米电器科技有限公司 | Motion posture correction method based on intelligent refrigerator, intelligent refrigerator and storage medium |
WO2022184569A1 (en) * | 2021-03-01 | 2022-09-09 | Signify Holding B.V. | Displaying an aggregation of data in dependence on a distance to a closest device in an image |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103517072A (en) * | 2012-06-18 | 2014-01-15 | 联想(北京)有限公司 | Video communication method and video communication equipment |
CN103999096A (en) * | 2011-12-16 | 2014-08-20 | 英特尔公司 | Reduced image quality for video data background regions |
CN104349142A (en) * | 2014-11-03 | 2015-02-11 | 南京航空航天大学 | Layered representation-based unmanned plane video adaptive transmission method |
CN104427291A (en) * | 2013-08-19 | 2015-03-18 | 华为技术有限公司 | Image processing method and device |
CN104732210A (en) * | 2015-03-17 | 2015-06-24 | 深圳超多维光电子有限公司 | Target human face tracking method and electronic equipment |
CN105608716A (en) * | 2015-12-21 | 2016-05-25 | 联想(北京)有限公司 | Information processing method and electronic equipment |
CN106530265A (en) * | 2016-11-08 | 2017-03-22 | 河海大学 | Adaptive image fusion method based on chromaticity coordinates |
CN107409203A (en) * | 2015-05-27 | 2017-11-28 | 谷歌公司 | Reduce the method and apparatus of the spherical video bandwidth of user's headphone |
CN107439010A (en) * | 2015-05-27 | 2017-12-05 | 谷歌公司 | The spherical video of streaming |
CN107547804A (en) * | 2017-09-21 | 2018-01-05 | 北京奇虎科技有限公司 | Realize the video data handling procedure and device, computing device of scene rendering |
CN107610149A (en) * | 2017-09-25 | 2018-01-19 | 北京奇虎科技有限公司 | Image segmentation result edge optimization processing method, device and computing device |
US20180075582A1 (en) * | 2016-09-15 | 2018-03-15 | International Business Machines Corporation | Image processing |
-
2018
- 2018-08-28 CN CN201810989621.5A patent/CN109089097A/en active Pending
Patent Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103999096A (en) * | 2011-12-16 | 2014-08-20 | 英特尔公司 | Reduced image quality for video data background regions |
CN103517072A (en) * | 2012-06-18 | 2014-01-15 | 联想(北京)有限公司 | Video communication method and video communication equipment |
CN104427291A (en) * | 2013-08-19 | 2015-03-18 | 华为技术有限公司 | Image processing method and device |
CN104349142A (en) * | 2014-11-03 | 2015-02-11 | 南京航空航天大学 | Layered representation-based unmanned plane video adaptive transmission method |
CN104732210A (en) * | 2015-03-17 | 2015-06-24 | 深圳超多维光电子有限公司 | Target human face tracking method and electronic equipment |
CN107409203A (en) * | 2015-05-27 | 2017-11-28 | 谷歌公司 | Reduce the method and apparatus of the spherical video bandwidth of user's headphone |
CN107439010A (en) * | 2015-05-27 | 2017-12-05 | 谷歌公司 | The spherical video of streaming |
CN105608716A (en) * | 2015-12-21 | 2016-05-25 | 联想(北京)有限公司 | Information processing method and electronic equipment |
US20180075582A1 (en) * | 2016-09-15 | 2018-03-15 | International Business Machines Corporation | Image processing |
CN106530265A (en) * | 2016-11-08 | 2017-03-22 | 河海大学 | Adaptive image fusion method based on chromaticity coordinates |
CN107547804A (en) * | 2017-09-21 | 2018-01-05 | 北京奇虎科技有限公司 | Realize the video data handling procedure and device, computing device of scene rendering |
CN107610149A (en) * | 2017-09-25 | 2018-01-19 | 北京奇虎科技有限公司 | Image segmentation result edge optimization processing method, device and computing device |
Non-Patent Citations (1)
Title |
---|
SHENXIAOLU1984: "【目标检测】RCNN算法详解", 《HTTPS://BLOG.CSDN.NET/SHENXIAOLU1984/ARTICLE/DETAILS/51066975》 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021073268A1 (en) * | 2019-10-15 | 2021-04-22 | 北京市商汤科技开发有限公司 | Augmented reality data presentation method and apparatus, electronic device, and storage medium |
CN113297876A (en) * | 2020-02-21 | 2021-08-24 | 佛山市云米电器科技有限公司 | Motion posture correction method based on intelligent refrigerator, intelligent refrigerator and storage medium |
WO2022184569A1 (en) * | 2021-03-01 | 2022-09-09 | Signify Holding B.V. | Displaying an aggregation of data in dependence on a distance to a closest device in an image |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Lee et al. | From big to small: Multi-scale local planar guidance for monocular depth estimation | |
TWI524734B (en) | Method and device for generating a depth map | |
CN106157307B (en) | A kind of monocular image depth estimation method based on multiple dimensioned CNN and continuous CRF | |
EP2639761B1 (en) | Depth information generator, depth information generation method, and stereoscopic image converter | |
CN101416520B (en) | Efficient encoding of multiple views | |
CN102741879B (en) | Method for generating depth maps from monocular images and systems using the same | |
US11840357B2 (en) | Method and device for dual-light image integration, and unmanned aerial vehicle | |
CN100571335C (en) | Image syncretizing effect real-time estimating method and device based on pixel space relativity | |
CN109089097A (en) | A kind of object of focus choosing method based on VR image procossing | |
CN110084757A (en) | A kind of infrared depth image enhancement method based on generation confrontation network | |
CN103430210A (en) | Information processing system, information processing device, imaging device, and information processing method | |
CN110335222B (en) | Self-correction weak supervision binocular parallax extraction method and device based on neural network | |
CN113159028A (en) | Saliency-aware image cropping method and apparatus, computing device, and storage medium | |
Cong et al. | Discrete haze level dehazing network | |
CN103260043A (en) | Binocular stereo image matching method and system based on learning | |
CN106934828A (en) | Depth image processing method and depth image processing system | |
CN103167306A (en) | Device and method for extracting high-resolution depth map in real time based on image matching | |
CN116665092A (en) | Method and system for identifying sewage suspended matters based on IA-YOLOV7 | |
WO2021205068A1 (en) | A method, an apparatus and a computer program product for volumetric video coding | |
CN109191381A (en) | A kind of method and system of calibration focus processing image | |
CN113628121B (en) | Method and device for processing and training multimedia data | |
CN112633198A (en) | Picture processing method and device, storage medium and electronic device | |
WO2019185983A1 (en) | A method, an apparatus and a computer program product for encoding and decoding digital volumetric video | |
Bae et al. | Efficient and scalable view generation from a single image using fully convolutional networks | |
EP4311234A1 (en) | A method, an apparatus and a computer program product for volumetric video synchronization using spatial neural attention network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20181225 |
|
RJ01 | Rejection of invention patent application after publication |