CN106997583A - A kind of underwater robot adaptive image enhancement and feature extracting method - Google Patents

A kind of underwater robot adaptive image enhancement and feature extracting method Download PDF

Info

Publication number
CN106997583A
CN106997583A CN201710141451.0A CN201710141451A CN106997583A CN 106997583 A CN106997583 A CN 106997583A CN 201710141451 A CN201710141451 A CN 201710141451A CN 106997583 A CN106997583 A CN 106997583A
Authority
CN
China
Prior art keywords
point
image
value
key point
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN201710141451.0A
Other languages
Chinese (zh)
Inventor
曾庆军
王建明
王倩
戴晓强
索文杰
张寿超
印少卿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu Diyi Group Co Ltd
Original Assignee
Jiangsu Diyi Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangsu Diyi Group Co Ltd filed Critical Jiangsu Diyi Group Co Ltd
Priority to CN201710141451.0A priority Critical patent/CN106997583A/en
Publication of CN106997583A publication Critical patent/CN106997583A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/90Dynamic range modification of images or parts thereof
    • G06T5/94Dynamic range modification of images or parts thereof based on local image properties, e.g. for local contrast enhancement
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/40Image enhancement or restoration using histogram techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20004Adaptive image processing
    • G06T2207/20012Locally adaptive

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a kind of underwater robot adaptive image enhancement in image processing field and feature extracting method, comprise the following steps:1)The video camera carried using underwater robot is carried out image taking and stored;2)Image procossing is carried out using CLAHE technologies;3)Image characteristics extraction is carried out using SIFT algorithms;4)Fixation and recognition is carried out using feature, the underwater identification of underwater robot and positional accuracy are improved by using the present invention, available in marine resources exploratory development.

Description

A kind of underwater robot adaptive image enhancement and feature extracting method
Technical field
The present invention relates to a kind of underwater robot, more particularly to a kind of underwater robot method for adaptive image enhancement.
Background technology
China is an ocean big country, and the exploration and exploitation of marine resources have weight to the national economy and national defense safety of China The meaning and value wanted.Energy has been put into machine under water by the research with the mankind to ocean, increasing domestic and foreign scholars The research of device people.Underwater robot has in fields such as exploration of ocean resources exploitation, marine ecology monitoring, and military affairs extensively should With.Present underwater robot is provided with greatly optical imagery and image procossing identifying system, and visible images are machines under water One of primary information resource of device people.But limited by severe imaging circumstances under water, the underwater picture generally existing got The unfavorable factors such as contrast is low, fuzzy, colour cast.The decline of image visual quality can have a strong impact on subsequent characteristics and extract and target knowledge The performance of process such as not.Therefore, original underwater picture visual quality is improved by image processing techniques, so as to preferably be schemed As feature extraction has important scientific meaning, the concern of more and more researchers is obtained, correlative study is increasing.
There are the appearance of some achievements, existing underwater picture processing method to underwater picture processing domestic and foreign scholars in recent years It is classified largely into image enhaucament and image recovers two classes, because the various methods that image recovers is required for the priori for using environment to know Know, such as scene depth, locus, optical length, illumination, temperature and electrical conductivity so that during the processing of Image restoration Between considerably beyond image enchancing method.Therefore image enchancing method has simple and quick, can effectively strengthen the advantage of image.And In terms of image enhaucament, the histogram equalization (abbreviation HE) based on airspace enhancement is a kind of the most frequently used method, and HE methods are examined The statistical distribution of each passage value is considered, but has not accounted for positional information, easily strengthened noise and image detail together.Cause This, in order to overcome the shortcomings of HE methods, is used under water using a kind of contrast limited adaptive histogram equalization CLAHE technologies Image procossing, its basic thought is the height by limiting local histogram, so as to limit noise amplification and weaken local Crossing for contrast strengthens.This method can not only stretch the dynamic range of gradation of image, some prominent details, while can pass through Limit contrast to reduce the distortion phenomenon of image, so as to obtain visual effect better image.
In terms of image characteristics extraction, SIFT (Scale Invariant Feature Transform), i.e. Scale invariant Eigentransformation, by the continuous exploration of researcher, SIFT description son tools are confirmed after a variety of description are carried out with experimental analysis There is most strong robustness, can be rotated in image, scale, translate, illumination, carried under the conditions of affine transformation and partial occlusion etc. Take stable feature.Based on above characteristic, SIFT is obtained for extensively in the fields such as image characteristics extraction, target recognition and tracking General application, especially for due to underwater complex environment causes shooting image distortion the problem of, can be good at using SIFT algorithms Solve.Therefore, SIFT algorithms turn into one of domestic and international image procossing and the focus in computer vision research field.
The content of the invention
The purpose of the present invention be for underwater picture due to there are numerous suspended materials in water, light ray energy decay and The scattering process of pipe, causing image blurring quality to reduce this phenomenon, there is provided a kind of underwater robot adaptive image enhancement And feature extracting method, to ensure that the view data collected becomes apparent from accurately.
The object of the present invention is achieved like this:A kind of underwater robot adaptive image enhancement and feature extracting method, Comprise the following steps:
1) video camera carried using underwater robot is carried out image taking and stored;
2) image procossing is carried out using CLAHE technologies;
3) image characteristics extraction is carried out using SIFT algorithms;
4) fixation and recognition is carried out using feature.
Be used as the present invention further restriction, step 2) specific method be:
2-1) by step 1) artwork be divided into k size be m × n continuous nonoverlapping subregion.The value of [m n] is determined Determine the details enhancing degree of image, generally its value is smaller, and enhancing effect is weaker;
2-2) calculate the grey level histogram per sub-regions;
2-3) determine per sub-regions grey level histogram " shearing " value;
The average value N that number of pixels in per sub-regions is evenly distributed to each gray level is calculated by formula (1) firstaver
Wherein, NgrayIt is the quantity of gray level in subregion;It is the pixel count in subregion x-axis direction;It is The pixel count in subregion y-axis direction.Actual " shearing " value is NCL
NCL=Nclip·Naver (2)
Wherein, NclipIt is interception limit coefficient, it is meant that the pixel count that is included of the limitation each gray level of subregion not More than the N of mean pixel numberclipTimes;
The grey level histogram per sub-regions 2-4) is sheared, the number of pixels that remainder is come is re-assigned to each histogrammic each In gray level;
If the sum of all pixels being sheared is NΣclip, then obtain the shearing pixel count N that each gray level is divided equallyacp
2-5) after redistributing, pixel remainder is shear off for NLP, the step value of pixel is allocated by as follows Formula is provided:
It is 0 to carry out cycle assignment to pixel count by above-mentioned step value since minimal gray level to residual pixel, final To new histogram;
2-6) grey level histogram after contrast " shearing " in every sub-regions is equalized;
2-7) according to new histogram distribution, gray scale bilinear interpolation is carried out to image, enhanced result images are generated.
Be used as the present invention further restriction, step 3) specific method be:
3-1) read in step 2-7) in obtained enhanced result images, generate Gaussian difference scale space:For spy Phenomenon a little unstable on metric space is levied, this problem is solved using Gaussian difference scale space, it is by different scale Gaussian difference pyrene and image carry out convolution and obtain, such as formula (5):
D (x, y, σ)=(G (x, y, k σ)-G (x, y, σ)) * I (x, y) (5)
Wherein, σ is that normal distribution standard is poor, and * is convolution symbol, and I (x, y) represents input picture, and G (x, y, σ) is yardstick Variable Gaussian function, (x, y) represent image pixel position, k be the adjacent yardstick of each two between be separated by linetype scale because Son;
3-2) metric space extreme point is detected:The acquisition of extreme point is needed by sampled point and certain limit adjacent thereto Interior point is compared, the scope be under same yardstick around 8 points and totally 26 points of the 2*9 in two neighborhoods up and down, see Whether the point is maximum or minimum value;
3-3) key point is accurately positioned:Being accurately positioned for key point can improve its matching accuracy, therefore, need to be to key Point does following processing:
1. position and the dimensional information (reaching sub-pixel) of key point are determined by the fitting of three-dimensional quadratic function;
2. by setting suitable threshold value (typically taking 0.03) to filter out the key point of low contrast;
3. given threshold (generally the taking 10) key point unstable to skirt response is filtered out;
3-4) determine the main gradient direction of key point:According to the scale-value σ for characteristic point and this feature point having been detected by, obtain To the Gaussian image close to this scale-value:
L (x, y, σ)=G (x, y, σ) * I (x, y) (6)
Gradient magnitude and the direction that the key point can then be obtained are as follows:
M (x, y)=[(L (x+1, y)-L (x-1, y))2+(L(x,y+1)-L(x,y-1))2]1/2 (7)
Wherein, L (x, y, σ) represents the metric space of image, and m (x, y) and θ (x, y) are respectively gaussian pyramid (x, y) place Gradient magnitude and direction, direction used in L be each crucial point scale under direction;
Obtain after m (x, y) and θ (x, y), sampled in the contiguous range centered on key point, it is counted after sampling Pixel gradient direction, is then represented with histogram.Histogrammic peak value represents the principal direction of the crucial neighborhood of a point gradient, as The direction of the key point;So far, feature point detection is finished, and each characteristic point is believed comprising position, yardstick and the aspect of direction 3 Breath;
3-5) generation feature point description:Feature point description is calculated by carrying out image block to key point peripheral region Histogram of gradients in each piece, generates unique vector descriptor, and specific method is:16 × 16 are taken centered on key point Sample area, divides the area into 16 seed points that size is 4 × 4, calculates 8 directions respectively in each seed point region Gradient accumulated value, draw gradient direction direction histogram;The Feature Descriptor of final 128 dimensions for obtaining one 4 × 4 × 8.
Compared with prior art, the beneficial effects of the present invention are:
1) the contrast limited adaptive histogram equalization CLAHE methods employed in this patent are schemed relative to others As for processing method, by stretching the dynamic range of gradation of image, prominent image detail, while noise amplification is limited, to water Picture quality can be improved after hypograph processing, and keeps the feature in terms of image detail;
2) the Scale invariant features transform SIFT methods that this patent is used relative to other image characteristic extracting methods come Say, the feature extracted has translation, rotates and scale the characteristic such as constant, while also having to illumination, noise, small visual angle change Certain adaptability, in addition, verification and measurement ratio of the characteristic vector extracted under target part circumstance of occlusion is also at a relatively high, it is a kind of The algorithm of the extraction local feature of robust;
3) image procossing and feature extracting method employed in this patent are under water in the application of robot, under water Image causes image blurring quality due to there is numerous suspended materials, light ray energy decay and the scattering process of pipe in water Reducing this phenomenon, there is provided a kind of efficiently feasible method, it is possible to increase the underwater identification of underwater robot and positioning are accurate Exactness.The present invention can be used in marine resources exploratory development.
Brief description of the drawings
Fig. 1 is workflow diagram of the present invention.
Fig. 2 carries out image processing method flow chart for CLAHE methods in the present invention.
Fig. 3 carries out image characteristic extracting method flow chart for SIFT methods in the present invention.
Embodiment
A kind of underwater robot adaptive image enhancement and feature extracting method as Figure 1-3, comprise the following steps:
1) video camera carried using underwater robot is carried out image taking and stored;
2) image procossing is carried out using CLAHE technologies, specific method is:
2-1) by step 1) artwork be divided into k size be m × n continuous nonoverlapping subregion.The value of [m n] is determined Determine the details enhancing degree of image, generally its value is smaller, and enhancing effect is weaker;
2-2) calculate the grey level histogram per sub-regions;
2-3) determine per sub-regions grey level histogram " shearing " value;
The average value N that number of pixels in per sub-regions is evenly distributed to each gray level is calculated by formula (1) firstaver
Wherein, NgrayIt is the quantity of gray level in subregion;It is the pixel count in subregion x-axis direction;It is The pixel count in subregion y-axis direction.Actual " shearing " value is NCL
NCL=Nclip·Naver(2)
Wherein, NclipIt is interception limit coefficient, it is meant that the pixel count that is included of the limitation each gray level of subregion not More than the N of mean pixel numberclipTimes;
The grey level histogram per sub-regions 2-4) is sheared, the number of pixels that remainder is come is re-assigned to each histogrammic each In gray level;
If the sum of all pixels being sheared is NΣclip, then obtain the shearing pixel count N that each gray level is divided equallyacp
2-5) after redistributing, pixel remainder is shear off for NLP, the step value of pixel is allocated by as follows Formula is provided:
It is 0 to carry out cycle assignment to pixel count by above-mentioned step value since minimal gray level to residual pixel, final To new histogram;
2-6) grey level histogram after contrast " shearing " in every sub-regions is equalized;
2-7) according to new histogram distribution, gray scale bilinear interpolation is carried out to image, enhanced result images are generated;
3) image characteristics extraction is carried out using SIFT algorithms, specific method is:
3-1) read in step 2-7) in obtained enhanced result images, generate Gaussian difference scale space:For spy Phenomenon a little unstable on metric space is levied, this problem is solved using Gaussian difference scale space, it is by different scale Gaussian difference pyrene and image carry out convolution and obtain, such as formula (5):
D (x, y, σ)=(G (x, y, k σ)-G (x, y, σ)) * I (x, y) (5)
Wherein, σ is that normal distribution standard is poor, and * is convolution symbol, and I (x, y) represents input picture, and G (x, y, σ) is yardstick Variable Gaussian function, (x, y) represent image pixel position, k be the adjacent yardstick of each two between be separated by linetype scale because Son;
3-2) metric space extreme point is detected:The acquisition of extreme point is needed by sampled point and certain limit adjacent thereto Interior point is compared, the scope be under same yardstick around 8 points and totally 26 points of the 2*9 in two neighborhoods up and down, see Whether the point is maximum or minimum value;
3-3) key point is accurately positioned;Key point:Because the extreme point detected is the extreme point of discrete space, and Gauss Difference operator is more sensitive to border, can produce very strong skirt response, so in the extreme point detected above, it may appear that Some unstable, low contrast extreme points;In order to remove these pseudo- extreme points, improve and adopted in noise resisting ability, SIFT algorithms The extreme point of low contrast is got rid of with the method for curve matching, and utilizes the unstable marginal point of Hessian matrix removing.It is surplus Under point be key point;
Being accurately positioned for key point can improve its matching accuracy, therefore, following processing need to be done to key point:
1. position and the dimensional information (reaching sub-pixel) of key point are determined by the fitting of three-dimensional quadratic function;
2. by setting suitable threshold value (typically taking 0.03) to filter out the key point of low contrast;
3. given threshold (generally the taking 10) key point unstable to skirt response is filtered out;
3-4) determine the main gradient direction of key point:According to the scale-value σ for characteristic point and this feature point having been detected by, obtain To the Gaussian image close to this scale-value:
L (x, y, σ)=G (x, y, σ) * I (x, y) (6)
Gradient magnitude and the direction that the key point can then be obtained are as follows:
M (x, y)=[(L (x+1, y)-L (x-1, y))2+(L(x,y+1)-L(x,y-1))2]1/2 (7)
Wherein, L (x, y, σ) represents the metric space of image, and m (x, y) and θ (x, y) are respectively gaussian pyramid (x, y) place Gradient magnitude and direction, direction used in L be each crucial point scale under direction;
Obtain after m (x, y) and θ (x, y), sampled in the contiguous range centered on key point, it is counted after sampling Pixel gradient direction, is then represented with histogram.Histogrammic peak value represents the principal direction of the crucial neighborhood of a point gradient, as The direction of the key point;So far, feature point detection is finished, and each characteristic point is believed comprising position, yardstick and the aspect of direction 3 Breath;
3-5) generation feature point description:Feature point description is calculated by carrying out image block to key point peripheral region Histogram of gradients in each piece, generates unique vector descriptor, and specific method is:16 × 16 are taken centered on key point Sample area, divides the area into 16 seed points that size is 4 × 4, calculates 8 directions respectively in each seed point region Gradient accumulated value, draw gradient direction direction histogram;The Feature Descriptor of final 128 dimensions for obtaining one 4 × 4 × 8;
4) fixation and recognition is carried out using feature.
The invention is not limited in above-described embodiment, on the basis of technical scheme disclosed by the invention, the skill of this area Art personnel are according to disclosed technology contents, it is not necessary to which performing creative labour just can make one to some of which technical characteristic A little to replace and deform, these are replaced and deformed within the scope of the present invention.

Claims (3)

1. a kind of underwater robot adaptive image enhancement and feature extracting method, it is characterised in that comprise the following steps:
1) video camera carried using underwater robot is carried out image taking and stored;
2) image procossing is carried out using CLAHE technologies;
3) image characteristics extraction is carried out using SIFT algorithms;
4) fixation and recognition is carried out using feature.
2. a kind of underwater robot adaptive image enhancement according to claim 1 and feature extracting method, its feature exist In step 2) specific method be:
2-1) by step 1) artwork be divided into k size be m × n continuous nonoverlapping subregion.The value of [m n] decides The details enhancing degree of image, generally its value is smaller, and enhancing effect is weaker;
2-2) calculate the grey level histogram per sub-regions;
2-3) determine per sub-regions grey level histogram " shearing " value;
The average value N that number of pixels in per sub-regions is evenly distributed to each gray level is calculated by formula (1) firstaver
N a v e r = N C R - X p N C R - Y p N g r a y - - - ( 1 )
Wherein, NgrayIt is the quantity of gray level in subregion;It is the pixel count in subregion x-axis direction;It is sub-district The pixel count in domain y-axis direction.Actual " shearing " value is NCL
NCL=Nclip·Naver (2)
Wherein, NclipIt is interception limit coefficient, it is flat that it is meant that the pixel count that the limitation each gray level of subregion is included is no more than The N of equal pixel countclipTimes;
The grey level histogram per sub-regions 2-4) is sheared, the number of pixels that remainder is come is re-assigned to each histogrammic each gray scale In level;
If the sum of all pixels being sheared is NΣclip, then obtain the shearing pixel count N that each gray level is divided equallyacp
N a c p = N Σ c l i p N g r a y - - - ( 3 )
2-5) after redistributing, pixel remainder is shear off for NLP, the step value of pixel is allocated by equation below Provide:
S = N g r a y N L P - - - ( 4 )
It is 0 to carry out cycle assignment to pixel count by above-mentioned step value since minimal gray level to residual pixel, is finally given new Histogram;
2-6) grey level histogram after contrast " shearing " in every sub-regions is equalized;
2-7) according to new histogram distribution, gray scale bilinear interpolation is carried out to image, enhanced result images are generated.
3. a kind of underwater robot adaptive image enhancement according to claim 2 and feature extracting method, its feature exist In step 3) specific method be:
3-1) read in step 2-7) in obtained enhanced result images, generate Gaussian difference scale space:For characteristic point The unstable phenomenon on metric space, this problem is solved using Gaussian difference scale space, it by different scale height This difference core carries out convolution with image and obtained, such as formula (5):
D (x, y, σ)=(G (x, y, k σ)-G (x, y, σ)) * I (x, y) (5)
Wherein, σ is that normal distribution standard is poor, and * is convolution symbol, and I (x, y) represents input picture, and G (x, y, σ) is changeable scale Gaussian function, (x, y) represent image pixel position, k be the adjacent yardstick of each two between be separated by the linetype scale factor;
3-2) metric space extreme point is detected:The acquisition of extreme point is needed by sampled point and adjacent thereto a range of Point is compared, the scope be under same yardstick around 8 points and totally 26 points of the 2*9 in two neighborhoods up and down, see the point Whether it is maximum or minimum value;
3-3) key point is accurately positioned:Being accurately positioned for key point can improve its matching accuracy, therefore, need to be done to key point Handle below:
1. position and the dimensional information of key point are determined by the fitting of three-dimensional quadratic function;
2. by setting suitable threshold value to filter out the key point of low contrast;
3. the given threshold key point unstable to skirt response is filtered out;
3-4) determine the main gradient direction of key point:According to the scale-value σ for characteristic point and this feature point having been detected by, connect The Gaussian image of this nearly scale-value:
L (x, y, σ)=G (x, y, σ) * I (x, y) (6)
Gradient magnitude and the direction that the key point can then be obtained are as follows:
M (x, y)=[(L (x+1, y)-L (x-1, y))2+(L(x,y+1)-L(x,y-1))2]1/2 (7)
θ ( x , y ) = a r c t a n L ( x + 1 , y ) - L ( x - 1 , y ) L ( x , y + 1 ) - L ( x , y - 1 ) - - - ( 8 )
Wherein, L (x, y, σ) represents the metric space of image, and m (x, y) and θ (x, y) are respectively the ladder at gaussian pyramid (x, y) place Size and Orientation is spent, direction used in L is the direction under each crucial point scale;
Obtain after m (x, y) and θ (x, y), sampled in the contiguous range centered on key point, its pixel is counted after sampling Gradient direction, is then represented with histogram.Histogrammic peak value represents the principal direction of the crucial neighborhood of a point gradient, is used as the pass The direction of key point.So far, feature point detection is finished, and each characteristic point includes position, yardstick and the aspect information of direction 3;
3-5) generation feature point description:Feature point description calculates each piece by carrying out image block to key point peripheral region Interior histogram of gradients, generates unique vector descriptor, and specific method is:Taken centered on key point 16 × 16 sampling Region, divides the area into 16 seed points that size is 4 × 4, calculates the ladder in 8 directions respectively in each seed point region Accumulated value is spent, the direction histogram of gradient direction is drawn;The Feature Descriptor of final 128 dimensions for obtaining one 4 × 4 × 8.
CN201710141451.0A 2017-03-10 2017-03-10 A kind of underwater robot adaptive image enhancement and feature extracting method Withdrawn CN106997583A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710141451.0A CN106997583A (en) 2017-03-10 2017-03-10 A kind of underwater robot adaptive image enhancement and feature extracting method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710141451.0A CN106997583A (en) 2017-03-10 2017-03-10 A kind of underwater robot adaptive image enhancement and feature extracting method

Publications (1)

Publication Number Publication Date
CN106997583A true CN106997583A (en) 2017-08-01

Family

ID=59431432

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710141451.0A Withdrawn CN106997583A (en) 2017-03-10 2017-03-10 A kind of underwater robot adaptive image enhancement and feature extracting method

Country Status (1)

Country Link
CN (1) CN106997583A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111127341A (en) * 2019-12-05 2020-05-08 Oppo广东移动通信有限公司 Image processing method and apparatus, and storage medium
CN112200019A (en) * 2020-09-22 2021-01-08 江苏大学 Rapid building night scene lighting light fault detection method
CN113962890A (en) * 2021-10-22 2022-01-21 大连海事大学 Affine transformation-based underwater color cast image sharpening method
CN113989628A (en) * 2021-10-27 2022-01-28 哈尔滨工程大学 Underwater signal lamp positioning method based on weak direction gradient

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103440637A (en) * 2013-09-22 2013-12-11 北京智诺英特科技有限公司 Method and device used for enhancing image sequences
CN103839265A (en) * 2014-02-26 2014-06-04 西安电子科技大学 SAR image registration method based on SIFT and normalized mutual information
CN104463816A (en) * 2014-12-02 2015-03-25 苏州大学 Image processing method and device
CN104794705A (en) * 2015-03-27 2015-07-22 努比亚技术有限公司 Image defogging method and method based on image partial content characteristics
CN105809626A (en) * 2016-03-08 2016-07-27 长春理工大学 Self-adaption light compensation video image splicing method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103440637A (en) * 2013-09-22 2013-12-11 北京智诺英特科技有限公司 Method and device used for enhancing image sequences
CN103839265A (en) * 2014-02-26 2014-06-04 西安电子科技大学 SAR image registration method based on SIFT and normalized mutual information
CN104463816A (en) * 2014-12-02 2015-03-25 苏州大学 Image processing method and device
CN104794705A (en) * 2015-03-27 2015-07-22 努比亚技术有限公司 Image defogging method and method based on image partial content characteristics
CN105809626A (en) * 2016-03-08 2016-07-27 长春理工大学 Self-adaption light compensation video image splicing method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
胡学友: "雾天降质图像的增强复原算法研究", 《中国博士学位论文全文数据库 信息科技辑》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111127341A (en) * 2019-12-05 2020-05-08 Oppo广东移动通信有限公司 Image processing method and apparatus, and storage medium
CN112200019A (en) * 2020-09-22 2021-01-08 江苏大学 Rapid building night scene lighting light fault detection method
CN112200019B (en) * 2020-09-22 2024-02-09 上海罗曼照明科技股份有限公司 Rapid building night scene lighting lamp fault detection method
CN113962890A (en) * 2021-10-22 2022-01-21 大连海事大学 Affine transformation-based underwater color cast image sharpening method
CN113962890B (en) * 2021-10-22 2024-05-10 大连海事大学 Affine transformation-based underwater color cast image sharpening method
CN113989628A (en) * 2021-10-27 2022-01-28 哈尔滨工程大学 Underwater signal lamp positioning method based on weak direction gradient

Similar Documents

Publication Publication Date Title
CN109242888B (en) Infrared and visible light image fusion method combining image significance and non-subsampled contourlet transformation
CN109948562B (en) Security check system deep learning sample generation method based on X-ray image
CN109872285B (en) Retinex low-illumination color image enhancement method based on variational constraint
CN106204509B (en) Infrared and visible light image fusion method based on regional characteristics
CN106997583A (en) A kind of underwater robot adaptive image enhancement and feature extracting method
CN110163818A (en) A kind of low illumination level video image enhancement for maritime affairs unmanned plane
CN106530271B (en) A kind of infrared image conspicuousness detection method
CN107808161A (en) A kind of Underwater targets recognition based on light vision
CN108564092A (en) Sunflower disease recognition method based on SIFT feature extraction algorithm
CN107798670A (en) A kind of dark primary prior image defogging method using image wave filter
CN109242787A (en) It paints in a kind of assessment of middle and primary schools' art input method
CN108664970A (en) A kind of fast target detection method, electronic equipment, storage medium and system
CN110619647A (en) Method for positioning fuzzy region of image based on combination of edge point frequency domain and spatial domain characteristics
CN108694349A (en) A kind of pantograph image extraction method and device based on line-scan digital camera
CN113657335A (en) Mineral phase identification method based on HSV color space
CN105354547A (en) Pedestrian detection method in combination of texture and color features
CN104992426B (en) A kind of multi-layer image blending algorithm for light field micro-imaging
CN111626944A (en) Video deblurring method based on space-time pyramid network and natural prior resistance
CN108960285B (en) Classification model generation method, tongue image classification method and tongue image classification device
CN111047546A (en) Infrared image super-resolution reconstruction method and system and electronic equipment
CN110345919A (en) Space junk detection method based on three-dimensional space vector and two-dimensional plane coordinate
CN114994098A (en) Foreign matter detection method and device
CN114582017A (en) Generation method and generation system of gesture data set and storage medium
Cheng et al. Research on image processing technology based on analog head reading
Tao et al. Measurement algorithm of notch length of plastic parts based on video

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication
WW01 Invention patent application withdrawn after publication

Application publication date: 20170801