CN113364963A - Imaging control method and related equipment - Google Patents

Imaging control method and related equipment Download PDF

Info

Publication number
CN113364963A
CN113364963A CN202110910045.2A CN202110910045A CN113364963A CN 113364963 A CN113364963 A CN 113364963A CN 202110910045 A CN202110910045 A CN 202110910045A CN 113364963 A CN113364963 A CN 113364963A
Authority
CN
China
Prior art keywords
imaging
target
target object
adjusting
aperture
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110910045.2A
Other languages
Chinese (zh)
Other versions
CN113364963B (en
Inventor
潘武
陈天钧
况璐
潘润发
赵军
李准
卢二利
周国民
刘强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang Dahua Technology Co Ltd
Original Assignee
Zhejiang Dahua Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang Dahua Technology Co Ltd filed Critical Zhejiang Dahua Technology Co Ltd
Priority to CN202110910045.2A priority Critical patent/CN113364963B/en
Publication of CN113364963A publication Critical patent/CN113364963A/en
Application granted granted Critical
Publication of CN113364963B publication Critical patent/CN113364963B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/50Constructional details
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/50Constructional details
    • H04N23/55Optical parts specially adapted for electronic image sensors; Mounting thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/70Circuitry for compensating brightness variation in the scene

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Studio Devices (AREA)
  • Exposure Control For Cameras (AREA)

Abstract

The application discloses an imaging control method and related equipment, wherein the imaging control method comprises the following steps: determining at least two target objects in an imaging picture of an imaging device; adjusting an imaging related element of the imaging device to a target position, wherein the target position enables an imaging plane of the imaging device to be arranged at an angle with respect to a lens plane; and adjusting the aperture of the imaging device to a target aperture value, wherein the target aperture value can enable the imaging definition conditions of at least two target objects to meet a first preset requirement. By the method, the target object in subsequent imaging can be clearer.

Description

Imaging control method and related equipment
Technical Field
The present application relates to the field of camera imaging technologies, and in particular, to an imaging control method and related devices.
Background
How to obtain a clear image for camera imaging is always an important issue in the technical field. In the process of camera imaging, if a plurality of target objects are located on a plane which is not parallel to the lens plane and the imaging plane, the depth of field of the imaging device needs to be adjusted to obtain a clear image of the target objects. However, the depth of field adjustment of the imaging apparatus has a limit, which makes it impossible to obtain a clear image about the target object by adjusting the depth of field in some cases.
Therefore, it is very important how to improve the imaging technology of the imaging device to obtain a clearer image, especially in the case that the planes of a plurality of target objects are not parallel to the lens plane and the imaging plane, and a clearer image about the target objects is obtained.
Disclosure of Invention
The application provides an imaging control method and related equipment.
A first aspect of the present application provides an imaging control method, including: determining at least two target objects in an imaging picture of an imaging device; adjusting an imaging related element of the imaging device to a target position, wherein the target position enables an imaging plane of the imaging device to be arranged at an angle with respect to a lens plane; and adjusting the aperture of the imaging device to a target aperture value, wherein the target aperture value can enable the imaging definition conditions of at least two target objects to meet a first preset requirement.
Therefore, the imaging plane of the imaging device and the lens plane are arranged at an angle by adjusting the imaging related elements of the imaging device to the target position, so that imaging by utilizing the Schlemm's law is realized, a clear image of a target object can be obtained under the condition that the planes of a plurality of target objects are not parallel to the lens plane and the imaging plane, and the depth of field of the imaging device is changed by adjusting the aperture of the imaging device to the target aperture value, so that the target object in subsequent imaging can be clearer.
A second aspect of the present application provides an imaging apparatus including an imaging-related element for forming an imaging screen, and an adjustment member; the adjusting component is used for adjusting the imaging related element to a target position, wherein the target position can enable an imaging plane of the imaging device to be arranged at an angle with the lens plane; and adjusting the aperture of the imaging device to a target aperture value, wherein the target aperture value can enable the imaging definition conditions of at least two target objects in the imaging picture to meet a first preset requirement.
A third aspect of the application provides an imaging control device comprising a processor and a memory coupled to each other, wherein the processor is configured to execute a computer program stored in the memory to perform the method described in the first aspect above.
According to the scheme, the imaging plane of the imaging device and the lens plane are arranged at an angle by adjusting the imaging related elements of the imaging device to the target position, so that imaging by utilizing the Schlemm's law is realized, a clear image about a target object can be obtained under the condition that the plane where a plurality of target objects are located is not parallel to the lens plane and the imaging plane, and the depth of field of the imaging device is changed by adjusting the aperture of the imaging device to the target aperture value, so that the target object in subsequent imaging can be clearer.
Drawings
FIG. 1 is a first flow chart of a first embodiment of an imaging control method of the present application;
FIG. 2a is a schematic diagram of an imaging plane parallel to a lens plane in the imaging control method of the present application;
FIG. 2b is a schematic diagram of an imaging plane and a lens plane arranged at an angle in the imaging control method of the present application;
FIG. 2c is another schematic diagram of the imaging plane and the lens plane being disposed at an angle in the imaging control method of the present application;
FIG. 2d is a schematic view of the imaging plane and the lens plane being disposed at an angle in the imaging control method of the present application;
FIG. 3 is a second flow chart of the first embodiment of the imaging control method of the present application;
FIG. 4 is a third flow chart of the first embodiment of the imaging control method of the present application;
FIG. 5 is a schematic flow chart diagram of a second embodiment of the imaging control method of the present application;
FIG. 6 is a block diagram of an embodiment of an imaging device of the present application;
FIG. 7 is a block diagram of an embodiment of an imaging control apparatus of the present application;
FIG. 8 is a block diagram of an embodiment of a computer-readable storage medium of the present application.
Detailed Description
The following describes in detail the embodiments of the present application with reference to the drawings attached hereto.
In the following description, for purposes of explanation and not limitation, specific details are set forth such as particular system structures, interfaces, techniques, etc. in order to provide a thorough understanding of the present application.
The terms "system" and "network" are often used interchangeably herein. The term "and/or" herein is merely an association describing an associated object, meaning that three relationships may exist, e.g., a and/or B, may mean: a exists alone, A and B exist simultaneously, and B exists alone. In addition, the character "/" herein generally indicates that the former and latter related objects are in an "or" relationship. Further, the term "plurality" herein means two or more than two.
In the present application, an execution subject executing the imaging control method of the present application may be an imaging apparatus, such as a camera, or the like, or may be a related control apparatus connected to the imaging apparatus, such as a computer connected to the camera, or the like.
Referring to fig. 1, fig. 1 is a first flowchart of a first embodiment of an imaging control method according to the present application. Specifically, the method may include the steps of:
step S11: at least two target objects in an imaging frame of an imaging device are determined.
The imaging device is, for example, a camera, a video camera, or the like that can form an image using the optical imaging principle. The imaging device may be used to acquire an imaging picture.
In the present application, at least two target objects in an imaging picture of an imaging device may be determined. Image recognition may be performed on the imaged picture to determine several objects present in the imaged picture, and then at least some of the objects may be determined as target objects. For example, after image recognition is performed on the imaged picture, it is determined that three persons and two vehicles exist in the imaged picture. At this time, three persons may be specified as the target object, or two vehicles may be specified as the target object. In one embodiment, the number of target objects is not less than 3.
In one embodiment, after image recognition is performed on the imaged picture and several objects existing in the imaged picture are determined, objects belonging to the same class may be determined as target objects. At this time, several classifications may be set in advance, then the objects determined by image recognition are classified, and finally the objects belonging to the same class are determined as the target objects. For example, after image recognition is performed on an imaging picture, it is determined that three persons, two vehicles and three dogs exist in the imaging picture, at this time, according to a preset classification, the persons may be classified into one category, the vehicles may be classified into one category, and the dogs may be classified into one category, and then, the three persons may be determined as a target object, or two vehicles may be determined as a target object, or two dogs may be determined as a target object.
In one embodiment, after performing image recognition on the imaged picture and determining several objects existing in the imaged picture, an object having a height difference within a preset range may be used as a target object. At this time, a height value may be set in advance for each object, for example, a height value of a person is set to 1.7 m, a height value of a dog is set to 0.55 m, a height value of a car is set to 1.5 m, and the like. If the preset range of the height difference is set to 0.3 m, both the vehicle and the person can be targeted.
In one embodiment, the target objects may be target recognition results obtained by target-recognizing the imaged picture by the image recognition apparatus, and the at least two target objects are determined based on the target recognition results. For example, the target objects are two persons, the imaged picture may be transmitted to an image recognition apparatus, target recognition is performed by the image recognition apparatus, target recognition results of the two persons are obtained, and then the target objects may be determined. In one embodiment, the image recognition device may be a part of an execution main body that executes the imaging control method of the present application, or may be another device connected to the execution main body. By using the image recognition device for target recognition, the target object in the imaging picture can be determined, and then the relevant imaging can be carried out on the target object.
In one embodiment, at least two target objects are determined in an imaging picture of an imaging device, each target object can be framed by a frame with a preset shape, and an image area to which the frame with the preset shape belongs is an area of the target object in the imaging picture.
In one embodiment, the image recognition apparatus recognizes a number of target objects greater than two, and at this time, two target objects having the largest area may be selected as the target objects to be finally confirmed. In another embodiment, a target object whose ratio of the area of the target object in the imaging screen to the entire imaging screen is not less than 1/4 may be selected as the target object for final confirmation.
Step S12: adjusting an imaging-related element of the imaging device to a target position.
The imaging-related element includes at least one of an imaging sensor and a lens. The target position enables an imaging plane of the imaging device to be disposed at an angle to the lens plane. In this application, the imaging plane being disposed at an angle to the lens plane means that the two planes are not parallel or that the angle between the two planes is not 0 ° or 180 °. The imaging plane and the lens plane can be arranged at an angle by adjusting at least one of the imaging sensor and the lens of the imaging device. At this time, since the imaging plane is disposed at an angle to the lens plane, imaging can be performed by using the schemer's law. In one embodiment, the imaging sensor may be adjusted such that the imaging plane is disposed at an angle to the lens plane. In another embodiment, the lens may be adjusted such that the imaging plane is disposed at an angle to the lens plane. In one embodiment, the position of the imaging-related element may be changed by driving the motor, so that the imaging plane is disposed at an angle to the lens plane.
Referring to fig. 2a, fig. 2a is a schematic view illustrating an imaging plane and a lens plane in the imaging control method of the present application are parallel. In fig. 2a, the imaging plane 211 is parallel to the lens plane 221. Referring to fig. 2b, fig. 2b is a schematic view illustrating an angle between an imaging plane and a lens plane in the imaging control method of the present application. In fig. 2b, the imaging plane 211 is set at an angle to the lens plane 221 by adjusting the imaging sensor 21. Referring to fig. 2c, fig. 2c is another schematic diagram of the imaging plane and the lens plane being disposed at an angle in the imaging control method of the present application. In fig. 2c, the lens 22 is adjusted so that the imaging plane 211 is at an angle to the lens plane 221. Referring to fig. 2d, fig. 2d is another schematic diagram of the imaging plane and the lens plane being arranged at an angle in the imaging control method of the present application. In fig. 2d, the imaging plane 211 is set at an angle to the lens plane 221 by adjusting the imaging sensor 21 and the lens 22.
Step S13: and adjusting the aperture of the imaging device to the target aperture value.
The aperture of the imaging apparatus is adjusted to a target aperture value, i.e., the size of the aperture is adjusted to adjust the aperture to the target aperture value. By adjusting the size of the aperture, different depths of field can be obtained, and the imaging clarity of the target object can be adjusted. The imaging definition condition may be imaging definition, and the imaging definition may be obtained according to a general imaging definition calculation method, which is not described herein again. In the present application, the target aperture value enables the imaging clarity of the at least two target objects to satisfy the first preset requirement. In one embodiment, the first preset requirement may be an aperture value corresponding to the optimal imaging clarity in adjusting the aperture. In other embodiments, the first preset requirement may be an aperture value corresponding to the case where the imaging clarity of a part of the target objects is optimal among all the target objects. In a specific embodiment, the imaging definition of the target object may be calculated by calculating the imaging definition of the region where the target object is located. The imaging definition condition corresponding to the candidate aperture value in the plurality of candidate aperture values is optimal by setting the first preset requirement, and the clearest image can be obtained by adjusting the aperture. It is understood that the first preset requirement can be set according to the requirement, and is not limited herein.
Therefore, the imaging plane of the imaging device and the lens plane are arranged at an angle by adjusting the imaging related elements of the imaging device to the target position, so that imaging by utilizing the Schlemm's law is realized, a clear image of a target object can be obtained under the condition that the planes of a plurality of target objects are not parallel to the lens plane and the imaging plane, and the depth of field of the imaging device is changed by adjusting the aperture of the imaging device to the target aperture value, so that the target object in subsequent imaging can be clearer.
Referring to fig. 3, fig. 3 is a second flow chart of the first embodiment of the imaging control method of the present application. In the present embodiment, the aforementioned step of "adjusting the imaging-related element of the imaging apparatus to the target position" specifically includes steps S121 to S124.
Step S121: and adjusting the imaging related elements to a plurality of candidate positions respectively.
In this embodiment, the candidate positions are all capable of setting the imaging plane at an angle to the lens plane.
In one embodiment, several candidate locations may be determined by an iterative optimization method. For example, the adjustment direction of the imaging sensor position may be determined by a hill climbing algorithm, and then the position of the imaging sensor may be adjusted to several candidate positions according to the adjustment direction. The adjustment direction of the imaging plane is determined, for example, by a hill-climbing algorithm. In other embodiments, the candidate position may be a predetermined position. It is to be understood that the determination of the candidate position is not limited and may be set as desired.
Step S122: and respectively acquiring the imaging definition of the target object when the imaging related element is adjusted to different candidate positions.
After adjusting the imaging related element to several candidate positions, the image sharpness of the target object may be calculated for each candidate position. That is, at this time, the imaging sharpness of the target object when the imaging related element is adjusted to different candidate positions can be acquired separately. The imaging definition may be obtained according to a general imaging definition calculation method, and is not described herein again.
Step S123: and determining candidate positions which enable the imaging definition conditions of at least two target objects to meet a second preset requirement as target positions.
In one embodiment, the second preset requirement is that the imaging clarity corresponding to the candidate position is optimal among several candidate positions. The imaging clarity condition corresponding to the candidate position in the candidate positions may specifically be the imaging clarity of an image area in the imaging picture corresponding to the target object in the candidate positions. In other embodiments, the second preset requirement may be that imaging clarity of at least one of the at least two target objects is optimal.
Step S124: and adjusting the imaging related element to the target position.
After the target position is determined, the imaging-related component may be adjusted to the target position, for example, the imaging sensor or the lens may be adjusted to the target position.
Therefore, by determining the target position according to the imaging clarity of the target object, the target position of the imaging-related element can be adjusted in a targeted manner according to the target object, whereby a clear image about the target object can be obtained.
Referring to fig. 4, fig. 4 is a third flow chart of the first embodiment of the imaging control method of the present application. In the present embodiment, the aforementioned step of "adjusting the aperture of the imaging apparatus to the target aperture value" specifically includes steps S131 to S134.
Step S131: and respectively adjusting the aperture to a plurality of candidate aperture values.
In one embodiment, the candidate aperture value may be any value within a size range in which the aperture can be adjusted. For example, the size range in which the aperture can be adjusted is F2.8-F11, and the candidate aperture value may be any one of F2.8-F11. In other embodiments, the candidate aperture value may be any value within a preset range, and the preset range may be set according to actual needs.
In one embodiment, because adjusting the aperture causes the brightness of the imaged picture to change, several candidate aperture values may be determined within a preset brightness loss range.
Step S132: and respectively acquiring the imaging definition of the target object when the aperture is adjusted to different candidate aperture values.
When the aperture is adjusted to a plurality of candidate aperture values, the imaging sharpness of the target object at different candidate aperture values may be obtained, and specifically, the imaging sharpness of an image area in an imaging picture corresponding to the target object may be obtained.
Step S133: and determining the candidate aperture values which enable the imaging definition conditions of at least two target objects to meet the first preset requirement as the target aperture values.
In an embodiment, the first preset requirement is that the imaging clarity corresponding to the candidate aperture value is optimal among the plurality of candidate aperture values, and specifically, the imaging clarity of all target objects is optimal, or the imaging clarity of a part of target objects is optimal. By determining the first preset requirement as the imaging definition condition corresponding to the candidate aperture value is optimal, a clear image about the target object can be obtained by adjusting the aperture.
Step S134: the aperture is adjusted to the target aperture value.
After the target aperture value is determined, the aperture of the imaging apparatus may be adjusted to the target aperture value, thereby obtaining a clear image about the target object.
Therefore, a clear image of the target object can be obtained by determining a target aperture value satisfying the first preset requirement at several candidate aperture values.
Referring to fig. 5, fig. 5 is a schematic flow chart of a second embodiment of the imaging control method of the present application. In this embodiment, steps S21 to S23 are specifically included.
Step S21: a distance relationship between each target object and the imaging device is acquired.
In one embodiment, the distance relationship between the target objects and the imaging device is a far-near relationship between the target objects and the imaging device, i.e., it is determined which target object is closer to the imaging device and which target object is farther from the imaging device between the target objects.
In one embodiment, the distance relationship between each target object and the imaging device may be determined by depth information between each target object and the imaging device. The depth information may be determined based on the depth image. In another embodiment, the distance relationship between each target object and the imaging device may be determined by other ranging methods, such as laser ranging.
In one embodiment, the distance relationship between each target object and the imaging device may be acquired through steps S211 and S212.
Step S211: and adjusting the focal length of the imaging equipment, and determining the imaging definition change trend of the target object in the process of adjusting the focal length.
When the focal length of the imaging device is adjusted, the focal length can be adjusted from long to short, or the focal length can be adjusted from short to long, and the definition of the target object is correspondingly changed in the process of adjusting the focal length, so that the change trend of the imaging definition of the target object can be obtained. The trend of the change in the image sharpness of the target object is in particular increasingly sharp or increasingly unclear.
Step S212: and determining the far-near relation between each target object and the imaging equipment according to the imaging definition change trend.
In the camera imaging principle, when the focal length is shortened from long to short, the imaging sharpness of a target object at a closer distance becomes better, and the imaging sharpness of a target object at a further distance becomes worse. For example, when the focal distance becomes smaller, a certain target object is more clear, and it may be determined that the target object is closer to the imaging device, and another target object is less clear, and it may be determined that the target object is farther from the imaging device. Conversely, when the focal length is larger, a certain target object is clearer, and it can be determined that the target object is farther from the imaging device, and another target object is less clear, and it can be determined that the target object is closer to the imaging device.
Therefore, by acquiring the trend of the change in the imaging sharpness of the target object in the process of adjusting the focal length, the distance relationship between each target object and the imaging apparatus can be determined.
In another embodiment, the distance relationship between each target object and the imaging device may also be acquired through steps S213 and S214.
Step S213: and determining the target area of each target object in the imaging picture.
In one embodiment, the target area of the target object in the imaging picture may be determined, for example, in the step of determining at least two target objects in the imaging picture of the imaging device, after the target object is determined, the target area of the target object in the imaging picture may also be determined. For example, the size of the imaged picture is 3200 x 6400, where the area of one target object may be 500 x 600.
Step S214: based on the target area of each target object, the distance relationship between the target objects and the imaging device is determined.
It will be appreciated that objects closer to the imaging device will generally have a larger area in the imaged frame. Therefore, the size of the target area of each target object can be judged to determine the distance relationship between the target objects and the imaging device. Specifically, the larger the target area of the target object, the closer to the imaging device may be considered, and the smaller the target area of the target object, the farther from the imaging device may be considered.
Therefore, by based on the target area of each target object, the distance relationship between the target objects to the imaging device can be determined.
Step S22: based on the distance relationship, a sharpness weight for each target object is determined.
In one embodiment, the depth of field of the imaging device approximates a cone when imaging using the Schlemm's Law, with the depth of field being smaller the closer the imaging device is. Therefore, when the imaging clarity of a target object at a close distance from the imaging device is better, the imaging clarity of a target object at a far distance from the imaging device is generally better. Therefore, the sharpness weight of a target object closer to the imaging device can be set to be larger than the sharpness weight of a target object farther from the imaging device.
Of course, in other embodiments, the sharpness weight of a target object closer to the imaging device may be less than the sharpness weight of a target object farther from the imaging device.
Step S23: and carrying out weighting processing on the imaging definition of each target object based on the definition weight to obtain the imaging definition conditions of at least two target objects.
The imaging definition of each target object is weighted based on the definition weight, the imaging definition of each target object can be calculated respectively, then the imaging definition of each target object is weighted based on the definition weight of each target object, and finally the imaging definition conditions of at least two target objects are obtained through addition, so that the imaging definition conditions of all the target objects are obtained.
In an embodiment, before the "weighting processing is performed on the imaging sharpness of each target object based on the sharpness weight to obtain the imaging sharpness of at least two target objects", the imaging control method of the present application may further include step S31 and step S2.
Step S31: a target region of each target object in the imaging sensor is determined.
The target area of each target object in the imaging sensor, i.e. in which area the target object is specifically imaged in the imaging sensor, can be determined.
In one embodiment, a target region of each target object in the imaging sensor may be determined through the following steps S311 to S313.
Step S311: the imaging picture is divided into a plurality of first areas, and the imaging area of the imaging sensor is divided into a plurality of second areas.
In one embodiment, the imaging screen may be divided into several first regions and the imaging region of the imaging sensor may be divided into several second regions, and then the second region corresponding to each of the first regions may be determined. For example, the imaging frame is divided into 64 first regions of 8 × 8, and in this case, the imaging region of the imaging sensor may be divided into 64 second regions of 8 × 8. Thus, each first region has a one-to-one correspondence with its second region.
Step S312: and determining a first area corresponding to each target object.
The first region corresponding to each target object is determined, and specifically, the first region corresponding to the target object is determined according to a region of the target object in the imaging picture.
In one embodiment, the area of the target object in the imaging frame is a rectangle, which may be determined by the image recognition device in recognizing the target object. Then, the first region corresponding to the target object can be determined according to the region of each rectangle in the imaging picture. In one embodiment, if the region of the target object in the imaging screen spans a plurality of first regions, a first region where the region of the target object in the imaging screen occupies the largest area among the first regions may be used as the first region corresponding to the target object.
Step S313: and taking the second area corresponding to the first area as a target area.
After the first region corresponding to each target object is determined, the target region of each target object in the imaging sensor can be determined according to the corresponding relation between each first region and the second region.
Therefore, by dividing the imaging picture into a plurality of first areas and dividing the imaging area of the imaging sensor into a plurality of second areas, the corresponding relation between the first areas and the second areas can be determined, and the target area of each target object in the imaging sensor can be further determined.
Step S32: and determining the imaging definition of the target object based on the imaging picture information of the target area.
After the target area of each target object in the imaging sensor is determined, the imaging definition of each target object can be calculated according to the imaging picture information of the target area, so that the imaging definition of each target object can be determined.
Therefore, by determining the target area of each target object in the imaging sensor and using the imaging picture information of the target area, the imaging definition of each target object can be determined more accurately without being interfered by the imaging picture information outside the target area.
Referring to fig. 6, fig. 6 is a schematic diagram of a framework of an embodiment of an image forming apparatus of the present application. In the present embodiment, the imaging device 60 includes an imaging-related element 61 and an adjustment assembly 62. The imaging-related element 61 is used to form an imaging screen. The adjusting component 62 is used for adjusting the imaging related element to a target position, wherein the target position enables an imaging plane of the imaging device to be arranged at an angle with respect to the lens plane; and adjusting the aperture of the imaging device to a target aperture value, wherein the target aperture value can enable the imaging definition conditions of at least two target objects in the imaging picture to meet a first preset requirement.
Referring to fig. 7, fig. 7 is a schematic diagram of a framework of an embodiment of the imaging control apparatus of the present application. The imaging control apparatus 70 comprises a memory 701 and a processor 702 coupled to each other, the processor 702 being configured to execute program instructions stored in the memory 701 to implement the steps of any of the above-described embodiments of the imaging control method. In one particular implementation scenario, imaging control device 70 may include, but is not limited to: the microcomputer, the server, and in addition, the imaging control device 70 may further include a mobile device such as a notebook computer, a tablet computer, and the like, which is not limited herein.
In one embodiment, the imaging control device 70 further comprises an imaging related element 703 and an adjustment component 704, the imaging related element 703 being used to form an imaging picture; the adjustment component 704 is configured to perform adjustments of the imaging-related elements and the aperture in response to control instructions of the processor.
Specifically, the processor 702 is configured to control itself and the memory 701 to implement the steps of any of the above-described imaging control method embodiments. Processor 702 may also be referred to as a CPU (Central Processing Unit). The processor 702 may be an integrated circuit chip having signal processing capabilities. The Processor 702 may also be a general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other Programmable logic device, discrete Gate or transistor logic, discrete hardware components. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like. In addition, the processor 702 may be collectively implemented by an integrated circuit chip.
Referring to fig. 8, fig. 8 is a block diagram illustrating an embodiment of a computer-readable storage medium according to the present application. The computer readable storage medium 80 stores program instructions 81 executable by the processor, the program instructions 81 being for implementing the steps of any of the imaging control method embodiments described above.
In some embodiments, functions of or modules included in the apparatus provided in the embodiments of the present disclosure may be used to execute the method described in the above method embodiments, and specific implementation thereof may refer to the description of the above method embodiments, and for brevity, will not be described again here.
The foregoing description of the various embodiments is intended to highlight various differences between the embodiments, and the same or similar parts may be referred to each other, and for brevity, will not be described again herein.
In the several embodiments provided in the present application, it should be understood that the disclosed method and apparatus may be implemented in other ways. For example, the above-described apparatus embodiments are merely illustrative, and for example, a division of a module or a unit is merely one type of logical division, and an actual implementation may have another division, for example, a unit or a component may be combined or integrated with another system, or some features may be omitted, or not implemented. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection of devices or units through some interfaces, and may be in an electrical, mechanical or other form.
Units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on network elements. Some or all of the units can be selected according to actual needs to achieve the purpose of the embodiment.
In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated unit, if implemented in the form of a software functional unit and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present application may be substantially implemented or contributed to by the prior art, or all or part of the technical solution may be embodied in a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, a network device, or the like) or a processor (processor) to execute all or part of the steps of the method according to the embodiments of the present application. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.

Claims (15)

1. An imaging control method, characterized by comprising:
determining at least two target objects in an imaging picture of an imaging device;
adjusting an imaging related element of the imaging device to a target position, wherein the target position enables an imaging plane of the imaging device to be arranged at an angle with a lens plane;
and adjusting the aperture of the imaging device to a target aperture value, wherein the target aperture value can enable the imaging definition conditions of the at least two target objects to meet a first preset requirement.
2. The method of claim 1, wherein said adjusting an imaging-related element of said imaging device to a target position comprises:
adjusting the imaging related elements to a plurality of candidate positions respectively, wherein the candidate positions can enable the imaging plane and the lens plane to be arranged at an angle;
respectively acquiring the imaging definition of the target object when the imaging related element is adjusted to different candidate positions;
determining candidate positions which enable the imaging definition conditions of the at least two target objects to meet a second preset requirement as the target positions;
adjusting the imaging-related element to the target position.
3. The method according to claim 2, wherein the second predetermined requirement is that the imaging resolution corresponding to the candidate position among the plurality of candidate positions is optimal.
4. The method of claim 1, wherein the adjusting the aperture of the imaging device to a target aperture value comprises:
respectively adjusting the aperture to a plurality of candidate aperture values;
respectively acquiring the imaging definition of the target object when the aperture is adjusted to different candidate aperture values;
determining a candidate aperture value which enables the imaging definition conditions of the at least two target objects to meet a first preset requirement as the target aperture value;
adjusting the aperture to the target aperture value.
5. The method according to claim 4, wherein the first predetermined requirement is that the imaging clarity corresponding to the candidate aperture value among the plurality of candidate aperture values is optimal.
6. The method according to claim 1, characterized in that the method further comprises the following steps to obtain imaging clarity of the at least two target objects:
acquiring a distance relationship between each target object and the imaging device;
determining a sharpness weight for each of the target objects based on the distance relationship;
and carrying out weighting processing on the imaging definition of each target object based on the definition weight to obtain the imaging definition conditions of the at least two target objects.
7. The method of claim 6, wherein the distance relationship between the target object and the imaging device is a distance relationship between the target object and the imaging device.
8. The method of claim 7, wherein said obtaining a distance relationship between each of said target objects and said imaging device comprises:
adjusting the focal length of the imaging equipment, and determining the imaging definition change trend of the target object in the process of adjusting the focal length; determining a distance relation between each target object and the imaging equipment based on the imaging definition variation trend; alternatively, the first and second electrodes may be,
determining the target area of each target object in the imaging picture; and determining the distance relation between the target objects and the imaging device based on the target area of each target object.
9. The method of claim 7, wherein the sharpness weight of a target object closer to the imaging device is greater than the sharpness weight of a target object further from the imaging device.
10. The method of claim 6, wherein before the weighting the imaging sharpness of each of the target objects based on the sharpness weights to obtain the imaging sharpness of the at least two target objects, the method further comprises:
determining a target area of each of the target objects in an imaging sensor;
and determining the imaging definition of the target object based on the imaging picture information of the target area.
11. The method of claim 10, wherein said determining a target area of each of said target objects in an imaging sensor comprises:
dividing the imaging picture into a plurality of first areas, and dividing the imaging area of the imaging sensor into a plurality of second areas;
determining the first area corresponding to each target object;
and taking a second area corresponding to the first area as the target area.
12. The method of claim 1, wherein determining at least two target objects in an imaging frame of an imaging device comprises:
acquiring a target recognition result obtained by performing target recognition on the imaging picture by image recognition equipment, and determining the at least two target objects based on the target recognition result;
and/or, the imaging-related element comprises at least one of an imaging sensor and a lens.
13. An image forming apparatus, characterized by comprising:
an imaging-related element for forming an imaged picture;
the adjusting component is used for adjusting the imaging related element to a target position, wherein the target position enables an imaging plane of the imaging device to be arranged at an angle with a lens plane; and adjusting the aperture of the imaging device to a target aperture value, wherein the target aperture value can enable the imaging definition conditions of at least two target objects in the imaging picture to meet a first preset requirement.
14. An imaging control apparatus comprising a processor and a memory coupled to each other, wherein,
the processor is configured to execute the memory-stored computer program to perform the method of any of claims 1 to 12.
15. The apparatus according to claim 14, characterized in that the imaging control apparatus further comprises:
an imaging-related element for forming an imaged picture;
an adjustment component for performing adjustment of the imaging-related element and aperture in response to control instructions of the processor.
CN202110910045.2A 2021-08-09 2021-08-09 Imaging control method and related equipment Active CN113364963B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110910045.2A CN113364963B (en) 2021-08-09 2021-08-09 Imaging control method and related equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110910045.2A CN113364963B (en) 2021-08-09 2021-08-09 Imaging control method and related equipment

Publications (2)

Publication Number Publication Date
CN113364963A true CN113364963A (en) 2021-09-07
CN113364963B CN113364963B (en) 2021-11-16

Family

ID=77540701

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110910045.2A Active CN113364963B (en) 2021-08-09 2021-08-09 Imaging control method and related equipment

Country Status (1)

Country Link
CN (1) CN113364963B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103813104A (en) * 2014-03-12 2014-05-21 爱佩仪光电技术(深圳)有限公司 Method for realizing shift photography and three-dimensional multi-region automatic focusing through touch screen operation
CN108881736A (en) * 2018-06-30 2018-11-23 浙江大华技术股份有限公司 A kind of iris correction method and device
EP3684046A1 (en) * 2019-01-17 2020-07-22 Canon Kabushiki Kaisha Autofocus in an image-capturing apparatus operating using the scheimpflug principle
CN112703532A (en) * 2020-12-03 2021-04-23 华为技术有限公司 Image processing method, device, equipment and storage medium
CN113225473A (en) * 2020-02-05 2021-08-06 浙江大华技术股份有限公司 Method, apparatus, device and medium for auto-focusing

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103813104A (en) * 2014-03-12 2014-05-21 爱佩仪光电技术(深圳)有限公司 Method for realizing shift photography and three-dimensional multi-region automatic focusing through touch screen operation
CN108881736A (en) * 2018-06-30 2018-11-23 浙江大华技术股份有限公司 A kind of iris correction method and device
EP3684046A1 (en) * 2019-01-17 2020-07-22 Canon Kabushiki Kaisha Autofocus in an image-capturing apparatus operating using the scheimpflug principle
CN113225473A (en) * 2020-02-05 2021-08-06 浙江大华技术股份有限公司 Method, apparatus, device and medium for auto-focusing
CN112703532A (en) * 2020-12-03 2021-04-23 华为技术有限公司 Image processing method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN113364963B (en) 2021-11-16

Similar Documents

Publication Publication Date Title
JP6844038B2 (en) Biological detection methods and devices, electronic devices and storage media
WO2020121973A1 (en) Object identification system, operation processing device, vehicle, lighting tool for vehicle, and training method for classifier
EP3534109B1 (en) Depth measuring method and system
US8306360B2 (en) Device and method for obtaining clear image
US10070038B2 (en) Image processing apparatus and method calculates distance information in a depth direction of an object in an image using two images whose blur is different
CN111598065B (en) Depth image acquisition method, living body identification method, apparatus, circuit, and medium
US8055016B2 (en) Apparatus and method for normalizing face image used for detecting drowsy driving
CN112529951A (en) Method and device for acquiring extended depth of field image and electronic equipment
CN106375666B (en) A kind of Atomatic focusing method and device based on license plate
CN108401457A (en) A kind of control method of exposure, device and unmanned plane
KR20160048140A (en) Method and apparatus for generating an all-in-focus image
JP2021503659A (en) Biodetection methods, devices and systems, electronic devices and storage media
CN108833795B (en) Focusing method and device of image acquisition equipment
CN103593647A (en) Biological characteristic imaging method and equipment
CN108234874B (en) Method and device for adjusting imaging precision of somatosensory camera
CN108702457B (en) Method, apparatus and computer-readable storage medium for automatic image correction
JP2012113460A (en) Information processor and method, and program
CN107101612A (en) A kind of target distance measurement method and system
CN116051391B (en) Image processing method and electronic equipment
CN111105370B (en) Image processing method, image processing apparatus, electronic device, and readable storage medium
WO2017101292A1 (en) Autofocusing method, device and system
CN116051736A (en) Three-dimensional reconstruction method, device, edge equipment and storage medium
CN113364963B (en) Imaging control method and related equipment
CN114363522A (en) Photographing method and related device
KR101635602B1 (en) Method and apparatus for iris scanning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant