CN113613024A - Video preprocessing method and device - Google Patents

Video preprocessing method and device Download PDF

Info

Publication number
CN113613024A
CN113613024A CN202110919840.8A CN202110919840A CN113613024A CN 113613024 A CN113613024 A CN 113613024A CN 202110919840 A CN202110919840 A CN 202110919840A CN 113613024 A CN113613024 A CN 113613024A
Authority
CN
China
Prior art keywords
video frame
preprocessing
current video
processing
mode
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110919840.8A
Other languages
Chinese (zh)
Other versions
CN113613024B (en
Inventor
鲁方波
汪贤
徐晓光
赵瑞前
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Kingsoft Cloud Network Technology Co Ltd
Original Assignee
Beijing Kingsoft Cloud Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Kingsoft Cloud Network Technology Co Ltd filed Critical Beijing Kingsoft Cloud Network Technology Co Ltd
Publication of CN113613024A publication Critical patent/CN113613024A/en
Application granted granted Critical
Publication of CN113613024B publication Critical patent/CN113613024B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/87Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving scene cut or scene change detection in combination with video compression

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The embodiment of the application relates to the technical field of multimedia, and provides a video preprocessing method and video preprocessing equipment. The video preprocessing method comprises the following steps: acquiring video frame attribute parameter information of a current video frame, determining a corresponding switched target preprocessing mode from at least two preset preprocessing modes under the condition that the video frame attribute parameter information meets a preprocessing mode switching condition, and further executing preprocessing on the current video frame according to the target preprocessing mode. It can be seen that in the technical solution of the embodiment of the present application, a plurality of preprocessing modes are configured, and then a preprocessing mode matched with a video frame is selected from the plurality of preprocessing modes according to the attribute parameter information of the video frame, so that the video frame is preprocessed by using the matched preprocessing mode. Therefore, the quality difference of the video after the preprocessing of each video frame of the video can be ensured to be relatively small, and the watching experience of a user can be improved.

Description

Video preprocessing method and device
Technical Field
The embodiment of the application relates to the technical field of multimedia, in particular to a video preprocessing method and video preprocessing equipment.
Background
Video pre-processing is the process of processing the video data obtained for encoding before video encoding. One common video pre-processing method is to process each video frame of the video according to a pre-configured processing flow and algorithm.
In an actual implementation scenario, changes in the displayed content of a video, changes in the network status, and the like may cause different qualities (e.g., image quality, noise, and the like) of different video frames of the same video. Based on this, the same processing flow and algorithm are adopted to process each video frame of the video, which results in larger quality difference after processing each video frame, thereby reducing the watching experience of the user.
Disclosure of Invention
The embodiment of the application provides a video preprocessing method and video preprocessing equipment, which are used for solving the problem that the quality difference of each processed video frame is large due to the existing video preprocessing method.
In a first aspect, an embodiment of the present application provides a video preprocessing method, where the method includes:
acquiring video frame attribute parameter information of a current video frame;
if the video frame attribute parameter information meets the switching condition of the preprocessing modes, determining a corresponding switched target preprocessing mode from at least two preset preprocessing modes;
and performing preprocessing on the current video frame according to the target preprocessing mode.
In a second aspect, an embodiment of the present application provides a video preprocessing apparatus, including:
the acquisition module is used for acquiring the video frame attribute parameter information of the current video frame;
the processing mode determining module is used for determining a corresponding switched target preprocessing mode from at least two preset preprocessing modes if the video frame attribute parameter information meets the preprocessing mode switching condition;
and the processing module is used for executing preprocessing on the current video frame according to the target preprocessing mode.
In some possible embodiments, the preset at least two preprocessing modes include: no processing, color enhancement processing mode, sharpness enhancement processing mode, and noise reduction processing mode.
In some possible embodiments, the video frame attribute parameter information includes a scene type corresponding to the current video frame, and the obtaining module is further configured to extract a first video frame feature of the current video frame and a second video frame feature of a previous video frame of the current video frame; similarity calculation is carried out on the first video frame characteristics and the second video frame characteristics, and the similarity of the current video frame and the previous video frame of the current video frame is determined; if the similarity is smaller than a preset similarity threshold, determining that the scene of the current video frame is switched relative to the previous video frame of the current video frame; and determining a scene category of the current video frame.
In some possible embodiments, the processing mode determining module is further configured to determine the color enhancement processing mode as the target processing mode if the scene type of the current video frame is the target scene type.
In some possible embodiments, the video frame attribute parameter information further comprises at least one of: the definition of the current video frame, the resolution of the current video frame and the corresponding frame rate of the current video frame;
the processing mode determining module is further configured to determine that no processing is performed as the target processing mode if the definition of the current video frame is greater than a first threshold, or the resolution of the current video frame is greater than a second threshold, or the frame rate corresponding to the current video frame is greater than a third threshold;
the processing mode determining module is further used for determining the definition enhancement processing as a target processing mode if the definition of the current video frame is smaller than a fourth threshold, wherein the fourth threshold is smaller than the first threshold;
and the processing mode determining module is further used for determining the noise reduction processing mode as the target processing mode if the noise of the current video frame is greater than the fifth threshold.
In some possible embodiments, the scene category of the current video frame is a target scene category, including: the current video frame is a game scene or an animation scene.
And the processing mode determining module is further used for determining a target preprocessing mode according to the priority of each preset condition under the condition that the video frame attribute parameter information meets at least two preprocessing mode switching conditions.
In a third aspect, an electronic device is provided that includes memory and one or more processors; wherein the memory is for storing a computer program; the computer program, when executed by the processor, causes the electronic device to perform the video pre-processing method of the first aspect.
In a fourth aspect, the present application provides a computer-readable storage medium, which stores instructions that, when executed on a computer, cause the computer to perform part or all of the steps of the video preprocessing method according to the first aspect.
In a fifth aspect, the present application provides a computer program product, where the computer program product includes computer program code, and when the computer program code runs on a computer, the computer is caused to implement the video preprocessing method according to the first aspect.
In order to solve the technical problem of the existing scheme, the embodiment of the application pre-configures at least two preprocessing modes. After acquiring the video frame attribute parameter information of the current video frame, if the video frame attribute parameter information meets the switching condition of the preprocessing modes, determining a corresponding switched target preprocessing mode from at least two pre-configured preprocessing modes, and then preprocessing the current video frame according to the determined target preprocessing mode. It can be seen that in the technical solution of the embodiment of the present application, a plurality of preprocessing modes are configured, and then, according to the attribute parameter information of the video frame, a preprocessing mode matched with the current video frame is selected from the plurality of preprocessing modes, so that the matching processing mode is used to execute preprocessing on the video frame. In this way, the preprocessing mode is adaptively selected from the plurality of preprocessing modes, so that the quality difference of the processed video frames of the video can be relatively small, and the watching experience of the user can be improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings required to be used in the embodiments of the present application will be briefly described below. It should be understood that other figures may be derived from these figures by those of ordinary skill in the art without inventive exercise.
Fig. 1 is a schematic diagram of an exemplary application scenario provided in an embodiment of the present application;
fig. 2 is a flowchart of an exemplary method of video pre-processing 100 provided in an embodiment of the present application;
FIG. 3 is a schematic diagram of an exemplary interface of a video frame provided by an embodiment of the present application;
FIG. 4 is a schematic diagram of another exemplary interface of a video frame provided by an embodiment of the present application;
fig. 5A is a schematic diagram illustrating an exemplary composition of a video preprocessing apparatus 50 according to an embodiment of the present disclosure;
fig. 5B is a schematic structural diagram of an electronic device 51 provided in the embodiment of the present application.
Detailed Description
The following describes technical solutions of the embodiments of the present application with reference to the drawings in the embodiments of the present application.
The terminology used in the following examples of the present application is for the purpose of describing particular embodiments and is not intended to be limiting of the technical solutions of the present application. As used in the specification of the present application and the appended claims, the singular forms "a", "an", "the" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It should also be understood that although the terms first, second, etc. may be used in the following embodiments to describe a class of objects, the objects should not be limited to these terms. These terms are used to distinguish between particular objects of the class of objects. For example, the terms first, second, etc. are used in the following embodiments to describe video frame characteristics, but the video frame characteristics are not limited to these terms. These terms are only used to distinguish features of different video frames. The following embodiments may adopt the terms first, second, etc. to describe other class objects in the same way, and are not described herein again.
The following describes application scenarios and electronic devices of embodiments of the present application, and embodiments of video preprocessing methods performed using such electronic devices.
The embodiment of the application can be applied to live broadcasting, screen projection and the like, and relates to a video transmission scene. The embodiments of the present application relate to the processing stage prior to video transmission. Referring to fig. 1, fig. 1 illustrates an exemplary application scenario diagram according to an embodiment of the present application. Before sending a video, the electronic device preprocesses the video to be sent to obtain video data for encoding, and then encodes the obtained video data to obtain encoded data for sending.
Optionally, the video to be sent according to the embodiment of the present application may include a video signal that is acquired or captured by an electronic device in real time through a video capture device, may also include a video file that is pre-stored locally by the electronic device, and may also include a video file that is received through a network. Video capture devices include, for example, cameras, Virtual Reality (VR) devices, and Augmented Reality (AR) devices, among others.
The video may be understood as a plurality of Frame images (which may also be described as video frames in the art) played according to a certain sequence and a Frame Rate (Frame Rate), and based on this, the video is preprocessed, and in an actual execution process, each video Frame of the video is preprocessed. The purpose of video preprocessing may be to ensure that the image quality corresponding to the video frame data used for encoding reaches a certain level and to reduce the code rate of the video frame data used for encoding. In view of this, video pre-processing may include cropping, color format conversion (e.g., converting a video frame from Red Green Blue (RGB) format to luma and chroma (YUV) format), toning or denoising, etc. of a video frame.
The video coding according to the embodiments of the present application may be an operation on pre-processed video data according to a video coding standard (e.g., high efficiency video coding h.265 standard) or an industry standard. The standards described herein may include ITU-T H.261, ISO/IECMPEG-1Visual, ITU-T H.262 or ISO/IECMPEG-2Visual, ITU-T H.263, ISO/IECMPEG-4Visual, ITU-T H.264 (also known as ISO/IECMPEG-4AVC), and the like. The embodiments of the present application do not limit this.
Embodiments of the present application relate to electronic devices implemented as video transmission devices, which may be implemented to include desktop computers, mobile computing devices, notebook (e.g., laptop) computers, tablet computers, smart phones, televisions, cameras, digital media players, video game consoles, in-vehicle computers, or other devices that support video pre-processing functions. Optionally, the electronic device may use a Central Processing Unit (CPU) or a Graphics Processing Unit (GPU) to implement part or all of the steps of the video preprocessing method according to the embodiment of the present application. Alternatively, in an actual implementation scenario, the video pre-processing function may be deployed in another device independent from the video encoding device, or may be deployed in the video encoding device. The embodiments of the present application do not limit this.
The embodiment of the application provides a video preprocessing method and video preprocessing equipment, wherein a plurality of preprocessing modes are configured in the equipment in advance. Then, the device selects a preprocessing mode matched with the video frame from the plurality of preprocessing modes according to the attribute parameter information of the video frame to be processed, so as to execute preprocessing on the video frame by using the matched preprocessing mode. Therefore, the quality difference of the video after the preprocessing of each video frame of the video can be ensured to be relatively small, and the watching experience of a user can be improved.
The technical solutions of the embodiments of the present application and the technical effects produced by the technical solutions of the present application will be described below through descriptions of exemplary embodiments.
Referring to fig. 2, fig. 2 illustrates an exemplary video pre-processing method 100 (hereinafter referred to as method 100). This implementation is illustrated by way of example as the sending device executing method 100, which may be an electronic device as described above. The method 100 includes the steps of:
in step S101, the transmitting device acquires video frame attribute parameter information of a current video frame.
Wherein, the current video frame is a video frame of the video to be processed. Optionally, the first video frame and the last video frame of the video to be processed generally have less influence on the overall appearance of the video to be processed, and based on this, the current video frame may be any video frame in the video to be processed except the first video frame and the last video frame. It should be understood that the video to be processed here has the same meaning as the video to be transmitted illustrated in fig. 1.
In combination with the foregoing description of the video source to be transmitted, optionally, the transmitting device may acquire or capture the current video frame in real time through the video capturing device, or obtain the current video frame from a locally pre-stored video file, or receive the current video frame through the network. Illustratively, the manner in which the sending device obtains the current video frame is relevant to the implementation scenario. For example, in a live scene, a sending device may capture or capture a current video frame in real time through a video capture device; for another example, in a screen-shot scene, the sending device may obtain a current video frame from a locally pre-stored video file.
Optionally, the video frame attribute parameter information may include at least one of a scene type corresponding to the current video frame, a definition of the current video frame, a resolution of the current video frame, and a frame rate corresponding to the current video frame.
Optionally, the "scene" described in the embodiment of the present application may refer to a context expressed by the picture content of the video frame, for example, the context expressed by the video frame illustrated in fig. 3 is a game, and then, what is shown in fig. 3 is a game scene; for another example, the scenario represented by the video frame illustrated in fig. 4 is a close-up, and then fig. 4 shows a close-up scene. In order to ensure the viewing experience of the user, the video frames corresponding to different scenes can be displayed with different image qualities. Based on this, the embodiments of the present application may classify the scene types according to the requirements for image quality, and use the type corresponding to the scene having a certain requirement for image quality as the target scene type. For example, in a live video, a game scene, an animation scene, and a feature scene are scenes that have a certain requirement for image quality, and then, scene categories corresponding to the game scene, the animation scene, and the feature scene may be set as target scene categories.
In an actual implementation scene, a plurality of consecutive video frames of a video to be processed usually show the same scene, and the sending device may determine whether the video to be processed switches the scene according to the current video frame, and if it is determined that the video to be processed switches the scene, the sending device may further determine the scene type of the current video frame.
For example, the sending device may extract a first video frame feature of the current video frame and a second video frame feature of a video frame previous to the current video frame, and then perform similarity calculation on the first video frame feature and the second video frame feature to determine the similarity between the current video frame and the video frame previous to the current video frame. And if the similarity is smaller than a preset similarity threshold, determining that the scene of the current video frame is switched relative to the previous video frame of the current video frame, and determining the scene type of the current video frame.
Illustratively, the first video frame feature and the second video frame feature may each be implemented as a gray histogram feature.
Illustratively, the similarity may be characterized by a Distance, where the Distance includes a Euclidean Distance (Euclidean Distance) or a cosine Distance. The larger the euclidean distance is, the higher the similarity between the two features is, and the smaller the euclidean distance is, the lower the similarity between the two features is. The smaller the cosine distance, the higher the similarity of the two features, and the larger the cosine distance, the lower the similarity of the two features.
The definition refers to the definition of each detail shadow and its boundary on the video frame, and can be expressed by resolution. The resolution refers to the total number of pixels (P) in the video frame, and is also expressed as the size or dimension of the video frame, and may be generally represented by the number of pixels in the width direction and the number of pixels in the height direction, for example, the resolution of the current video frame is 1024P 768P, which represents an image in which the width of the current video frame is 1024 pixels and the height of the current video frame is 768 pixels. The Frame rate refers to the number of video Frames (FPS) included in a video per second.
In step S102, if the video frame attribute parameter information satisfies the pre-processing mode switching condition, a corresponding switched target pre-processing mode is determined from at least two pre-processing modes.
At least two preprocessing modes can be configured in advance in the sending device, and optionally, the at least two preprocessing modes can include no processing, a color enhancement processing mode, a definition enhancement processing mode, a noise reduction processing mode, and the like. The color enhancement processing mode, the definition enhancement processing mode and the noise reduction processing mode can respectively realize related processing functions through processing algorithms or combination of the processing algorithms. The mode of no processing does not contain any processing algorithm. Optionally, if two preprocessing modes include the same processing algorithm, the parameters of the same processing algorithm in the two preprocessing modes may be different. Optionally, the processing algorithm may include a noise reduction algorithm, a color enhancement algorithm, a sharpening algorithm, a contrast enhancement algorithm, and the like.
It should be understood that the above is only an illustrative description of the processing mode and the processing algorithm, and the embodiments of the present application are not limited thereto. In other embodiments, the at least two pre-processing modes may also include more or fewer processing modes and processing algorithms.
As can be seen from the foregoing description of the video frame attribute parameter information, the video frame attribute parameter information includes at least one item of information, based on which the video frame attribute parameter information satisfies the preprocessing mode switching condition, including at least one condition, and correspondingly, the target preprocessing mode after switching is determined from at least two preset preprocessing modes, which may also include multiple conditions.
Optionally, if the scene type of the current video frame is the target scene type, the color enhancement processing mode is determined as the target processing mode. And if the definition of the current video frame is greater than a first threshold, or the resolution of the current video frame is greater than a second threshold, or the frame rate corresponding to the current video frame is greater than a third threshold, determining that no processing is performed as the target processing mode. And if the definition of the current video frame is smaller than a fourth threshold value, determining the definition enhancement processing as a target processing mode, wherein the fourth threshold value is smaller than the first threshold value. And if the noise of the current video frame is larger than the fifth threshold, determining the noise reduction processing mode as the target processing mode.
Illustratively, the scene category of the current video frame is a target scene category, including: the current video frame is a game scene or an animation scene.
In step S103, the transmitting device performs preprocessing on the current video frame according to the target preprocessing mode.
After determining the target processing mode, the transmitting device may pre-process the current video frame according to an algorithm of the target processing mode.
Optionally, when the video frame attribute parameter information satisfies at least two conditions for switching the pre-processing modes, the sending device may determine the target pre-processing mode according to the priorities of the preset conditions.
It is to be understood that the above priority of each information is only an illustrative description, and does not limit the embodiments of the present application. In the actual implementation process, the priority of each piece of information can be flexibly set according to the requirement. The embodiments of the present application do not limit this.
For example:
the first pretreatment mode is as follows: the current video frame satisfies at least one of: the resolution of the current video frame is greater than 540p, the frame rate of the current video frame is greater than 25, or the definition of the current video frame is greater than 50, and the sending device does not pre-process the current video frame according to the target processing mode (i.e. does not process).
And a second pretreatment mode: the resolution of the current video frame is less than 540p, the frame rate of the current video frame is equal to 25 and the definition of the current video frame is less than 20, and the sending device: and a definition enhancement processing mode for performing definition enhancement processing on the current video frame.
A third pretreatment mode: and (3) scenes in which the conditions of the first preprocessing mode and the conditions of the second preprocessing mode are not satisfied. In the third preprocessing mode, the current video frame shows the cartoon and the game, and the sending equipment is used for processing the video frame according to the target processing mode: and the color enhancement processing mode is used for carrying out color enhancement processing on the current video frame.
The pretreatment mode is four: and (3) a scene in which the condition of the first preprocessing mode, the condition of the second preprocessing mode and the condition of the third preprocessing mode are not satisfied. In the fourth preprocessing mode, the noise intensity of the current video frame is greater than 5, and the sending equipment: and the noise reduction processing mode is used for carrying out noise reduction processing on the current video frame.
It should be understood that the above first to fourth preprocessing modes are schematic descriptions and do not limit the embodiments of the present application. In some other embodiments, the threshold of each condition may be another value, and if the current video frame meets another condition, the sending device may further perform another preprocessing operation on the current video frame. The embodiments of the present application do not limit this.
In addition, the at least two preprocessing modes, the content of the attribute information, various conditions corresponding to the attribute information, and the like are all schematically described, and the embodiments of the present application are not limited. In the implementation process of the present technical solution, the at least two preprocessing modes, the content of the attribute information, and various conditions corresponding to the attribute information may include other examples. The embodiments of the present application are not illustrated.
In summary, in the implementation manner of the embodiment of the present application, at least two preprocessing modes are preconfigured. After acquiring the video frame attribute parameter information of the current video frame, if the video frame attribute parameter information meets the switching condition of the preprocessing modes, determining a corresponding switched target preprocessing mode from at least two pre-configured preprocessing modes, and then preprocessing the current video frame according to the determined target preprocessing mode. It can be seen that in the technical solution of the embodiment of the present application, a plurality of preprocessing modes are configured, and then, according to the attribute parameter information of the video frame, a preprocessing mode matched with the current video frame is selected from the plurality of preprocessing modes, so that the matching processing mode is used to execute preprocessing on the video frame. In this way, the preprocessing mode is adaptively selected from the plurality of preprocessing modes, so that the quality difference of the processed video frames of the video can be relatively small, and the watching experience of the user can be improved.
The foregoing embodiments describe various embodiments of the video preprocessing method provided in the embodiments of the present application from the perspective of actions performed by devices, such as obtaining attribute parameter information of a video frame, detecting whether a switching condition is satisfied, and determining a target preprocessing model. It should be understood that, in the embodiments of the present application, the functions described above may be implemented in hardware or a combination of hardware and computer software corresponding to the processing steps of obtaining the attribute parameter information of the video frame, detecting whether the switching condition is satisfied, determining the target preprocessing model, and the like. Whether a function is performed as hardware or computer software drives hardware depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
For example, if the above implementation steps implement the corresponding functions through software modules, as shown in fig. 5A, the embodiment of the present application provides a video preprocessing apparatus 50. The video pre-processing apparatus 50 may include an acquisition module 501, a processing mode determination module 502, and a processing module 503. The video pre-processing device 50 may be used to perform some or all of the operations of the method 100 described above.
For example: the obtaining module 501 may be configured to obtain video frame attribute parameter information of a current video frame. The processing mode determining module 502 may be configured to determine a corresponding switched target preprocessing mode from at least two preset preprocessing modes if the video frame attribute parameter information satisfies the preprocessing mode switching condition. The processing module 503 may be configured to perform pre-processing on the current video frame according to the target pre-processing mode.
Therefore, the video preprocessing apparatus 50 provided in the embodiment of the present application is preconfigured with at least two preprocessing modes. After acquiring the video frame attribute parameter information of the current video frame, if the video frame attribute parameter information meets the switching condition of the preprocessing modes, determining a corresponding switched target preprocessing mode from at least two pre-configured preprocessing modes, and then preprocessing the current video frame according to the determined target preprocessing mode. In this way, the preprocessing mode is adaptively selected from the plurality of preprocessing modes, so that the quality difference of the processed video frames of the video can be relatively small, and the watching experience of the user can be improved.
Optionally, the preset at least two preprocessing modes include: no processing, color enhancement processing mode, sharpness enhancement processing mode, and noise reduction processing mode.
Optionally, the video frame attribute parameter information includes a scene type corresponding to the current video frame, and the obtaining module 501 is further configured to extract a first video frame feature of the current video frame and a second video frame feature of a previous video frame of the current video frame; similarity calculation is carried out on the first video frame characteristics and the second video frame characteristics, and the similarity of the current video frame and the previous video frame of the current video frame is determined; if the similarity is smaller than a preset similarity threshold, determining that the scene of the current video frame is switched relative to the previous video frame of the current video frame; and determining a scene category of the current video frame.
Optionally, the processing mode determining module 502 is further configured to determine the color enhancement processing mode as the target processing mode if the scene type of the current video frame is the target scene type.
Optionally, the video frame attribute parameter information further includes at least one of: the definition of the current video frame, the resolution of the current video frame, and the corresponding frame rate of the current video frame. In this example, the processing mode determining module 502 is further configured to determine that no processing is performed as the target processing mode if the definition of the current video frame is greater than a first threshold, or the resolution of the current video frame is greater than a second threshold, or the frame rate corresponding to the current video frame is greater than a third threshold. The processing mode determining module 502 is further configured to determine the sharpness enhancement processing as the target processing mode if the sharpness of the current video frame is smaller than a fourth threshold, where the fourth threshold is smaller than the first threshold. The processing mode determining module 502 is further configured to determine the noise reduction processing mode as the target processing mode if the noise of the current video frame is greater than the fifth threshold.
Optionally, the scene category of the current video frame is a target scene category, and includes: the current video frame is a game scene or an animation scene.
Optionally, the processing mode determining module 502 is further configured to determine the target preprocessing mode according to the priority of each preset condition when the video frame attribute parameter information satisfies at least two preprocessing mode switching conditions.
It is understood that the above division of each module/unit is only a division of a logic function, and in actual implementation, the functions of the above modules may be integrated into a hardware entity, for example, the functions of the processing mode determining module 502 and the processing module 503 may be integrated into a processor, the function of the obtaining module 501 may be integrated into a transceiver, and programs and instructions for implementing the functions of the above modules may be maintained in a memory.
For example, fig. 5B provides an electronic device 51, and the electronic device 51 may implement the functions of the aforementioned transmitting device. The electronic device 51 includes a processor 511, a transceiver 512, and a memory 513. The transceiver 512 is used for performing the information acquisition in the method 100. Alternatively, the transceiver 512 may be implemented as an image receiving device, such as a camera. The memory 513 may be used to store the pre-installed programs/codes of the video pre-processing apparatus 50, or store the codes for the processor 301 to execute, and the like. The processor 511, when executing the code stored in the memory 513, causes the electronic device 51 to perform some or all of the operations of the video pre-processing method of the method 100 described above.
The specific implementation process is described in the above exemplary embodiment of the method 100, and is not described in detail here.
In a specific implementation, corresponding to the foregoing electronic device, an embodiment of the present application further provides a computer storage medium, where the computer storage medium disposed in the electronic device may store a program, and when the program is executed, part or all of the steps in each embodiment of the video preprocessing method including the method 100 may be implemented. The storage medium may be a magnetic disk, an optical disk, a read-only memory (ROM), a Random Access Memory (RAM), or the like.
One or more of the above modules or units may be implemented in software, hardware or a combination of both. When any of the above modules or units are implemented in software, which is present as computer program instructions and stored in a memory, a processor may be used to execute the program instructions and implement the above method flows. The processor may include, but is not limited to, at least one of: various computing devices that run software, such as a Central Processing Unit (CPU), a microprocessor, a Digital Signal Processor (DSP), a Microcontroller (MCU), or an artificial intelligence processor, may each include one or more cores for executing software instructions to perform operations or processing. The processor may be built in an SoC (system on chip) or an Application Specific Integrated Circuit (ASIC), or may be a separate semiconductor chip. The processor may further include a necessary hardware accelerator such as a Field Programmable Gate Array (FPGA), a PLD (programmable logic device), or a logic circuit for implementing a dedicated logic operation, in addition to a core for executing software instructions to perform an operation or a process.
When the above modules or units are implemented in hardware, the hardware may be any one or any combination of a CPU, a microprocessor, a DSP, an MCU, an artificial intelligence processor, an ASIC, an SoC, an FPGA, a PLD, a dedicated digital circuit, a hardware accelerator, or a discrete device that is not integrated, which may run necessary software or is independent of software to perform the above method flows.
Further, a bus interface may also be included in FIG. 5B, which may include any number of interconnected buses and bridges, with one or more processors, represented by a processor, and various circuits of memory, represented by memory, linked together. The bus interface may also link together various other circuits such as peripherals, voltage regulators, power management circuits, and the like, which are well known in the art, and therefore, will not be described any further herein. The bus interface provides an interface. The transceiver provides a means for communicating with various other apparatus over a transmission medium. The processor is responsible for managing the bus architecture and the usual processing, and the memory may store data used by the processor in performing operations.
When the above modules or units are implemented using software, they may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When loaded and executed on a computer, cause the processes or functions described in accordance with the embodiments of the invention to occur, in whole or in part. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored in a computer readable storage medium or transmitted from one computer readable storage medium to another, for example, from one website site, computer, server, or data center to another website site, computer, server, or data center via wired (e.g., coaxial cable, fiber optic, Digital Subscriber Line (DSL)) or wireless (e.g., infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that incorporates one or more of the available media. The usable medium may be a magnetic medium (e.g., floppy Disk, hard Disk, magnetic tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., Solid State Disk (SSD)), among others.
It should be understood that, in the various embodiments of the present application, the size of the serial number of each process does not mean the execution sequence, and the execution sequence of each process should be determined by the function and the inherent logic thereof, and should not constitute any limitation to the implementation process of the embodiments.
All parts of the specification are described in a progressive mode, the same and similar parts of all embodiments can be referred to each other, and each embodiment is mainly introduced to be different from other embodiments. In particular, as to the apparatus and system embodiments, since they are substantially similar to the method embodiments, the description is relatively simple and reference may be made to the description of the method embodiments in relevant places.
While alternative embodiments of the present application have been described, additional variations and modifications of these embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. Therefore, it is intended that the appended claims be interpreted as including preferred embodiments and all alterations and modifications as fall within the scope of the application.
The above-mentioned embodiments, objects, technical solutions and advantages of the present application are further described in detail, it should be understood that the above-mentioned embodiments are only examples of the present application, and are not intended to limit the scope of the present application, and any modifications, equivalent substitutions, improvements and the like made on the basis of the technical solutions of the present application should be included in the scope of the present invention.

Claims (10)

1. A method for video pre-processing, the method comprising:
acquiring video frame attribute parameter information of a current video frame;
if the video frame attribute parameter information meets the switching condition of the preprocessing modes, determining a corresponding switched target preprocessing mode from at least two preset preprocessing modes;
and performing preprocessing on the current video frame according to the target preprocessing mode.
2. The method of claim 1, wherein the preset at least two pre-processing modes comprise: no processing, color enhancement processing mode, sharpness enhancement processing mode, and noise reduction processing mode.
3. The method according to claim 1 or 2, wherein the video frame attribute parameter information comprises a scene category corresponding to the current video frame,
the acquiring of the video frame attribute parameter information of the current video frame includes:
extracting first video frame characteristics of a current video frame and second video frame characteristics of a previous video frame of the current video frame;
similarity calculation is carried out on the first video frame characteristics and the second video frame characteristics, and the similarity of the current video frame and a video frame before the current video frame is determined;
if the similarity is smaller than a preset similarity threshold, determining that the current video frame is switched to a scene relative to a previous video frame of the current video frame;
determining a scene category of the current video frame.
4. The method according to claim 3, wherein if the video frame attribute parameter information satisfies a pre-processing mode switching condition, determining a corresponding switched target pre-processing mode from at least two pre-processing modes, comprising:
and if the scene type of the current video frame is the target scene type, determining a color enhancement processing mode as the target processing mode.
5. The method according to claim 1 or 2, wherein the video frame property parameter information further comprises at least one of: the definition of the current video frame, the resolution of the current video frame and the corresponding frame rate of the current video frame;
if the video frame attribute parameter information meets the preprocessing mode switching condition, determining a corresponding switched target preprocessing mode from at least two preset preprocessing modes, wherein the target preprocessing mode comprises at least one of the following items:
if the definition of the current video frame is greater than a first threshold, or the resolution of the current video frame is greater than a second threshold, or the frame rate corresponding to the current video frame is greater than a third threshold, determining that no processing is performed as the target processing mode;
if the definition of the current video frame is smaller than a fourth threshold value, determining definition enhancement processing as the target processing mode, wherein the fourth threshold value is smaller than the first threshold value;
and if the noise of the current video frame is larger than a fifth threshold value, determining a noise reduction processing mode as the target processing mode.
6. The method according to claim 3 or 4, wherein the scene category of the current video frame is a target scene category, comprising:
the current video frame is a game scene or an animation scene.
7. The method according to claim 3 or 4, wherein if the video frame attribute parameter information satisfies a pre-processing mode switching condition, determining a corresponding switched target pre-processing mode from at least two pre-processing modes, further comprising:
and when the video frame attribute parameter information meets at least two preprocessing mode switching conditions, determining the target preprocessing mode according to the priority of each preset condition.
8. A video pre-processing apparatus, the apparatus comprising:
the acquisition module is used for acquiring the video frame attribute parameter information of the current video frame;
a processing mode determining module, configured to determine, if the video frame attribute parameter information satisfies a pre-processing mode switching condition, a corresponding switched target pre-processing mode from at least two pre-processing modes;
and the processing module is used for executing preprocessing on the current video frame according to the target preprocessing mode.
9. An electronic device, wherein the electronic device comprises memory and one or more processors; wherein the memory is for storing a computer program; the computer program, when executed by the processor, causes the electronic device to perform the video pre-processing method of any of claims 1 to 7.
10. A computer-readable storage medium, characterized in that it stores a computer program which, when run on a computer, causes the computer to perform the video pre-processing method according to any one of claims 1 to 7.
CN202110919840.8A 2021-08-09 2021-08-11 Video preprocessing method and device Active CN113613024B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110909960X 2021-08-09
CN202110909960 2021-08-09

Publications (2)

Publication Number Publication Date
CN113613024A true CN113613024A (en) 2021-11-05
CN113613024B CN113613024B (en) 2023-04-25

Family

ID=78308242

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110919840.8A Active CN113613024B (en) 2021-08-09 2021-08-11 Video preprocessing method and device

Country Status (1)

Country Link
CN (1) CN113613024B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114302175A (en) * 2021-12-01 2022-04-08 阿里巴巴(中国)有限公司 Video processing method and device

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104581199A (en) * 2014-12-12 2015-04-29 百视通网络电视技术发展有限责任公司 Video processing system and method
CN109905711A (en) * 2019-02-28 2019-06-18 深圳英飞拓智能技术有限公司 A kind of processing method of image, system and terminal device
US20200012854A1 (en) * 2017-09-08 2020-01-09 Tencent Technology (Shenzhen) Company Ltd Processing method for augmented reality scene, terminal device, system, and computer storage medium
CN110958469A (en) * 2019-12-13 2020-04-03 联想(北京)有限公司 Video processing method and device, electronic equipment and storage medium
CN111698553A (en) * 2020-05-29 2020-09-22 维沃移动通信有限公司 Video processing method and device, electronic equipment and readable storage medium
CN112312231A (en) * 2019-07-31 2021-02-02 北京金山云网络技术有限公司 Video image coding method and device, electronic equipment and medium
CN112805990A (en) * 2018-11-15 2021-05-14 深圳市欢太科技有限公司 Video processing method and device, electronic equipment and computer readable storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104581199A (en) * 2014-12-12 2015-04-29 百视通网络电视技术发展有限责任公司 Video processing system and method
US20200012854A1 (en) * 2017-09-08 2020-01-09 Tencent Technology (Shenzhen) Company Ltd Processing method for augmented reality scene, terminal device, system, and computer storage medium
CN112805990A (en) * 2018-11-15 2021-05-14 深圳市欢太科技有限公司 Video processing method and device, electronic equipment and computer readable storage medium
CN109905711A (en) * 2019-02-28 2019-06-18 深圳英飞拓智能技术有限公司 A kind of processing method of image, system and terminal device
CN112312231A (en) * 2019-07-31 2021-02-02 北京金山云网络技术有限公司 Video image coding method and device, electronic equipment and medium
CN110958469A (en) * 2019-12-13 2020-04-03 联想(北京)有限公司 Video processing method and device, electronic equipment and storage medium
CN111698553A (en) * 2020-05-29 2020-09-22 维沃移动通信有限公司 Video processing method and device, electronic equipment and readable storage medium

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114302175A (en) * 2021-12-01 2022-04-08 阿里巴巴(中国)有限公司 Video processing method and device

Also Published As

Publication number Publication date
CN113613024B (en) 2023-04-25

Similar Documents

Publication Publication Date Title
US10977809B2 (en) Detecting motion dragging artifacts for dynamic adjustment of frame rate conversion settings
CN111681167B (en) Image quality adjusting method and device, storage medium and electronic equipment
US10887614B2 (en) Adaptive thresholding for computer vision on low bitrate compressed video streams
US20150117540A1 (en) Coding apparatus, decoding apparatus, coding data, coding method, decoding method, and program
JP7359521B2 (en) Image processing method and device
US10354394B2 (en) Dynamic adjustment of frame rate conversion settings
CN112788235B (en) Image processing method, image processing device, terminal equipment and computer readable storage medium
CN113518185A (en) Video conversion processing method and device, computer readable medium and electronic equipment
CN112954393A (en) Target tracking method, system, storage medium and terminal based on video coding
US11562772B2 (en) Video processing method, electronic device, and storage medium
US8630500B2 (en) Method for the encoding by segmentation of a picture
CN111738951B (en) Image processing method and device
CN110691246B (en) Video coding method and device and electronic equipment
CN113613024B (en) Video preprocessing method and device
CN109120979B (en) Video enhancement control method and device and electronic equipment
CN116980604A (en) Video encoding method, video decoding method and related equipment
CN111383289A (en) Image processing method, image processing device, terminal equipment and computer readable storage medium
CN113542864B (en) Video splash screen area detection method, device and equipment and readable storage medium
US20240153033A1 (en) Method and system of automatic content-dependent image processing algorithm selection
CN108933945B (en) GIF picture compression method, device and storage medium
CN112488933A (en) Video detail enhancement method and device, mobile terminal and storage medium
CN110969672A (en) Image compression method and device
US20230306687A1 (en) Mesh zippering
CN115965616B (en) Iris image processing method and device and electronic equipment
US20230025378A1 (en) Task-driven machine learning-based representation and compression of point cloud geometry

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant