WO2020077914A1 - 图像处理方法、装置、硬件装置 - Google Patents

图像处理方法、装置、硬件装置 Download PDF

Info

Publication number
WO2020077914A1
WO2020077914A1 PCT/CN2019/073084 CN2019073084W WO2020077914A1 WO 2020077914 A1 WO2020077914 A1 WO 2020077914A1 CN 2019073084 W CN2019073084 W CN 2019073084W WO 2020077914 A1 WO2020077914 A1 WO 2020077914A1
Authority
WO
WIPO (PCT)
Prior art keywords
audio
attribute data
image
image processing
time node
Prior art date
Application number
PCT/CN2019/073084
Other languages
English (en)
French (fr)
Inventor
范旭
李琰
Original Assignee
北京微播视界科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京微播视界科技有限公司 filed Critical 北京微播视界科技有限公司
Priority to GB2105456.4A priority Critical patent/GB2593059B/en
Priority to US17/286,193 priority patent/US20210350545A1/en
Priority to JP2021521029A priority patent/JP7199527B2/ja
Publication of WO2020077914A1 publication Critical patent/WO2020077914A1/zh

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/149Segmentation; Edge detection involving deformable models, e.g. active contour models
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/686Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title or artist information, time, location or usage information, user ratings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/564Depth or shape recovery from multiple images from contours
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/361Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems
    • G10H1/368Recording/reproducing of accompaniment for use with an external source, e.g. karaoke systems displaying animated or moving pictures synchronized with the music or audio part
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • G10H1/40Rhythm
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/80Camera processing pipelines; Components thereof
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/57Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for processing of video signals

Definitions

  • the present disclosure relates to the field of image processing, and in particular, to an image processing method, device, and hardware device.
  • APP Application, referred to as: APP
  • an APP that can realize functions such as dark light detection, beauty camera and super pixels.
  • the beautification function of the smart terminal usually includes beautification processing effects such as skin tone adjustment, dermabrasion, big eyes, and thin face, and can perform the same degree of beautification processing on all the faces recognized in the image.
  • beautification processing effects such as skin tone adjustment, dermabrasion, big eyes, and thin face, and can perform the same degree of beautification processing on all the faces recognized in the image.
  • APPs that can achieve simple special effects.
  • the current special effects function can only pre-set the special effects and synthesize them into the video or image. If you need to modify the special effects, you need to re-create the special effects and then synthesize them into the video or image, making the special effects very inflexible.
  • An image processing method includes: acquiring audio and preprocessing the audio to obtain audio attribute data of each first time node of the audio; acquiring first audio attribute data corresponding to the current time node; The first audio attribute data performs preset processing on the image to be processed.
  • the acquiring audio and preprocessing the audio to obtain the audio attribute data of each first time node of the audio includes: acquiring audio and parsing the audio to obtain each of the audio A mapping table of first time nodes and audio attribute data.
  • the obtaining the first audio attribute data corresponding to the current time node includes: obtaining the current playing time node of the audio; obtaining the first audio attribute data corresponding to the current playing time node according to the mapping table.
  • the preset processing of the image to be processed according to the first audio attribute data includes: presetting the target object in the image to be processed according to the first audio attribute data.
  • the acquiring audio and preprocessing the audio to obtain audio attribute data of each first time node of the audio includes: acquiring audio associated with the image to be processed; parsing the audio, A mapping table between each first time node and audio attribute data in the audio is obtained.
  • the acquiring first audio attribute data corresponding to the current time node includes: acquiring a second time node corresponding to the current image frame of the image to be processed; acquiring the second time according to the mapping table The first audio attribute data corresponding to the node.
  • the preset processing of the image to be processed according to the first audio attribute data includes: presetting the target object in the current image frame according to the first audio attribute data.
  • the method further includes preprocessing the image to be processed, and preprocessing the image to be processed includes segmenting the image to be processed to Obtain the outline of the target object; generate the inner outline feature point of the target object according to the outline of the target object; generate the outer outline feature point according to the inner outline feature point.
  • the preset processing of the image to be processed according to the first audio attribute data includes: setting the attribute of a preset material according to the first audio attribute data; filling the preset material Into the area between the inner contour feature point and the outer contour feature point.
  • the first audio attribute data is audio rhythm intensity information
  • setting the attribute of the preset material according to the first audio attribute data includes: setting the pre-set according to the audio rhythm intensity information One or more of the set width, brightness and display duration of the material.
  • An image processing device including:
  • An audio attribute data obtaining module configured to obtain audio and preprocess the audio to obtain audio attribute data of each first time node of the audio
  • a first audio attribute data obtaining module configured to obtain first audio attribute data corresponding to the current time node
  • the image processing module is configured to perform preset processing on the image to be processed according to the first audio attribute data.
  • the audio attribute data acquisition module is used to acquire audio and parse the audio to obtain a mapping table between each first time node and audio attribute data in the audio.
  • the first audio attribute data acquisition module is used to acquire the current playback time node of the audio; and acquire the first audio attribute data corresponding to the current playback time node according to the mapping table.
  • the image processing module is configured to perform preset processing on the target object in the image to be processed according to the first audio attribute data.
  • the audio attribute data acquisition module is used to acquire audio associated with the image to be processed; analyze the audio to obtain a mapping table between each first time node and audio attribute data in the audio.
  • the first audio attribute data acquisition module is configured to acquire a second time node corresponding to the current image frame of the image to be processed; and obtain a first time node corresponding to the second time node according to the mapping table Audio attribute data.
  • the image processing module is configured to perform preset processing on the target object in the current image frame according to the first audio attribute data.
  • the image processing device further includes a to-be-processed image preprocessing module for segmenting the to-be-processed image to obtain an outline of the target object; generating an inner outline feature point of the target object according to the outline of the target object; The inner contour feature points generate outer contour feature points.
  • the image processing module further includes a material setting module and a filling module, wherein the material setting module is used to set the properties of the preset material according to the first audio attribute data; the filling module is used to The preset material is filled into the area between the inner contour feature point and the outer contour feature point.
  • the first audio attribute data is audio rhythm intensity information
  • the material setting module is configured to set one or more of the preset material width, brightness, and display duration according to the audio rhythm intensity information Pcs.
  • An electronic device includes: a memory for storing non-transitory computer-readable instructions; and a processor for running the computer-readable instructions so that the processor executes any of the above image processing methods when executed A step of.
  • a computer-readable storage medium is used to store non-transitory computer-readable instructions.
  • the non-transitory computer-readable instructions are executed by a computer, the computer is caused to perform the steps described in any of the above methods.
  • the present disclosure discloses an image processing method, device, and hardware device.
  • the image processing method includes: acquiring audio and preprocessing the audio to obtain audio attribute data of each first time node of the audio; acquiring first audio attribute data corresponding to the current time node; The first audio attribute data performs preset processing on the image to be processed.
  • the image processing method of the embodiment of the present disclosure performs preset processing on the image to be processed according to the audio attribute data.
  • the image processing can be completed only by setting the relationship between the audio attribute data and the image processing operation, which improves the image processing. Flexibility and efficiency.
  • FIG. 1 is a schematic flowchart of an image processing method according to an embodiment of the present disclosure
  • FIGS. 2a-2d are schematic diagrams of an example of image processing effects according to the present disclosure.
  • FIG. 3 is a schematic structural diagram of an image processing apparatus according to an embodiment of the present disclosure.
  • FIG. 4 is a schematic structural diagram of an electronic device according to an embodiment of the present disclosure.
  • An embodiment of the present disclosure provides an image processing method.
  • the image processing method provided in this embodiment may be executed by a computing device, which may be implemented as software or a combination of software and hardware.
  • the computing device may be integrated in a server, a terminal device, or the like.
  • the image processing method mainly includes the following steps S101 to S103. among them:
  • Step S101 Obtain audio and preprocess the audio to obtain audio attribute data of each first time node of the audio;
  • the audio and the image to be processed are independent of each other, and the image may be a dynamic image such as a video or a moving picture, or a static image such as a picture.
  • the audio to be used is obtained, and the audio is pre-processed.
  • the pre-processing may be to analyze the audio by time node, and the time node may be set according to the sampling frequency.
  • the length between time nodes can be set to 10ms, that is to say, a time node is sampled every 10ms, and the audio attribute data on the time node is parsed.
  • the audio attribute data can be any attribute of audio. Intensity, pitch, timbre, duration, rhythm, etc.
  • taking the audio attribute data as the strength of the rhythm for example, you can set the strength of the rhythm to 5 levels.
  • the rhythm intensity level on the time node can be obtained, and after the audio is parsed, the correspondence between the time node and its corresponding rhythm intensity level can be obtained.
  • a mapping table can be used to save the correspondence.
  • the audio is associated with the image to be processed
  • the image may be a dynamic image such as a video or a moving picture, or a static image such as a picture
  • the correlation may typically be
  • the audio has a corresponding relationship with the image to be processed, for example, when the video is played or the picture is opened, the audio is played with it.
  • the audio is typically the audio in the video or the audio built into the picture.
  • the audio to be used is obtained, and the audio is pre-processed.
  • the pre-processing may be to analyze the audio by time node, and the time node may be set according to the sampling frequency.
  • the sampling rate can be related to the attributes of the image.
  • the length between the time nodes can be set to 33ms, which means that a time node is sampled every 33ms.
  • the node will correspond to the video frame, or the image is a moving picture, then there are also multiple frames of pictures in the moving picture, and the appearance time of each frame of the picture is used as the time node. If the image is a static picture, you can set the time node arbitrarily. When the image is a video or a moving picture, the time node can also be set arbitrarily, which will not be repeated here. Analyze the audio attribute data on the time node.
  • the audio attribute data may be any attribute of audio, typically such as tone intensity, pitch, timbre, tone length, rhythm, etc.
  • the audio attribute data is used as the rhythm
  • a mapping table may be used to save the corresponding relationship.
  • the time node includes a time node ID and audio attribute data corresponding to the ID.
  • the time node ID may typically be the serial number of the time node in the time node sequence.
  • Step S102 Obtain first audio attribute data corresponding to the current time node
  • the audio and the image to be processed are independent of each other.
  • the time node to which the audio is currently played is obtained. If the current playing time is not on the time node, it can be rounded up or down Determine the current time node.
  • the mapping table generated in step S102 is searched according to the current time node, and the audio attribute data corresponding to the current time node is acquired and recorded as the first audio attribute data.
  • the serial number of the time node may be used to search the audio attribute data corresponding to the time node with the same serial number in the mapping table, and use the audio attribute data as the first audio attribute data.
  • the audio is associated with the to-be-processed image, for example, the audio is audio in a video, and then according to the sampling method in step S101, the frequency of appearance of the video frame may be used to sample the audio Time node.
  • the serial number of the video frame and the audio time node are one-to-one correspondence.
  • the current time node can be directly obtained through the video frame. It is a second time node; use the sequence number of the second time node to search for the audio attribute data corresponding to the time node with the same sequence number in the mapping table, and use the audio attribute data as the first audio attribute data.
  • Step S103 Perform preset processing on the image to be processed according to the first audio attribute data.
  • the target object in the image to be processed is preset processed according to the first audio attribute data.
  • the image to be processed may include multiple objects, the object to be processed may be preset as the target object, the relationship between the audio attribute data and the preset processing may be preset, and the relationship may be set through a mapping relationship, typically , You can directly set the direct correspondence between different audio attribute data and preset processing, or you can indirectly set the relationship between audio attribute data and preset processing. Typically, you can use audio attribute data as a parameter to calculate the Set a first parameter required for processing, and use the first parameter to set the preset processing.
  • the image may be a static picture. At this time, as the audio plays, different processing effects may be presented on static objects on the static picture.
  • the target object in the current image frame is preset according to the first audio attribute data.
  • the current image frame may include multiple objects, and the image frame may be processed as in the processing method in the above embodiment.
  • the image to be processed may be a video.
  • the shape of the target object in each video frame may change, and the processing of each video frame will also follow the audio The time node of changes, showing the effect of the special effects on the target object in the video changing with the audio.
  • step S103 before step S103, it further includes a step of preprocessing the image to be processed.
  • the preprocessing of the image to be processed includes:
  • An outer contour feature point is generated according to the inner contour feature point.
  • the image to be processed may be any image.
  • the image to be processed is a picture.
  • the target image includes a target object.
  • the target object may be any object.
  • the target object is human body.
  • the target image is segmented, and the objects in the image are segmented and separated from other objects to obtain the outline of the target object.
  • when the target image is a video it is necessary to obtain the video first and segment the video frame image in the video; separate the target object in the video frame image from other objects; obtain the target object's profile.
  • an inner outline feature point of the target object is generated, and the inner outline feature point may be located directly on the outline line of the target object, or maintain a predetermined distance from the outline line, such as the inner outline feature point A distance of 0.1 cm can be maintained from the outline.
  • the distance between the inner contour feature points is the same, that is to say, the inner contour feature points are evenly distributed with respect to the contour of the target object.
  • the outer contour feature points are generated according to the inner contour feature points generated in the previous step.
  • the outer contour features are interpolated in the direction of the inner contour feature points away from the target object according to the inner contour feature points Point, taking the target object as a human body as an example, the inner contour feature points are located on the contour line of the human body, and for each inner contour feature point, the corresponding outer contour feature point is interpolated on the outside of the human body.
  • the preset processing of the image to be processed according to the first audio attribute data includes: setting the attribute of the preset material according to the first audio attribute data; setting the preset Material is filled into the area between the inner contour feature point and the outer contour feature point.
  • the first audio attribute data is audio rhythm intensity information
  • setting the attribute of the preset material according to the first audio attribute data includes: setting according to the audio rhythm intensity information One or more of the preset material width, brightness and display duration.
  • the rhythm intensity of the audio includes 5 levels.
  • the width, brightness, and display duration of the corresponding material can be set. After the rhythm intensity level of the audio is obtained, the width and brightness of the material And the display duration is also set, and then the set material is used to fill the area between the inner contour feature point and the outer contour feature point to form a processing effect of using the material to stroke the target object.
  • the materials may be color cards, pictures, etc., and the disclosure does not specifically limit the types of materials.
  • the preset processing described in the present disclosure may be any processing, which is not limited to the above-mentioned processing manner, and the above-mentioned processing manner is just an example for easy understanding.
  • FIGS. 2a-2d it is an example of the effect of the image processing method disclosed in the present disclosure on processing an image.
  • the target object is a human body
  • the human body is stroked to highlight the human body in the image.
  • Position, as shown in Figures 2a-2d, the human body in the image is doing dance movements, and there are music accompaniments.
  • 2a-2d respectively show the different stroke effects presented by the four rhythm levels. 2c, 2d, 2a, 2b.
  • the present disclosure discloses an image processing method, device, and hardware device.
  • the image processing method includes: acquiring audio and preprocessing the audio to obtain audio attribute data of each first time node of the audio; acquiring first audio attribute data corresponding to the current time node; The first audio attribute data performs preset processing on the image to be processed.
  • the image processing method of the embodiment of the present disclosure performs preset processing on the image to be processed according to the audio attribute data.
  • the image processing can be completed only by setting the relationship between the audio attribute data and the image processing operation, which improves the image processing. Flexibility and efficiency.
  • the following is a device embodiment of the present disclosure.
  • the device embodiment of the present disclosure can be used to perform the steps implemented by the method embodiments of the present disclosure.
  • Only parts related to the embodiment of the present disclosure are shown. Specific technical details are not disclosed. Please Refer to the method embodiment of the present disclosure.
  • An embodiment of the present disclosure provides an image processing device.
  • the device may perform the steps described in the above embodiments of the image processing method.
  • the device 300 mainly includes an audio attribute data acquisition module 301, a first audio attribute data acquisition module 302, and an image processing module module 303. among them,
  • the audio attribute data obtaining module 301 is used to obtain audio and preprocess the audio to obtain audio attribute data of each first time node of the audio;
  • the first audio attribute data obtaining module 302 is used to obtain first audio attribute data corresponding to the current time node
  • the image processing module 303 is configured to perform preset processing on the image to be processed according to the first audio attribute data.
  • the audio attribute data obtaining module 301 is configured to obtain audio and parse the audio to obtain a mapping table between each first time node and audio attribute data in the audio.
  • the first audio attribute data obtaining module 302 is configured to obtain a current playing time node of the audio; obtain first audio attribute data corresponding to the current playing time node according to the mapping table.
  • the image processing module 303 is configured to perform preset processing on the target object in the image to be processed according to the first audio attribute data.
  • the audio attribute data obtaining module 301 is configured to obtain audio associated with the image to be processed; analyze the audio to obtain a mapping table between each first time node and audio attribute data in the audio.
  • the first audio attribute data obtaining module 302 is configured to obtain a second time node corresponding to the current image frame of the image to be processed; and obtain a second time node corresponding to the second time node according to the mapping table An audio attribute data.
  • the image processing module 303 is configured to perform preset processing on the target object in the current image frame according to the first audio attribute data.
  • the image processing device further includes a to-be-processed image preprocessing module for segmenting the to-be-processed image to obtain an outline of the target object; generating an inner outline feature point of the target object according to the outline of the target object; The inner contour feature points generate outer contour feature points.
  • the image processing module 303 further includes a material setting module and a filling module, wherein the material setting module is used to set the attributes of the preset material according to the first audio attribute data; the filling module is used to Filling the preset material into the area between the inner contour feature point and the outer contour feature point.
  • the first audio attribute data is audio rhythm intensity information
  • the material setting module is configured to set one or more of the preset material width, brightness, and display duration according to the audio rhythm intensity information Pcs.
  • the device shown in FIG. 3 can execute the method of the embodiment shown in FIG. 1.
  • the device shown in FIG. 3 can execute the method of the embodiment shown in FIG. 1.
  • the relevant description of the embodiment shown in FIG. For the parts not described in detail in this embodiment, reference may be made to the relevant description of the embodiment shown in FIG.
  • the execution process and technical effect of the technical solution refer to the description in the embodiment shown in FIG. 1, and details are not described herein again.
  • FIG. 4 shows a schematic structural diagram of an electronic device 400 suitable for implementing embodiments of the present disclosure.
  • Electronic devices in the embodiments of the present disclosure may include, but are not limited to, such as mobile phones, notebook computers, digital broadcast receivers, PDAs (personal digital assistants), PADs (tablet computers), PMPs (portable multimedia players), in-vehicle terminals (e.g. Mobile terminals such as car navigation terminals) and fixed terminals such as digital TVs, desktop computers and the like.
  • PDAs personal digital assistants
  • PADs tablet computers
  • PMPs portable multimedia players
  • in-vehicle terminals e.g. Mobile terminals such as car navigation terminals
  • fixed terminals such as digital TVs, desktop computers and the like.
  • the electronic device shown in FIG. 4 is only an example, and should not bring any limitation to the functions and use scope of the embodiments of the present disclosure.
  • the electronic device 400 may include a processing device (such as a central processing unit, a graphics processor, etc.) 401, which may be loaded into random access according to a program stored in a read-only memory (ROM) 402 or from the storage device 408
  • the program in the memory (RAM) 403 performs various appropriate operations and processes.
  • various programs and data necessary for the operation of the electronic device 400 are also stored.
  • the processing device 401, ROM 402, and RAM 403 are connected to each other via a bus 404.
  • An input / output (I / O) interface 405 is also connected to the bus 404.
  • the following devices can be connected to the I / O interface 405: including input devices 406 such as touch screen, touch pad, keyboard, mouse, image sensor, microphone, accelerometer, gyroscope, etc .; including, for example, liquid crystal display (LCD), speaker, An output device 407 such as a vibrator; a storage device 408 including, for example, a magnetic tape, a hard disk, etc .; and a communication device 409.
  • the communication device 409 may allow the electronic device 400 to perform wireless or wired communication with other devices to exchange data.
  • FIG. 4 shows an electronic device 400 having various devices, it should be understood that it is not required to implement or have all the devices shown. More or fewer devices may be implemented or provided instead.
  • the process described above with reference to the flowchart may be implemented as a computer software program.
  • embodiments of the present disclosure include a computer program product that includes a computer program carried on a computer-readable medium, the computer program containing program code for performing the method shown in the flowchart.
  • the computer program may be downloaded and installed from the network through the communication device 409, or from the storage device 408, or from the ROM 402.
  • the processing device 401 When the computer program is executed by the processing device 401, the above-mentioned functions defined in the method of the embodiments of the present disclosure are executed.
  • the above-mentioned computer-readable medium in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium or any combination of the two.
  • the computer-readable storage medium may be, for example, but not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, device, or device, or any combination of the above. More specific examples of computer-readable storage media may include, but are not limited to: electrical connections with one or more wires, portable computer diskettes, hard drives, random access memory (RAM), read-only memory (ROM), erasable Programmable read-only memory (EPROM or flash memory), optical fiber, portable compact disk read-only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the foregoing.
  • the computer-readable storage medium may be any tangible medium that contains or stores a program, and the program may be used by or in combination with an instruction execution system, apparatus, or device.
  • the computer-readable signal medium may include a data signal that is propagated in baseband or as part of a carrier wave, in which computer-readable program code is carried. This propagated data signal can take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the above.
  • the computer-readable signal medium may also be any computer-readable medium other than a computer-readable storage medium, and the computer-readable signal medium may send, propagate, or transmit a program for use by or in combination with an instruction execution system, apparatus, or device .
  • the program code contained on the computer readable medium may be transmitted using any appropriate medium, including but not limited to: electric wires, optical cables, RF (radio frequency), etc., or any suitable combination of the foregoing.
  • the computer-readable medium may be included in the electronic device; or it may exist alone without being assembled into the electronic device.
  • the computer-readable medium carries one or more programs, and when the one or more programs are executed by the electronic device, the electronic device is caused to: acquire audio and preprocess the audio to obtain each of the audio Audio attribute data of the first time node; acquiring first audio attribute data corresponding to the current time node; and performing preset processing on the image to be processed according to the first audio attribute data.
  • the computer program code for performing the operations of the present disclosure can be written in one or more programming languages or a combination thereof.
  • the above programming languages include object-oriented programming languages such as Java, Smalltalk, C ++, and also include conventional Procedural programming language-such as "C" language or similar programming language.
  • the program code may execute entirely on the user's computer, partly on the user's computer, as an independent software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server.
  • the remote computer can be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or it can be connected to an external computer (for example, through an Internet service provider Internet connection).
  • LAN local area network
  • WAN wide area network
  • Internet service provider Internet connection for example, AT&T, MCI, Sprint, EarthLink, MSN, GTE, etc.
  • each block in the flowchart or block diagram may represent a module, program segment, or part of code that contains one or more logic functions Executable instructions.
  • the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession can actually be executed in parallel, and sometimes they can also be executed in reverse order, depending on the functions involved.
  • each block in the block diagrams and / or flowcharts, and combinations of blocks in the block diagrams and / or flowcharts can be implemented with dedicated hardware-based systems that perform specified functions or operations Or, it can be realized by a combination of dedicated hardware and computer instructions.
  • the units described in the embodiments of the present disclosure may be implemented in software or hardware.
  • the name of the unit does not constitute a limitation on the unit itself.
  • the first obtaining unit may also be described as "a unit that obtains at least two Internet protocol addresses.”

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Signal Processing (AREA)
  • Acoustics & Sound (AREA)
  • Library & Information Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Processing Or Creating Images (AREA)
  • Television Signal Processing For Recording (AREA)
  • Management Or Editing Of Information On Record Carriers (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

本公开公开一种图像处理方法、装置、硬件装置。其中,该图像处理方法包括:获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据;获取与当前时间节点对应的第一音频属性数据;根据所述第一音频属性数据对待处理图像进行预设的处理。本公开实施例的图像处理方法,根据音频的属性数据来对待处理图像进行预设的处理,只需通过设置音频属性数据和图像处理操作之间的关系即可完成图像处理,提高了图像处理的灵活性和效率。

Description

图像处理方法、装置、硬件装置
交叉引用
本公开引用于2018年10月19日递交的名称为“图像处理方法、装置、硬件装置”的、申请号为201811222642.0的中国专利申请,其通过引用被全部并入本申请。
技术领域
本公开涉及图像处理领域,特别是涉及一种图像处理方法、装置、硬件装置。
背景技术
随着计算机技术的发展,智能终端的应用范围得到了广泛的提高,例如可以通过智能终端听音乐、玩游戏、上网聊天和拍照等。对于智能终端的拍照技术来说,其拍照像素已经达到千万像素以上,具有较高的清晰度和媲美专业相机的拍照效果。
目前在采用智能终端进行拍照时,不仅可以使用出厂时内置的拍照软件实现传统功能的拍照效果,还可以通过从网络端下载应用程序(Application,简称为:APP)来实现具有附加功能的拍照效果,例如可以实现暗光检测、美颜相机和超级像素等功能的APP。智能终端的美颜功能通常包括肤色调整、磨皮、大眼和瘦脸等美颜处理效果,能对图像中已识别出的所有人脸进行相同程度的美颜处理。目前也有APP可以实现简单的特效。
然而目前的特效功能,只能预先设置好特效的效果,并合成到视频或者图像中,如果需要修改特效,则需要重新制作特效后再合成到视频或者图像中,使得特效的生成很不灵活。
发明内容
根据本公开的一个方面,提供以下技术方案:
一种图像处理方法,包括:获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据;获取与当前时间节点对应的第一音频属性数据;根据所述第一音频属性数据对待处理图像进行预设的处理。
进一步的,所述获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据,包括:获取音频并对所述音频进行解析,得到所述音频中每个第一时间节点与音频属性数据的映射表。
进一步的,所述获取与当前时间节点对应的第一音频属性数据,包括:获取所述音频当前的播放时间节点;根据所述映射表获取与当前的播放时间节点对应的第一音频属性数据。
进一步的,所述根据所述第一音频属性数据对待处理图像进行预设的处理,包括:根据所述第一音频属性数据对待处理图像中的目标对象进行预设处理。
进一步的,所述获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据,包括:获取与待处理图像关联的音频;对所述音频进行解析,得到所述音频中每个第一时间节点与音频属性数据的映射表。
进一步的,所述获取与当前时间节点对应的第一音频属性数据,包括:获取所述待处理图像的当前图像帧所对应的第二时间节点;根据所述映射表获取与所述第二时间节点对应的第一音频属性数据。
进一步的,所述根据所述第一音频属性数据对待处理图像进行预设的处理,包括:根据所述第一音频属性数据对所述当前图像帧中的目标对象进行预设的处理。
进一步的,其中在根据所述第一音频属性数据对待处理图像进行预设的处理之前,所述方法还包括对待处理图像进行预处理,该对待处理图像进行预处理包括:对待处理图像进行分割以得到目标对象的轮廓;根据所述目标对象的轮廓生成所述目标对象的内轮廓特征点;根据所述内轮廓特征点生成外轮廓特征点。
进一步的,所述根据所述第一音频属性数据对待处理图像进行预设的处理,包括:根据所述第一音频属性数据对预设的素材的属性进行设置;将所述预设的素材填充到所述内轮廓特征点和所述外轮廓特征点之间的区域 中。
进一步的,所述第一音频属性数据为音频节奏强度信息,所述根据所述第一音频属性数据对预设的素材的属性进行设置,包括:根据所述音频节奏强度信息,设置所述预设的素材的宽度、亮度和显示时长中的一个或多个。
根据本公开的另一个方面,还提供以下技术方案:
一种图像处理装置,包括:
音频属性数据获取模块,用于获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据;
第一音频属性数据获取模块,用于获取与当前时间节点对应的第一音频属性数据;
图像处理模块,用于根据所述第一音频属性数据对待处理图像进行预设的处理。
进一步的,所述音频属性数据获取模块,用于获取音频并对所述音频进行解析,得到所述音频中每个第一时间节点与音频属性数据的映射表。
进一步的,所述第一音频属性数据获取模块,用于获取所述音频当前的播放时间节点;根据所述映射表获取与当前的播放时间节点对应的第一音频属性数据。
进一步的,所述图像处理模块,用于根据所述第一音频属性数据对待处理图像中的目标对象进行预设处理。
进一步的,所述音频属性数据获取模块,用于获取与待处理图像关联的音频;对所述音频进行解析,得到所述音频中每个第一时间节点与音频属性数据的映射表。
进一步的,所述第一音频属性数据获取模块,用于获取所述待处理图像的当前图像帧所对应的第二时间节点;根据所述映射表获取与所述第二时间节点对应的第一音频属性数据。
进一步的,所述图像处理模块,用于根据所述第一音频属性数据对所述当前图像帧中的目标对象进行预设的处理。
进一步的,所述图像处理装置还包括待处理图像预处理模块,用于对待处理图像进行分割以得到目标对象的轮廓;根据所述目标对象的轮廓生成所述目标对象的内轮廓特征点;根据所述内轮廓特征点生成外轮廓特征点。
进一步的,所述图像处理模块还包括素材设置模块和填充模块,其中所 述素材设置模块用于根据所述第一音频属性数据对预设的素材的属性进行设置;所述填充模块用于将所述预设的素材填充到所述内轮廓特征点和所述外轮廓特征点之间的区域中。
进一步的,所述第一音频属性数据为音频节奏强度信息,所述素材设置模块用于根据所述音频节奏强度信息,设置所述预设的素材的宽度、亮度和显示时长中的一个或多个。
根据本公开的又一个方面,还提供以下技术方案:
一种电子设备,包括:存储器,用于存储非暂时性计算机可读指令;以及处理器,用于运行所述计算机可读指令,使得所述处理器执行时实现上述任一图像处理方法所述的步骤。
根据本公开的又一个方面,还提供以下技术方案:
一种计算机可读存储介质,用于存储非暂时性计算机可读指令,当所述非暂时性计算机可读指令由计算机执行时,使得所述计算机执行上述任一方法中所述的步骤。
本公开公开一种图像处理方法、装置、硬件装置。其中,该图像处理方法包括:获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据;获取与当前时间节点对应的第一音频属性数据;根据所述第一音频属性数据对待处理图像进行预设的处理。本公开实施例的图像处理方法,根据音频的属性数据来对待处理图像进行预设的处理,只需通过设置音频属性数据和图像处理操作之间的关系即可完成图像处理,提高了图像处理的灵活性和效率。
上述说明仅是本公开技术方案的概述,为了能更清楚了解本公开的技术手段,而可依照说明书的内容予以实施,并且为让本公开的上述和其他目的、特征和优点能够更明显易懂,以下特举较佳实施例,并配合附图,详细说明如下。
附图说明
图1为根据本公开一个实施例的图像处理方法的流程示意图;
图2a-2d为根据本公开的图像处理效果的一个实例的示意图;
图3为根据本公开一个实施例的图像处理装置的结构示意图;
图4为根据本公开实施例提供的电子设备的结构示意图。
具体实施方式
以下通过特定的具体实例说明本公开的实施方式,本领域技术人员可由本说明书所揭露的内容轻易地了解本公开的其他优点与功效。显然,所描述的实施例仅仅是本公开一部分实施例,而不是全部的实施例。本公开还可以通过另外不同的具体实施方式加以实施或应用,本说明书中的各项细节也可以基于不同观点与应用,在没有背离本公开的精神下进行各种修饰或改变。需说明的是,在不冲突的情况下,以下实施例及实施例中的特征可以相互组合。基于本公开中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本公开保护的范围。
需要说明的是,下文描述在所附权利要求书的范围内的实施例的各种方面。应显而易见,本文中所描述的方面可体现于广泛多种形式中,且本文中所描述的任何特定结构及/或功能仅为说明性的。基于本公开,所属领域的技术人员应了解,本文中所描述的一个方面可与任何其它方面独立地实施,且可以各种方式组合这些方面中的两者或两者以上。举例来说,可使用本文中所阐述的任何数目个方面来实施设备及/或实践方法。另外,可使用除了本文中所阐述的方面中的一或多者之外的其它结构及/或功能性实施此设备及/或实践此方法。
还需要说明的是,以下实施例中所提供的图示仅以示意方式说明本公开的基本构想,图式中仅显示与本公开中有关的组件而非按照实际实施时的组件数目、形状及尺寸绘制,其实际实施时各组件的型态、数量及比例可为一种随意的改变,且其组件布局型态也可能更为复杂。
另外,在以下描述中,提供具体细节是为了便于透彻理解实例。然而,所属领域的技术人员将理解,可在没有这些特定细节的情况下实践所述方面。
本公开实施例提供一种图像处理方法。本实施例提供的该图像处理方法可以由一计算装置来执行,该计算装置可以实现为软件,或者实现为软件和硬件的组合,该计算装置可以集成设置在服务器、终端设备等中。如图1所示,该图像处理方法主要包括如下步骤S101至步骤S103。其中:
步骤S101:获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据;
在一个实施例中,所述音频与待处理的图像相互独立,所述的图像可以 是动态的图像如视频或动图,也可以是静态的图像如图片。在该实施例中,获取所要使用的音频,并对所述音频进行预处理,所述预处理可以是对音频进行逐时间节点的解析,所述时间节点可以根据采样频率设置,在该实施例中,可以设置时间节点之间的长度为10ms,也就是说每隔10ms采样一个时间节点,解析该时间节点上的音频属性数据,所述的音频属性数据可以是音频的任何属性,典型的如音强、音调、音色、音长、节奏等,在该实施例中以音频属性数据为节奏的强度为例,可以设置节奏的强度有5个等级,对于每个时间节点,经过对音频的解析,可以得到该时间节点上的节奏强度等级,将音频解析完之后,可以得到时间节点与其对应的节奏强度等级的对应关系,在该实施例中,可以使用映射表来保存该对应关系。
在另一个实施例中,所述音频与待处理图像相关联,所述的图像可以是动态的图像如视频或动图,也可以是静态的图像如图片,所述的相关联典型的可以是所述音频与待处理图像具有对应关系,如当视频播放或者图片打开的时候所述音频与之一起播放,这种情况典型的所述音频为视频中的音频或图片内置的音频。在该实施例中,获取所要使用的音频,并对所述音频进行预处理,所述预处理可以是对音频进行逐时间节点的解析,所述时间节点可以根据采样频率设置,在该实施例中,采样率可以跟图像的属性相关,如图像为视频,视频每秒有30帧视频帧,则可以设置时间节点之间的长度为33ms,也就是说每隔33ms采样一个时间节点,这样时间节点将与视频帧对应,或者图像为动图,则动图中同样有多帧图片,将每帧图片的出现时间作为时间节点,如果图像为静态的图片,则可以任意设置时间节点,当然当图像为视频或动图的时候,同样可以任意设置时间节点,在此不再赘述。解析该时间节点上的音频属性数据,所述的音频属性数据可以是音频的任何属性,典型的如音强、音调、音色、音长、节奏等,在该实施例中以音频属性数据为节奏的强度为例,可以设置节奏的强度有5个等级,对于每个时间节点,经过对音频的解析,可以得到该时间节点上的节奏强度等级,将音频解析完之后,可以得到时间节点与其对应的节奏强度等级的对应关系,在该实施例中,可以使用映射表来保存该对应关系。
在所述映射表表中,所述时间节点包括时间节点ID和与该ID对应的音频属性数据,所述时间节点ID典型的可以是所述时间节点在时间节点序列中的序号。
步骤S102:获取与当前时间节点对应的第一音频属性数据;
在一个实施例中,所述音频与所述待处理图像相互独立,此时获取所述 音频当前所播放到的时间节点,如果当前播放的时间不在时间节点上,可以通过向上或者向下取整的方式确定当前的时间节点。根据当前的时间节点查找步骤S102中所生成的映射表,获取所述当前时间节点所对应的音频属性数据,记为第一音频属性数据。典型的,可以使用所述时间节点的序号,到所述映射表中查找与该序号相同的时间节点所对应的音频属性数据,将该音频属性数据作为第一音频属性数据。
在另一个实施例中,所述音频与所述待处理图像相关联,如所述音频为视频中的音频,此时根据步骤S101中的采样方法,可以使用视频帧出现的频率来采样音频的时间节点,此时视频帧和音频的时间节点是有对应关系的,因此视频帧的序号与音频的时间节点的序号是一一对应的,此时通过视频帧可以直接获取当前的时间节点,记为第二时间节点;使用所述第二时间节点的序号在映射表中查找与该序号相同的时间节点所对应的音频属性数据,将该音频属性数据作为第一音频属性数据。
步骤S103:根据所述第一音频属性数据对待处理图像进行预设的处理。
在一个实施例中,根据所述第一音频属性数据对待处理图像中的目标对象进行预设处理。所述待处理图像中可以包括多个对象,可以预先设置需要进行处理的对象为目标对象,预先设置音频属性数据与预设处理之间的关系,可以通过映射关系来设置所述关系,典型的,可以直接设置不同的音频属性数据与预设处理之间的直接对应关系,也可以间接设置音频属性数据与预设处理之间的关系,典型的,可以将音频属性数据作为参数计算所述预设处理所需要的第一参数,并使用该第一参数设置所述预设处理。在该实施例中,所述图像可以是静态的图片,此时可以呈现随着音频的播放,在静态图片上的静态对象上呈现出不同的处理效果。
在另一个实施例中,根据所述第一音频属性数据对所述当前图像帧中的目标对象进行预设的处理。所述当前图像帧中可以包括多个对象,可以如上述实施例中的处理方法处理图像帧。在该实施例中,所述待处理图像可以是视频,随着视频的播放,在每一帧视频帧中的目标对象的形态可能发生变化,对每一帧视频帧的处理也会随着音频的时间节点发生变化,呈现出视频中的目标对象上的特效随着音频的变化而变化的效果。
在一个实施例中,在步骤S103之前,还包括对待处理图像进行预处理的步骤,所述对待处理图像进行预处理包括:
对待处理图像进行分割以得到目标对象的轮廓;
根据所述目标对象的轮廓生成所述目标对象的内轮廓特征点;
根据所述内轮廓特征点生成外轮廓特征点。
其中待处理图像可以是任何图像,在一个实施例中,所述待处理图像为图片,其中目标图像中包括目标对象,所述目标对象可以是任何物体,在一个实施例中所述目标对象为人体。对目标图像进行分割,将图像中的对象分割出来,与其他对象分离,得到目标对象的轮廓。在一个实施例中,所述目标图像为视频时,此时需要首先获取视频,对视频中的视频帧图像进行分割;将所述视频帧图像中的目标对象与其他对象分离;得到目标对象的轮廓。
根据上一步骤得到的目标对象的轮廓,生成目标对象的内轮廓特征点,该内轮廓特征点可以直接位于目标对象的轮廓线上,或者与轮廓线保持一预定的距离,如内轮廓特征点可以与所述轮廓线保持0.1cm的距离。在一个实施例中,所述内轮廓特征点之间的距离相同,也就是说所述内轮廓特征点相对于目标对象的轮廓均匀分布。
根据上一步骤中生成的内轮廓特征点,生成外轮廓特征点,在一个实施例中,根据所述内轮廓特征点,在所述内轮廓特征点的远离目标对象的方向上插值出外轮廓特征点,以目标对象为人体为例,内轮廓特征点位于人体轮廓线上,对每个内轮廓特征点,在人体的外侧插值出于之对应的外轮廓特征点。
通过上述三个步骤生成一个由内轮廓特征点和外轮廓特征点组成的区域,该区域位于内轮廓特征点和外轮廓特征点之间。在该实施例中,所述根据所述第一音频属性数据对待处理图像进行预设的处理,包括:据所述第一音频属性数据对预设的素材的属性进行设置;将所述预设的素材填充到所述内轮廓特征点和所述外轮廓特征点之间的区域中。在一个实施例中,所述第一音频属性数据为音频节奏强度信息,所述根据所述第一音频属性数据对预设的素材的属性进行设置,包括:根据所述音频节奏强度信息,设置所述预设的素材的宽度、亮度和显示时长中的一个或多个。在一个实施例中,所述音频的节奏强度包括5个等级,对于每个等级可以设置对应素材的宽度、亮度和显示时长,当获取到音频的节奏强度等级之后,所述素材的宽度、亮度和显示时长也被设置好,之后使用设置好的素材填充所述内轮廓特征点和所述外轮廓特征点之间的区域,形成利用素材对目标对象描边的处理效果。在该实施例中,所述素材可以是色卡、图片等等,本公开不对素材的种类做具体限制。
当然本公开所记载的预设的处理可以是任何处理,不局限于上述处理方式,上述处理方式仅仅是为了方便理解的举例。
如图2a-2d所示,为本公开所公开的图像处理方法对图像进行处理的效果实例,在该实例中,目标对象为人体,对人体进行描边处理,以在图像中突出显示人体的位置,如图2a-2d所示,图像中的人体在做舞蹈动作,并且有音乐伴奏,2a-2d分别展示了4个节奏等级所呈现出的不同的描边效果,节奏的快慢排序分别为2c、2d、2a、2b。
本公开公开一种图像处理方法、装置、硬件装置。其中,该图像处理方法包括:获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据;获取与当前时间节点对应的第一音频属性数据;根据所述第一音频属性数据对待处理图像进行预设的处理。本公开实施例的图像处理方法,根据音频的属性数据来对待处理图像进行预设的处理,只需通过设置音频属性数据和图像处理操作之间的关系即可完成图像处理,提高了图像处理的灵活性和效率。
在上文中,虽然按照上述的顺序描述了上述方法实施例中的各个步骤,本领域技术人员应清楚,本公开实施例中的步骤并不必然按照上述顺序执行,其也可以倒序、并行、交叉等其他顺序执行,而且,在上述步骤的基础上,本领域技术人员也可以再加入其他步骤,这些明显变型或等同替换的方式也应包含在本公开的保护范围之内,在此不再赘述。
下面为本公开装置实施例,本公开装置实施例可用于执行本公开方法实施例实现的步骤,为了便于说明,仅示出了与本公开实施例相关的部分,具体技术细节未揭示的,请参照本公开方法实施例。
本公开实施例提供一种图像处理装置。该装置可以执行上述图像处理方法实施例中所述的步骤。如图3所示,该装置300主要包括:音频属性数据获取模块301、第一音频属性数据获取模块302和图像处理模块模块303。其中,
音频属性数据获取模块301,用于获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据;
第一音频属性数据获取模块302,用于获取与当前时间节点对应的第一音频属性数据;
图像处理模块303,用于根据所述第一音频属性数据对待处理图像进行预设的处理。
进一步的,所述音频属性数据获取模块301,用于获取音频并对所述音频进行解析,得到所述音频中每个第一时间节点与音频属性数据的映射表。
进一步的,所述第一音频属性数据获取模块302,用于获取所述音频当前的播放时间节点;根据所述映射表获取与当前的播放时间节点对应的第一音频属性数据。
进一步的,所述图像处理模块303,用于根据所述第一音频属性数据对待处理图像中的目标对象进行预设处理。
进一步的,所述音频属性数据获取模块301,用于获取与待处理图像关联的音频;对所述音频进行解析,得到所述音频中每个第一时间节点与音频属性数据的映射表。
进一步的,所述第一音频属性数据获取模块302,用于获取所述待处理图像的当前图像帧所对应的第二时间节点;根据所述映射表获取与所述第二时间节点对应的第一音频属性数据。
进一步的,所述图像处理模块303,用于根据所述第一音频属性数据对所述当前图像帧中的目标对象进行预设的处理。
进一步的,所述图像处理装置还包括待处理图像预处理模块,用于对待处理图像进行分割以得到目标对象的轮廓;根据所述目标对象的轮廓生成所述目标对象的内轮廓特征点;根据所述内轮廓特征点生成外轮廓特征点。
进一步的,所述图像处理模块303还包括素材设置模块和填充模块,其中所述素材设置模块用于根据所述第一音频属性数据对预设的素材的属性进行设置;所述填充模块用于将所述预设的素材填充到所述内轮廓特征点和所述外轮廓特征点之间的区域中。
进一步的,所述第一音频属性数据为音频节奏强度信息,所述素材设置模块用于根据所述音频节奏强度信息,设置所述预设的素材的宽度、亮度和显示时长中的一个或多个。
图3所示装置可以执行图1所示实施例的方法,本实施例未详细描述的部分,可参考对图1所示实施例的相关说明。该技术方案的执行过程和技术效果参见图1所示实施例中的描述,在此不再赘述。
下面参考图4,其示出了适于用来实现本公开实施例的电子设备400的结构示意图。本公开实施例中的电子设备可以包括但不限于诸如移动电话、笔记本电脑、数字广播接收器、PDA(个人数字助理)、PAD(平板电脑)、PMP(便携式多媒体播放器)、车载终端(例如车载导航终端)等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。图4示出的电子设备仅仅是一个示例,不应对本公开实施例的功能和使用范围带来任何限制。
如图4所示,电子设备400可以包括处理装置(例如中央处理器、图形处理器等)401,其可以根据存储在只读存储器(ROM)402中的程序或者从存储装置408加载到随机访问存储器(RAM)403中的程序而执行各种适当的动作和处理。在RAM 403中,还存储有电子设备400操作所需的各种程序和数据。处理装置401、ROM 402以及RAM 403通过总线404彼此相连。输入/输出(I/O)接口405也连接至总线404。
通常,以下装置可以连接至I/O接口405:包括例如触摸屏、触摸板、键盘、鼠标、图像传感器、麦克风、加速度计、陀螺仪等的输入装置406;包括例如液晶显示器(LCD)、扬声器、振动器等的输出装置407;包括例如磁带、硬盘等的存储装置408;以及通信装置409。通信装置409可以允许电子设备400与其他设备进行无线或有线通信以交换数据。虽然图4示出了具有各种装置的电子设备400,但是应理解的是,并不要求实施或具备所有示出的装置。可以替代地实施或具备更多或更少的装置。
特别地,根据本公开的实施例,上文参考流程图描述的过程可以被实现为计算机软件程序。例如,本公开的实施例包括一种计算机程序产品,其包括承载在计算机可读介质上的计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。在这样的实施例中,该计算机程序可以通过通信装置409从网络上被下载和安装,或者从存储装置408被安装,或者从ROM 402被安装。在该计算机程序被处理装置401执行时,执行本公开实施例的方法中限定的上述功能。
需要说明的是,本公开上述的计算机可读介质可以是计算机可读信号介质或者计算机可读存储介质或者是上述两者的任意组合。计算机可读存储介质例如可以是——但不限于——电、磁、光、电磁、红外线、或半导体的***、装置或器件,或者任意以上的组合。计算机可读存储介质的更具体的例子可以包括但不限于:具有一个或多个导线的电连接、便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、光纤、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本公开中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行***、装置或者器件使用或者与其结合使用。而在本公开中,计算机可读信号介质可以包括在基带中或者作为载波一部分传播的数据信号,其中承载了计算机可读的程序代码。这种传播的数据信号可以采用多种形式,包括但不限于电磁信号、光信号或上述的任意合适的组合。计算机可读信号介质还可以是计算机可读存储介质以外的任何计算机可读介质,该计算机 可读信号介质可以发送、传播或者传输用于由指令执行***、装置或者器件使用或者与其结合使用的程序。计算机可读介质上包含的程序代码可以用任何适当的介质传输,包括但不限于:电线、光缆、RF(射频)等等,或者上述的任意合适的组合。
上述计算机可读介质可以是上述电子设备中所包含的;也可以是单独存在,而未装配入该电子设备中。
上述计算机可读介质承载有一个或者多个程序,当上述一个或者多个程序被该电子设备执行时,使得该电子设备:获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据;获取与当前时间节点对应的第一音频属性数据;根据所述第一音频属性数据对待处理图像进行预设的处理。
可以以一种或多种程序设计语言或其组合来编写用于执行本公开的操作的计算机程序代码,上述程序设计语言包括面向对象的程序设计语言—诸如Java、Smalltalk、C++,还包括常规的过程式程序设计语言—诸如“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络——包括局域网(LAN)或广域网(WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。
附图中的流程图和框图,图示了按照本公开各种实施例的***、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,该模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的***来实现,或者可以用专用硬件与计算机指令的组合来实现。
描述于本公开实施例中所涉及到的单元可以通过软件的方式实现,也可以通过硬件的方式来实现。其中,单元的名称在某种情况下并不构成对该单元本身的限定,例如,第一获取单元还可以被描述为“获取至少两个网际 协议地址的单元”。
以上描述仅为本公开的较佳实施例以及对所运用技术原理的说明。本领域技术人员应当理解,本公开中所涉及的公开范围,并不限于上述技术特征的特定组合而成的技术方案,同时也应涵盖在不脱离上述公开构思的情况下,由上述技术特征或其等同特征进行任意组合而形成的其它技术方案。例如上述特征与本公开中公开的(但不限于)具有类似功能的技术特征进行互相替换而形成的技术方案。

Claims (13)

  1. 一种图像处理方法,包括:
    获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据;
    获取与当前时间节点对应的第一音频属性数据;
    根据所述第一音频属性数据对待处理图像进行预设的处理。
  2. 如权利要求1所述的图像处理方法,其中所述获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据,包括:
    获取音频并对所述音频进行解析,得到所述音频中每个第一时间节点与音频属性数据的映射关系表。
  3. 如权利要求2所述的图像处理方法,其中所述获取与当前时间节点对应的第一音频属性数据,包括:
    获取所述音频当前的播放时间节点;
    根据所述映射表获取与当前的播放时间节点对应的第一音频属性数据。
  4. 如权利要求3所述的图像处理方法,其中所述根据所述第一音频属性数据对待处理图像进行预设的处理,包括:
    根据所述第一音频属性数据对待处理图像中的目标对象进行预设处理。
  5. 如权利要求1所述的图像处理方法,其中所述获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据,包括:
    获取与待处理图像关联的音频;
    对所述音频进行解析,得到所述音频中每个第一时间节点与音频属性数据的映射表。
  6. 如权利要求5所述的图像处理方法,其中所述获取与当前时间节点对应的第一音频属性数据,包括:
    获取所述待处理图像的当前图像帧所对应的第二时间节点;
    根据所述映射表获取与所述第二时间节点对应的第一音频属性数据。
  7. 如权利要求5所述的图像处理方法,其中所述根据所述第一音频属 性数据对待处理图像进行预设的处理,包括:
    根据所述第一音频属性数据对所述当前图像帧中的目标对象进行预设的处理。
  8. 如权利要求1-7所述的图像处理方法中的任一项,其中在根据所述第一音频属性数据对待处理图像进行预设的处理之前,所述方法还包括对待处理图像进行预处理,该对待处理图像进行预处理包括:
    对待处理图像进行分割以得到目标对象的轮廓;
    根据所述目标对象的轮廓生成所述目标对象的内轮廓特征点;
    根据所述内轮廓特征点生成外轮廓特征点。
  9. 如权利要求8所述的图像处理方法,其中所述根据所述第一音频属性数据对待处理图像进行预设的处理,包括:
    根据所述第一音频属性数据对预设的素材的属性进行设置;
    将所述预设的素材填充到所述内轮廓特征点和所述外轮廓特征点之间的区域中。
  10. 如权利要求9所述的图像处理方法,其中所述第一音频属性数据为音频节奏强度信息,所述根据所述第一音频属性数据对预设的素材的属性进行设置,包括:
    根据所述音频节奏强度信息,设置所述预设的素材的宽度、亮度和显示时长中的一个或多个。
  11. 一种图像处理装置,包括:
    音频属性数据获取模块,用于获取音频并对所述音频进行预处理,得到所述音频的每个第一时间节点的音频属性数据;
    第一音频属性数据获取模块,用于获取与当前时间节点对应的第一音频属性数据;
    图像处理模块,用于根据所述第一音频属性数据对待处理图像进行预设的处理。
  12. 一种电子设备,包括:
    存储器,用于存储非暂时性计算机可读指令;以及
    处理器,用于运行所述计算机可读指令,使得所述处理器执行时实现根据权利要求1-10中任意一项所述的图像处理方法。
  13. 一种计算机可读存储介质,用于存储非暂时性计算机可读指令,当所述非暂时性计算机可读指令由计算机执行时,使得所述计算机执行权利要求1-10中任意一项所述的图像处理方法。
PCT/CN2019/073084 2018-10-19 2019-01-25 图像处理方法、装置、硬件装置 WO2020077914A1 (zh)

Priority Applications (3)

Application Number Priority Date Filing Date Title
GB2105456.4A GB2593059B (en) 2018-10-19 2019-01-25 Image processing method and apparatus, and hardware apparatus
US17/286,193 US20210350545A1 (en) 2018-10-19 2019-01-25 Image processing method and apparatus, and hardware apparatus
JP2021521029A JP7199527B2 (ja) 2018-10-19 2019-01-25 画像処理方法、装置、ハードウェア装置

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201811222642.0 2018-10-19
CN201811222642.0A CN110070896B (zh) 2018-10-19 2018-10-19 图像处理方法、装置、硬件装置

Publications (1)

Publication Number Publication Date
WO2020077914A1 true WO2020077914A1 (zh) 2020-04-23

Family

ID=67365889

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/073084 WO2020077914A1 (zh) 2018-10-19 2019-01-25 图像处理方法、装置、硬件装置

Country Status (5)

Country Link
US (1) US20210350545A1 (zh)
JP (1) JP7199527B2 (zh)
CN (1) CN110070896B (zh)
GB (1) GB2593059B (zh)
WO (1) WO2020077914A1 (zh)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111597476A (zh) * 2020-05-06 2020-08-28 北京金山云网络技术有限公司 图像处理方法及装置
US12040000B2 (en) 2021-02-07 2024-07-16 Lemon Inc. Special effect processing method and apparatus

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111489769B (zh) * 2019-01-25 2022-07-12 北京字节跳动网络技术有限公司 图像处理方法、装置和硬件装置
US11580785B1 (en) 2019-06-10 2023-02-14 Amazon Technologies, Inc. Detecting interactions with non-discretized items and associating interactions with actors using digital images
CN111081285B (zh) * 2019-11-30 2021-11-09 咪咕视讯科技有限公司 一种调整特效的方法、电子设备及存储介质
CN113055738B (zh) * 2019-12-26 2022-07-29 北京字节跳动网络技术有限公司 视频特效处理方法及装置
CN112052358A (zh) * 2020-09-07 2020-12-08 北京字节跳动网络技术有限公司 显示图像的方法、装置、电子设备和计算机可读介质
CN112348926B (zh) * 2020-11-23 2024-06-21 杭州优频科技有限公司 一种基于安卓的视频剪接类app的处理方法和装置
US20220406337A1 (en) * 2021-06-21 2022-12-22 Lemon Inc. Segmentation contour synchronization with beat

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015194509A1 (ja) * 2014-06-20 2015-12-23 株式会社ソニー・コンピュータエンタテインメント 動画像生成装置、動画像生成方法、プログラム、及び情報記憶媒体
CN108124101A (zh) * 2017-12-18 2018-06-05 北京奇虎科技有限公司 视频采集方法、装置、电子设备和计算机可读存储介质
CN108495036A (zh) * 2018-03-29 2018-09-04 维沃移动通信有限公司 一种图像处理方法及移动终端

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002112113A (ja) 2000-09-27 2002-04-12 Nippon Columbia Co Ltd 映像編集装置および記憶媒体
KR20050045826A (ko) * 2003-11-10 2005-05-17 삼성전자주식회사 재생제어정보를 갱신하여 그에 따라 av 데이터를재생하는 장치 및 방법
US8554057B2 (en) * 2004-08-10 2013-10-08 Sony Corporation Information signal processing method and apparatus, and computer program product
CN101458821B (zh) * 2008-12-09 2011-02-02 三星电子(中国)研发中心 一种对图像和视频进行卡通化处理的方法
JP5235644B2 (ja) 2008-12-16 2013-07-10 キヤノン株式会社 撮像装置、画像処理方法、及び撮像装置の制御方法
JP2011250100A (ja) 2010-05-26 2011-12-08 Sony Corp 画像処理装置および方法、並びにプログラム
US9358456B1 (en) * 2010-06-11 2016-06-07 Harmonix Music Systems, Inc. Dance competition game
KR20140114238A (ko) * 2013-03-18 2014-09-26 삼성전자주식회사 오디오와 결합된 이미지 표시 방법
CN105574814A (zh) * 2016-01-06 2016-05-11 华南理工大学 一种肖像剪纸特效的生成方法
US10453494B2 (en) * 2017-01-10 2019-10-22 Adobe Inc. Facilitating synchronization of motion imagery and audio
CN107644423B (zh) * 2017-09-29 2021-06-15 北京奇虎科技有限公司 基于场景分割的视频数据实时处理方法、装置及计算设备
CN108257609A (zh) * 2017-12-05 2018-07-06 北京小唱科技有限公司 音频内容修正的方法及其智能装置
CN108769535B (zh) * 2018-07-04 2021-08-10 腾讯科技(深圳)有限公司 图像处理方法、装置、存储介质和计算机设备

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015194509A1 (ja) * 2014-06-20 2015-12-23 株式会社ソニー・コンピュータエンタテインメント 動画像生成装置、動画像生成方法、プログラム、及び情報記憶媒体
CN108124101A (zh) * 2017-12-18 2018-06-05 北京奇虎科技有限公司 视频采集方法、装置、电子设备和计算机可读存储介质
CN108495036A (zh) * 2018-03-29 2018-09-04 维沃移动通信有限公司 一种图像处理方法及移动终端

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111597476A (zh) * 2020-05-06 2020-08-28 北京金山云网络技术有限公司 图像处理方法及装置
CN111597476B (zh) * 2020-05-06 2023-08-22 北京金山云网络技术有限公司 图像处理方法及装置
US12040000B2 (en) 2021-02-07 2024-07-16 Lemon Inc. Special effect processing method and apparatus

Also Published As

Publication number Publication date
GB202105456D0 (en) 2021-06-02
CN110070896A (zh) 2019-07-30
JP7199527B2 (ja) 2023-01-05
CN110070896B (zh) 2020-09-01
JP2022505118A (ja) 2022-01-14
US20210350545A1 (en) 2021-11-11
GB2593059A (en) 2021-09-15
GB2593059B (en) 2023-08-09

Similar Documents

Publication Publication Date Title
WO2020077914A1 (zh) 图像处理方法、装置、硬件装置
CN110070496B (zh) 图像特效的生成方法、装置和硬件装置
US11037600B2 (en) Video processing method and apparatus, terminal and medium
WO2020151491A1 (zh) 图像形变的控制方法、装置和硬件装置
WO2021139372A1 (zh) 图像的处理方法、装置、可读介质和电子设备
US11776209B2 (en) Image processing method and apparatus, electronic device, and storage medium
US11849211B2 (en) Video processing method, terminal device and storage medium
WO2019227429A1 (zh) 多媒体内容生成方法、装置和设备/终端/服务器
WO2021057740A1 (zh) 视频生成方法、装置、电子设备和计算机可读介质
WO2020077913A1 (zh) 图像处理方法、装置、硬件装置
US12019669B2 (en) Method, apparatus, device, readable storage medium and product for media content processing
CN110070592B (zh) 特效包的生成方法、装置和硬件装置
US20240040069A1 (en) Image special effect configuration method, image recognition method, apparatus and electronic device
CN114598815B (zh) 一种拍摄方法、装置、电子设备和存储介质
WO2020077912A1 (zh) 图像处理方法、装置、硬件装置
CN111352560B (zh) 分屏方法、装置、电子设备和计算机可读存储介质
JP7471510B2 (ja) ピクチャのビデオへの変換の方法、装置、機器および記憶媒体
CN110069641B (zh) 图像处理方法、装置和电子设备
CN112017261B (zh) 贴纸生成方法、装置、电子设备及计算机可读存储介质
CN109636917B (zh) 三维模型的生成方法、装置、硬件装置
US20230237625A1 (en) Video processing method, electronic device, and storage medium
WO2022227996A1 (zh) 图像处理方法、装置、电子设备以及可读存储介质
JP2023550970A (ja) 画面の中の背景を変更する方法、機器、記憶媒体、及びプログラム製品
CN111199519B (zh) 特效包的生成方法和装置
CN114647472B (zh) 图片处理方法、装置、设备、存储介质和程序产品

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19873979

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2021521029

Country of ref document: JP

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 202105456

Country of ref document: GB

Kind code of ref document: A

Free format text: PCT FILING DATE = 20190125

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 06/08/2021)

122 Ep: pct application non-entry in european phase

Ref document number: 19873979

Country of ref document: EP

Kind code of ref document: A1