CN115190356B - Multimedia data processing method and device, electronic equipment and storage medium - Google Patents

Multimedia data processing method and device, electronic equipment and storage medium Download PDF

Info

Publication number
CN115190356B
CN115190356B CN202210654770.2A CN202210654770A CN115190356B CN 115190356 B CN115190356 B CN 115190356B CN 202210654770 A CN202210654770 A CN 202210654770A CN 115190356 B CN115190356 B CN 115190356B
Authority
CN
China
Prior art keywords
slice
multimedia data
target
processed
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202210654770.2A
Other languages
Chinese (zh)
Other versions
CN115190356A (en
Inventor
陈沛康
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Dajia Internet Information Technology Co Ltd
Original Assignee
Beijing Dajia Internet Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Dajia Internet Information Technology Co Ltd filed Critical Beijing Dajia Internet Information Technology Co Ltd
Priority to CN202210654770.2A priority Critical patent/CN115190356B/en
Publication of CN115190356A publication Critical patent/CN115190356A/en
Application granted granted Critical
Publication of CN115190356B publication Critical patent/CN115190356B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/44016Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving splicing one content stream with another content stream, e.g. for substituting a video clip
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44213Monitoring of end-user related data
    • H04N21/44218Detecting physical presence or behaviour of the user, e.g. using sensors to detect if the user is leaving the room or changes his face expression during a TV program
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/488Data services, e.g. news ticker
    • H04N21/4884Data services, e.g. news ticker for displaying subtitles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8549Creating video summaries, e.g. movie trailer

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Social Psychology (AREA)
  • Databases & Information Systems (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computer Security & Cryptography (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The disclosure relates to a multimedia data processing method, a device, an electronic device and a storage medium, comprising: acquiring multimedia data to be processed and operation behavior data corresponding to the multimedia data to be processed; based on preset time length, aggregating the operation behavior data corresponding to the multimedia data to be processed, and determining the aggregation behavior data corresponding to each preset time length; dividing the multimedia data to be processed to obtain a plurality of slices, wherein the duration of the slices is longer than the preset duration; selecting a target slice from the plurality of slices according to the aggregation behavior data and the voice data in the multimedia data to be processed; and generating target multimedia data corresponding to the multimedia data to be processed according to the target slice. Therefore, the finally obtained target multimedia data can effectively capture the highlight content in the multimedia data to be processed, so that the highlight segment positioning logic is optimized, and the user requirement is met.

Description

Multimedia data processing method and device, electronic equipment and storage medium
Technical Field
The disclosure relates to the field of data processing, and in particular relates to a multimedia data processing method, a device, electronic equipment and a storage medium.
Background
The intelligent clipping technology can accurately predict the abstract segments in the multimedia data to be processed. Specifically, some segments in the to-be-processed multimedia data can be extracted, and the extracted segments are combined in series to generate abstract to-be-processed multimedia data, so that the abstract to-be-processed multimedia data can help a user to quickly know the content of the to-be-processed multimedia data.
At present, segments in the multimedia data to be processed are generally extracted based on a keyword recognition algorithm, namely, slicing time points of the multimedia data to be processed are recognized according to a keyword library set based on service characteristics, and then slicing is carried out on the multimedia data to be processed according to the slicing time points, so that the required segments are obtained.
However, the segments extracted according to the keywords are not necessarily highlight segments in the multimedia data to be processed, so that the segments of the multimedia data to be processed extracted in the prior art cannot effectively capture highlight content in the multimedia data to be processed, and it is difficult to meet user requirements.
Disclosure of Invention
The disclosure provides a method, a device, an electronic device and a storage medium for processing multimedia data to be processed, so as to at least solve the problem that a fragment of the multimedia data to be processed extracted in the related technology cannot effectively capture the wonderful content in the multimedia data to be processed, and is difficult to meet the user demand. The technical scheme of the present disclosure is as follows:
According to a first aspect of an embodiment of the present disclosure, there is provided a multimedia data processing method, including:
acquiring multimedia data to be processed and operation behavior data corresponding to the multimedia data to be processed;
based on preset time length, aggregating the operation behavior data corresponding to the multimedia data to be processed, and determining the aggregation behavior data corresponding to each preset time length;
dividing the multimedia data to be processed to obtain a plurality of slices, wherein the duration of the slices is longer than the preset duration;
selecting a target slice from the plurality of slices according to the aggregation behavior data and the voice data in the multimedia data to be processed;
and generating target multimedia data corresponding to the multimedia data to be processed according to the target slice.
Optionally, the aggregate behavior data includes a number of live-broadcast participants, and the selecting a target slice from the plurality of slices according to the aggregate behavior data and the voice data in the to-be-processed multimedia data includes:
according to the aggregation behavior data, selecting a slice corresponding to the maximum number of people in the live broadcasting room from the plurality of slices as a first slice;
and selecting a keyword slice from the plurality of slices according to voice data in the multimedia data to be processed, and taking the keyword slice and the first slice as target slices.
Optionally, the aggregate behavior data further includes a virtual article click number, and the selecting, according to the aggregate behavior data and the voice data in the to-be-processed multimedia data, a target slice from the plurality of slices further includes:
and selecting a slice corresponding to the maximum value of the virtual object click number from the plurality of slices according to the aggregation behavior data, and taking the keyword slice, the first slice and the second slice as target slices.
Optionally, the aggregate behavior data further includes an abnormal value of the interactive data, and the selecting a target slice from the plurality of slices according to the aggregate behavior data and the voice data in the to-be-processed multimedia data further includes:
and selecting a slice corresponding to the maximum abnormal value of the interaction data from the plurality of slices according to the aggregation behavior data, taking the slice as a third slice, and taking the keyword slice, the first slice, the second slice and the third slice as target slices.
Optionally, the generating, according to the target slice, target multimedia data corresponding to the multimedia data to be processed includes:
identifying voice data in the target slice to obtain subtitle information and key word information;
Generating a text special effect according to the subtitle information and the key word information;
and generating target multimedia data corresponding to the multimedia data to be processed based on the text special effect and the target slice.
Optionally, the generating the text special effect according to the subtitle information and the key word information includes:
if the key word information exists in the preset starting time of the target slice, generating a character special effect matched with the key word information;
and if the key word information does not exist in the preset starting time of the target slice, randomly selecting the character special effects from a preset special effect library.
Optionally, the generating the text special effect according to the subtitle information and the key word information includes:
determining the display time, display position and display style corresponding to each key word;
the generating the target multimedia data corresponding to the multimedia data to be processed based on the text special effect and the target slice comprises the following steps:
and displaying the key words in the form of the display style at the display time and the display position of the target slice to obtain the target multimedia data.
According to a second aspect of embodiments of the present disclosure, there is provided a multimedia data processing apparatus comprising:
The device comprises an acquisition unit, a processing unit and a processing unit, wherein the acquisition unit is configured to acquire to-be-processed multimedia data and operation behavior data corresponding to the to-be-processed multimedia data;
the aggregation unit is configured to perform aggregation on the operation behavior data corresponding to the multimedia data to be processed based on preset time lengths, and determine aggregation behavior data corresponding to each preset time length;
the slicing unit is configured to divide the multimedia data to be processed to obtain a plurality of slices, and the duration of the slices is longer than the preset duration;
a screening unit configured to perform screening of a target slice from the plurality of slices according to the aggregate behavior data and the voice data in the multimedia data to be processed;
and the generating unit is configured to execute the generation of target multimedia data corresponding to the multimedia data to be processed according to the target slice.
Optionally, the aggregate behavior data includes a live room number of people, and the screening unit is configured to perform:
according to the aggregation behavior data, selecting a slice corresponding to the maximum number of people in the live broadcasting room from the plurality of slices as a first slice;
and selecting a keyword slice from the plurality of slices according to voice data in the multimedia data to be processed, and taking the keyword slice and the first slice as target slices.
Optionally, the aggregate behavior data further includes a virtual item click number, and the filtering unit is further configured to perform:
and selecting a slice corresponding to the maximum value of the virtual object click number from the plurality of slices according to the aggregation behavior data, and taking the keyword slice, the first slice and the second slice as target slices.
Optionally, the aggregate behavior data further includes an interaction data outlier, and the filtering unit is further configured to perform:
and selecting a slice corresponding to the maximum abnormal value of the interaction data from the plurality of slices according to the aggregation behavior data, taking the slice as a third slice, and taking the keyword slice, the first slice, the second slice and the third slice as target slices.
Optionally, the generating unit is further configured to perform:
identifying voice data in the target slice to obtain subtitle information and key word information;
generating a text special effect according to the subtitle information and the key word information;
and generating target multimedia data corresponding to the multimedia data to be processed based on the text special effect and the target slice.
Optionally, the generating unit is configured to perform:
if the key word information exists in the preset starting time of the target slice, generating a character special effect matched with the key word information;
and if the key word information does not exist in the preset starting time of the target slice, randomly selecting the character special effects from a preset special effect library.
Optionally, the generating unit is configured to perform:
determining the display time, display position and display style corresponding to each key word;
and displaying the key words in the form of the display style at the display time and the display position of the target slice to obtain the target multimedia data.
According to a third aspect of embodiments of the present disclosure, there is provided a multimedia data processing electronic device comprising:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the multimedia data processing method of any of the above.
According to a fourth aspect of embodiments of the present disclosure, there is provided a computer readable storage medium, which when executed by a processor of a multimedia data processing electronic device, enables the multimedia data processing electronic device to perform any one of the above-described multimedia data processing methods.
According to a fifth aspect of embodiments of the present disclosure, there is provided a computer program product comprising a computer program/instruction which, when executed by a processor, implements the method of multimedia data processing of any of the above.
The technical scheme provided by the embodiment of the disclosure at least brings the following beneficial effects:
acquiring multimedia data to be processed and operation behavior data corresponding to the multimedia data to be processed; based on preset time length, aggregating the operation behavior data corresponding to the multimedia data to be processed, and determining the aggregation behavior data corresponding to each preset time length; dividing the multimedia data to be processed to obtain a plurality of slices, wherein the duration of the slices is longer than the preset duration; selecting a target slice from the plurality of slices according to the aggregation behavior data and the voice data in the multimedia data to be processed; and generating target multimedia data corresponding to the multimedia data to be processed according to the target slice.
In this way, the operation behavior data are aggregated, when the slices of the multimedia data to be processed are clipped, the corresponding aggregation behavior data are referred to, and the target slices are screened out from the slices, in other words, the target slices are fragments which are more likely to contain the highlight in the multimedia data to be processed, so that the finally obtained target multimedia data can effectively capture the highlight in the multimedia data to be processed, thereby optimizing the highlight positioning logic and meeting the user requirements.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the disclosure and together with the description, serve to explain the principles of the disclosure and do not constitute an undue limitation on the disclosure.
Fig. 1 is a flowchart illustrating a multimedia data processing method according to an exemplary embodiment.
Fig. 2 is a logic diagram illustrating a multimedia data processing method according to an exemplary embodiment.
Fig. 3 is a schematic diagram illustrating a scheme for generating a text effect according to subtitle information and accent word information according to an exemplary embodiment.
Fig. 4 is a schematic diagram of a scheme for processing multimedia data to be processed to obtain target multimedia data based on a live stream scene according to an exemplary embodiment.
Fig. 5 is a block diagram illustrating a multimedia data processing apparatus according to an exemplary embodiment.
Fig. 6 is a block diagram of an electronic device for multimedia data processing, according to an example embodiment.
Fig. 7 is a block diagram illustrating an apparatus for multimedia data processing according to an exemplary embodiment.
Detailed Description
In order to enable those skilled in the art to better understand the technical solutions of the present disclosure, the technical solutions of the embodiments of the present disclosure will be clearly and completely described below with reference to the accompanying drawings.
It should be noted that the terms "first," "second," and the like in the description and claims of the present disclosure and in the foregoing figures are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used may be interchanged where appropriate such that the embodiments of the disclosure described herein may be capable of operation in sequences other than those illustrated or described herein. The implementations described in the following exemplary examples are not representative of all implementations consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with some aspects of the present disclosure as detailed in the accompanying claims.
Fig. 1 is a flowchart illustrating a multimedia data processing method according to an exemplary embodiment, and the multimedia data processing method to be processed includes the following steps as shown in fig. 1.
In step S11, the multimedia data to be processed and the operation behavior data corresponding to the multimedia data to be processed are obtained.
The multimedia data to be processed may be audio data or video data, and may be any format of multimedia data to be processed, which is not limited in particular. The operation behavior data corresponding to the multimedia data to be processed are used for indicating the operation executed by the user based on the multimedia data to be processed.
For example, the multimedia data to be processed may be live multimedia data to be processed in a live broadcast room, may be a multimedia data stream to be processed acquired in real time, or may be a multimedia data file to be processed acquired after live broadcast is completed, which is not limited in particular. Correspondingly, the operation behavior data may be behavior data executed by a viewer in the live broadcast room, for example, entering or exiting the live broadcast room at a certain time in the live broadcast room, clicking on a virtual object displayed in the live broadcast room, and the like, which is not particularly limited.
In step S12, based on the preset duration, the operation behavior data corresponding to the multimedia data to be processed are aggregated, and the aggregated behavior data corresponding to each preset duration is determined.
In the present disclosure, the preset duration may be set according to a user's requirement, or may also be set according to a duration of multimedia data to be processed, which is not specifically limited. For example, the preset duration may be 1 second, that is, the operation behavior data is aggregated in units of seconds, to obtain the aggregated behavior data.
For example, in a live broadcasting room scene, operation behavior data such as entering a live broadcasting room or exiting the live broadcasting room, clicking virtual objects displayed in the live broadcasting room, and the like, can be aggregated, and if the preset duration is 1 second, the data such as the number of people entering the live broadcasting room, the number of people exiting the live broadcasting room, or the number of people clicking a certain virtual object displayed in the live broadcasting room in each second can be obtained as the aggregated behavior data.
In the case that the multimedia data to be processed is a multimedia data file to be processed, the preset duration may also be determined according to the duration of the multimedia data file to be processed, for example, if the duration of the multimedia data to be processed is longer and is several hours, the preset duration may be set to 1 minute, so that the number of times of data aggregation may be reduced, and the efficiency of processing the multimedia data to be processed may be further improved.
In step S13, the multimedia data to be processed is segmented to obtain a plurality of slices, and the duration of the slices is longer than a preset duration.
In this step, the number of the segments to be performed on the media data to be processed may be preset, and then the segments to be processed may be equally divided based on the preset number, for example, the segments to be performed on the multimedia data to be processed may be preset to be divided into a preset number of slices with the same duration. Or, the duration of each slice may be preset, and then the multimedia data to be processed is segmented to obtain a plurality of slices with preset slice durations.
The time length of the slice is longer than the preset time length, that is, each obtained slice corresponds to a plurality of pieces of aggregation behavior data, in other words, each slice can be further analyzed based on the aggregation behavior data, so that subsequent processing of the multimedia data to be processed is realized.
In step S14, a target slice is selected from the plurality of slices according to the aggregate behavior data and the speech data in the multimedia data to be processed.
It will be appreciated that the multimedia data to be processed includes video data and voice data, the voice data generally includes some important information, and the voice data corresponding to the multimedia data to be processed can be identified by voice recognition, and further, based on the identified information, a target slice can be determined from the slices obtained by segmentation.
For example, in a live broadcast room scenario, the voice data is usually a main broadcast comment, and the main broadcast comment usually includes some key information about the content of the multimedia data, for example, the key information may be introduction information of activities or virtual objects in the live broadcast room, or information guiding the operation of the audience, etc., so that the target slice may be selected from multiple slices according to the aggregated behavior data and the voice data in the multimedia data to be processed.
In one implementation, the aggregate behavior data includes a number of live-room entrances, and selecting a target slice from a plurality of slices according to the aggregate behavior data and voice data in the multimedia data to be processed, includes: according to the aggregation behavior data, selecting a slice corresponding to the maximum number of people in the live broadcasting room from a plurality of slices as a first slice; and selecting a keyword slice from the plurality of slices according to voice data in the multimedia data to be processed, and taking the keyword slice and the first slice as target slices.
The slice corresponding to the maximum number of people in the live broadcasting room means that the slice has strong attraction to the audience, in other words, the slice can be regarded as the slice with the highest heat in the live broadcasting room, so that the slice is taken as the first slice, and the subsequent heat of the obtained target multimedia data is improved.
In one implementation, the aggregate behavior data further includes a virtual object click number, and the method further includes selecting a target slice from the plurality of slices according to the aggregate behavior data and the voice data in the multimedia data to be processed, and further includes: and selecting a slice corresponding to the maximum value of the virtual object click number from the plurality of slices according to the aggregation behavior data, taking the slice as a second slice, and taking the keyword slice, the first slice and the second slice as target slices.
The slice corresponding to the maximum value of the click number of the virtual article means that the possibility that the audience is interested in the virtual article in the current slice is high, and the consumption will is high, so that the virtual article is used as the second slice, and the subsequent gain of the target multimedia data is improved.
In one implementation, the aggregate behavior data further includes an interaction data outlier, and the method further includes selecting a target slice from the plurality of slices according to the aggregate behavior data and the voice data in the multimedia data to be processed, and further includes: and selecting a slice corresponding to the maximum abnormal value of the interaction data from the plurality of slices according to the aggregation behavior data, taking the slice as a third slice, and taking the keyword slice, the first slice, the second slice and the third slice as target slices.
The slice corresponding to the maximum value of the abnormal value of the interactive data means that the current slice has some special contents, which may cause feedback of the audience, and increase the heat of the multimedia data, so that the current slice can be used as a third slice for subsequent processing, thereby increasing the wonderful degree of the target multimedia data.
In step S15, target multimedia data corresponding to the multimedia data to be processed is generated according to the target slice.
It will be appreciated that a plurality of types of slices that may contain critical information are included in the target slice, such that the resulting target multimedia data will be more likely to contain highlights in the multimedia data to be processed. For example, in a live broadcasting room scene, the identification of the highlight is required to consider live broadcasting room information, operation behavior information and the like, and the information extraction in the multimedia data to be processed is more comprehensive, so that highlight positioning logic can be optimized, and the user requirement is met.
Fig. 2 is a schematic logic diagram of the present embodiment in one implementation. Firstly, to-be-processed multimedia data and operation behavior data corresponding to the to-be-processed multimedia data can be obtained, wherein the granularity of obtaining the to-be-processed multimedia data can be specifically obtained from a live stream once every 10 minutes, and the obtained video stream is subjected to end-to-end voice recognition and explanation detection service to identify a voice fragment used for explanation in the live stream as the to-be-processed multimedia data. Then, the operation behavior data corresponding to the multimedia data to be processed can be aggregated based on the preset duration, the aggregation behavior data corresponding to each preset duration, that is, the fragment preliminary aggregation service, is determined, and meta (metadata) information of the fragment is obtained and used as the aggregation behavior data. And then, the material optimizing group is used for dividing the multimedia data to be processed to obtain a plurality of slices, and the duration of each slice is longer than the preset duration, for example, the duration of each slice can be 2-3 minutes. Then, a target slice can be selected from a plurality of slices according to the aggregate behavior data and the voice data in the multimedia data to be processed, for example, an automatic caption service is utilized to identify the voice text in each slice, then the identified caption is subjected to important word identification, and the caption and the important word information are subjected to material optimization group processing. Finally, according to the target slice, generating target multimedia data corresponding to the multimedia data to be processed, for example, audio data of 40-60 seconds. Meanwhile, the subtitle and key word recognition results corresponding to the target multimedia data can be generated, so that corresponding video and key word results can be obtained through processing of the prime manufacturing service group, and the film effect is finally obtained after Y-tech special effects and word special effects are added.
In this step, generating target multimedia data corresponding to the multimedia data to be processed according to the target slice may include: identifying voice data in the target slice to obtain subtitle information and key word information; generating a text special effect according to the subtitle information and the key word information; and generating target multimedia data corresponding to the multimedia data to be processed based on the text special effect and the target slice.
Therefore, the obtained target multimedia data can also comprise some text special effects, and compared with a mode of only comprising the target slice content, the method can better display the user and attract the user to watch.
The process of generating the text special effect according to the subtitle information and the key word information is shown in fig. 3. Firstly, adding a prefix for n seconds before a target slice, taking the prefix for n seconds as a slice header, determining a corresponding slice header special effect according to algorithm recommendation, and selecting the slice header special effect from a statement list in a slice header template; similarly, a suffix n seconds may be added after the target slice, and as a tail, the tail special effect may be selected from the sentence list in the tail template, which is not described herein. Then, in the first 3 seconds of the target slice, if the keywords exist, corresponding keyword special effects are generated, namely, hit keyword special effects are judged, and if the keywords do not exist in the first 3 seconds, 10 character special effects are randomly generated according to the spam strategy. In the remaining time of the target slice, the limitation between the special effect type and the keyword and tone adjustment play (tonettype) can be canceled, and the keyword matching and the special effect matching can be simultaneously carried out, wherein the keyword matching refers to the full text recognition of the caption of the target slice, the keyword and the corresponding time list are recognized, then the secondary fine adjustment is carried out, the keyword replacement is carried out according to a hard replacement rule, the keyword with too low reliability is screened, the keyword sticker special effect with too short interval time length is screened, the keyword sticker special effect with too long interval time is supplemented, and the like; the special effect matching means setting a keyword alternative list, and presetting different patterns for different keywords, for example, special effect patterns of different keywords can be key subtitles or sticker special effects respectively, or special effects of keywords can be self-adaptive according to the length of the keywords and the special effect patterns, can be directly hard matched according to specific stickers, and the like, after the special effect patterns are determined, fine adjustment can be performed according to the patterns and positions of the special effects in time, for example, avoidance rules can be observed, characters and subtitles in a special effect target slice can be avoided, special effect scattering rules can be observed, overlapping among special effects can be avoided, rhythm adjustment rules can be observed, the special effects can be changed in a click or a subtitle, the display effect is better, and the like.
In one implementation, generating a text effect according to subtitle information and key word information includes: if the key word information exists in the preset starting time of the target slice, generating a character special effect with the key word information matched; if the key word information does not exist in the preset starting time of the target slice, randomly selecting the character special effects from a preset special effect library.
It can be appreciated that through recognition of the key words, the finally generated text special effects are more in line with the content of the target multimedia data, so that the experience of the user is better.
In one implementation, generating a text effect according to subtitle information and key word information includes: determining the display time, display position and display style corresponding to each key word; correspondingly, based on the text special effect and the target slice, generating target multimedia data corresponding to the multimedia data to be processed, including: and displaying the key words in a display style form at the display time and the display position of the target slice to obtain the target multimedia data.
That is, for different target slices, the generated text special effects can be displayed in a diversified manner, so that the content of the finally obtained target multimedia data is also richer, thereby displaying the highlight content in the multimedia data to be processed in an all-around manner, and further improving the user experience.
Fig. 4 is a schematic diagram of a scheme for processing multimedia data to be processed based on a live stream scene to obtain target multimedia data. Firstly, to-be-processed multimedia data and operation behavior data corresponding to the to-be-processed multimedia data can be acquired, wherein each to-be-processed multimedia data has corresponding slicing time, the operation behavior data comprises live broadcasting room entrance data and dolly click data, the live broadcasting room entrance data can be aggregated according to seconds, the live broadcasting room entrance data comprises entrance time and entrance number, and the dolly click data can be aggregated according to seconds, the dolly click data comprises click time and click number. Then, based on a material processing algorithm, the operation behavior data corresponding to the multimedia data to be processed can be aggregated based on preset time length, and the aggregation behavior data corresponding to each preset time length is determined, wherein the aggregation behavior data specifically comprises a commodity explanation recognition algorithm, a benefit point recognition algorithm, a background music recognition algorithm and the like. Furthermore, the multimedia data to be processed is segmented based on a slicing algorithm to obtain a plurality of slices, wherein the duration of the slices is longer than a preset duration, for example, the slices can be fragments of commodity explanation, and each slice comprises commodity explanation time, benefit point fragment time, information about whether background music is contained or not, and the like. And screening target slices from the plurality of slices according to the aggregate behavior data and the voice data in the multimedia data to be processed, wherein the rule for selecting the target slices can be determined according to commodity explanation time, the maximum number of live broadcasting room in-process people, the maximum number of clicks of the live broadcasting room in-process small yellow cars and the abnormal value of the interactive data, and specifically, the target slices comprise benefit point fragments, the maximum number of live broadcasting room in-process people, the maximum number of clicks of the live broadcasting room in-process small yellow cars and the abnormal value of the interactive data. Finally, generating target multimedia data corresponding to the multimedia data to be processed according to the target slice, wherein the target multimedia data can comprise a subtitle and a key word recognition result.
From the above, it can be seen that, according to the technical solution provided by the embodiments of the present disclosure, the operation behavior data are aggregated, and when the slice of the multimedia data to be processed is clipped, the target slice is selected from the plurality of slices with reference to the corresponding aggregation behavior data, in other words, the target slice is a segment more likely to include the highlight in the multimedia data to be processed, so that the finally obtained target multimedia data can effectively capture the highlight in the multimedia data to be processed, thereby optimizing the highlight positioning logic and meeting the user requirements.
Fig. 5 is a block diagram of a multimedia data processing apparatus according to an exemplary embodiment, the apparatus comprising:
an obtaining unit 201 configured to perform obtaining to-be-processed multimedia data and operation behavior data corresponding to the to-be-processed multimedia data;
an aggregation unit 202, configured to perform aggregation of operation behavior data corresponding to the multimedia data to be processed based on a preset duration, and determine aggregation behavior data corresponding to each preset duration;
a slicing unit 203, configured to perform slicing on the multimedia data to be processed, so as to obtain a plurality of slices, where a duration of the slices is greater than the preset duration;
A screening unit 204 configured to perform screening of a target slice from the plurality of slices according to the aggregate behavior data and the voice data in the multimedia data to be processed;
and a generating unit 205 configured to generate target multimedia data corresponding to the multimedia data to be processed according to the target slice.
Optionally, the aggregate behavior data includes a live room number of people, and the screening unit is configured to perform:
according to the aggregation behavior data, selecting a slice corresponding to the maximum number of people in the live broadcasting room from the plurality of slices as a first slice;
and selecting a keyword slice from the plurality of slices according to voice data in the multimedia data to be processed, and taking the keyword slice and the first slice as target slices.
Optionally, the aggregate behavior data further includes a virtual item click number, and the filtering unit is further configured to perform:
and selecting a slice corresponding to the maximum value of the virtual object click number from the plurality of slices according to the aggregation behavior data, and taking the keyword slice, the first slice and the second slice as target slices.
Optionally, the aggregate behavior data further includes an interaction data outlier, and the filtering unit is further configured to perform:
and selecting a slice corresponding to the maximum abnormal value of the interaction data from the plurality of slices according to the aggregation behavior data, taking the slice as a third slice, and taking the keyword slice, the first slice, the second slice and the third slice as target slices.
Optionally, the generating unit is further configured to perform:
identifying voice data in the target slice to obtain subtitle information and key word information;
generating a text special effect according to the subtitle information and the key word information;
and generating target multimedia data corresponding to the multimedia data to be processed based on the text special effect and the target slice.
Optionally, the generating unit is configured to perform:
if the key word information exists in the preset starting time of the target slice, generating a character special effect matched with the key word information;
and if the key word information does not exist in the preset starting time of the target slice, randomly selecting the character special effects from a preset special effect library.
Optionally, the generating unit is configured to perform:
Determining the display time, display position and display style corresponding to each key word;
and displaying the key words in the form of the display style at the display time and the display position of the target slice to obtain the target multimedia data.
From the above, it can be seen that, according to the technical solution provided by the embodiments of the present disclosure, the operation behavior data are aggregated, and when the slice of the multimedia data to be processed is clipped, the target slice is selected from the plurality of slices with reference to the corresponding aggregation behavior data, in other words, the target slice is a segment more likely to include the highlight in the multimedia data to be processed, so that the finally obtained target multimedia data can effectively capture the highlight in the multimedia data to be processed, thereby optimizing the highlight positioning logic and meeting the user requirements.
The specific manner in which the various modules perform the operations in the apparatus of the above embodiments have been described in detail in connection with the embodiments of the method, and will not be described in detail herein.
FIG. 6 is a block diagram of an electronic device for multimedia data processing, including a processor and a memory for storing a computer program, according to an exemplary embodiment; the processor is used for executing programs stored on the memory.
The memory may include random access memory (Random Access Memory, RAM) or non-volatile memory (non-volatile memory), such as at least one disk memory. Optionally, the memory may also be at least one memory device located remotely from the aforementioned processor.
The processor may be a general-purpose processor, including a central processing unit (Central Processing Unit, CPU for short), a network processor (Network Processor, NP for short), etc.; but also digital signal processors (Digital Signal Processing, DSP for short), application specific integrated circuits (Application Specific Integrated Circuit, ASIC for short), field-programmable gate arrays (Field-Programmable Gate Array, FPGA for short) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
In an exemplary embodiment, a computer-readable storage medium is also provided, such as a memory, comprising instructions executable by a processor of an electronic device to perform the above-described method. Alternatively, the computer readable storage medium may be ROM, random Access Memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, etc.
In an exemplary embodiment, a computer program product is also provided which, when run on a computer, causes the computer to carry out the above-mentioned method of processing multimedia data to be processed.
From the above, it can be seen that, according to the technical solution provided by the embodiments of the present disclosure, the operation behavior data are aggregated, and when the slice of the multimedia data to be processed is clipped, the target slice is selected from the plurality of slices with reference to the corresponding aggregation behavior data, in other words, the target slice is a segment more likely to include the highlight in the multimedia data to be processed, so that the finally obtained target multimedia data can effectively capture the highlight in the multimedia data to be processed, thereby optimizing the highlight positioning logic and meeting the user requirements.
Fig. 7 is a block diagram illustrating an apparatus 800 for multimedia data processing according to an exemplary embodiment.
For example, apparatus 800 may be a mobile phone, computer, digital broadcast electronic device, messaging device, game console, tablet device, medical device, exercise device, personal digital assistant, or the like.
Referring to fig. 7, apparatus 800 may include one or more of the following components: a processing component 802, a memory 804, a power component 806, a multimedia component 808, an audio component 810, an input/output (I/O) interface 812, a sensor component 814, and a communication component 816.
The processing component 802 generally controls overall operation of the apparatus 800, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations. The processing component 802 may include one or more processors 820 to execute instructions to perform all or part of the steps of the methods described above. Further, the processing component 802 can include one or more modules that facilitate interactions between the processing component 802 and other components. For example, the processing component 802 can include a multimedia module to facilitate interaction between the multimedia component 808 and the processing component 802.
The memory 804 is configured to store various types of data to support operations at the device 800. Examples of such data include instructions for any application or method operating on the device 800, contact data, phonebook data, messages, pictures, videos, and the like. The memory 804 may be implemented by any type or combination of volatile or nonvolatile memory devices such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disk.
Power supply component 807 provides power to the various components of device 800. Power supply component 807 can include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for device 800.
The multimedia component 808 includes a screen between the device 800 and the user that provides an output interface. In some embodiments, the screen may include a Liquid Crystal Display (LCD) and a Touch Panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from a user. The touch panel includes one or more touch sensors to sense touches, swipes, and gestures on the touch panel. The touch sensor may sense not only the boundary of a touch or slide action, but also the duration and pressure associated with the touch or slide operation. In some embodiments, the multimedia component 808 includes a front camera and/or a rear camera. When the device 800 is in an operational mode, such as a shooting mode or a video mode, the front camera and/or the rear camera may receive external multimedia data to be processed. Each front camera and rear camera may be a fixed optical lens system or have focal length and optical zoom capabilities.
The audio component 810 is configured to output and/or input audio signals. For example, the audio component 810 includes a Microphone (MIC) configured to receive external audio signals when the device 800 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode. The received audio signals may be further stored in the memory 804 or transmitted via the communication component 816. In some embodiments, audio component 810 further includes a speaker for outputting audio signals.
The I/O interface 812 provides an interface between the processing component 802 and peripheral interface modules, which may be a keyboard, click wheel, buttons, etc. These buttons may include, but are not limited to: homepage button, volume button, start button, and lock button.
The sensor assembly 814 includes one or more sensors for providing status assessment of various aspects of the apparatus 800. For example, the sensor assembly 814 may detect an on/off state of the device 800, a relative positioning of the components, such as a display and keypad of the apparatus 800, the sensor assembly 814 may also detect a change in position of the apparatus 800 or one component of the apparatus 800, the presence or absence of user contact with the apparatus 800, an orientation or acceleration/deceleration of the apparatus 800, and a change in temperature of the apparatus 800. The sensor assembly 814 may include a proximity sensor configured to detect the presence of nearby objects without any physical contact. The sensor assembly 814 may also include a light sensor, such as a CMOS or CCD image sensor, for use in imaging applications. In some embodiments, the sensor assembly 814 may also include an acceleration sensor, a gyroscopic sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
The communication component 816 is configured to facilitate communication between the apparatus 800 and other devices, either in a wired or wireless manner. The device 800 may access a wireless network based on a communication standard, such as WiFi, an operator network (e.g., 2G, 3G, 4G, or 5G), or a combination thereof. In one exemplary embodiment, the communication component 816 receives broadcast signals or broadcast related information from an external broadcast management system via a broadcast channel. In one exemplary embodiment, the communication component 816 further includes a Near Field Communication (NFC) module to facilitate short range communications. For example, the NFC module may be implemented based on Radio Frequency Identification (RFID) technology, infrared data association (IrDA) technology, ultra Wideband (UWB) technology, bluetooth (BT) technology, and other technologies.
In an exemplary embodiment, the apparatus 800 may be implemented by one or more Application Specific Integrated Circuits (ASICs), digital Signal Processors (DSPs), digital Signal Processing Devices (DSPDs), programmable Logic Devices (PLDs), field Programmable Gate Arrays (FPGAs), controllers, microcontrollers, microprocessors or other electronic elements for executing the methods described in the first and second aspects.
In an exemplary embodiment, a non-transitory computer readable storage medium is also provided, such as memory 804 including instructions executable by processor 820 of apparatus 800 to perform the above-described method. Alternatively, for example, the storage medium may be a non-transitory computer-readable storage medium, which may be, for example, ROM, random Access Memory (RAM), CD-ROM, magnetic tape, floppy disk, optical data storage device, and the like.
In an exemplary embodiment, a computer program product comprising instructions which, when run on a computer, causes the computer to perform the multimedia data processing method of any of the above embodiments is also provided.
From the above, it can be seen that, according to the technical solution provided by the embodiments of the present disclosure, the operation behavior data are aggregated, and when the slice of the multimedia data to be processed is clipped, the target slice is selected from the plurality of slices with reference to the corresponding aggregation behavior data, in other words, the target slice is a segment more likely to include the highlight in the multimedia data to be processed, so that the finally obtained target multimedia data can effectively capture the highlight in the multimedia data to be processed, thereby optimizing the highlight positioning logic and meeting the user requirements.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This application is intended to cover any adaptations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.
It is to be understood that the present disclosure is not limited to the precise arrangements and instrumentalities shown in the drawings, and that various modifications and changes may be effected without departing from the scope thereof. The scope of the present disclosure is limited only by the appended claims.

Claims (14)

1. A method of multimedia data processing, comprising:
acquiring multimedia data to be processed and operation behavior data corresponding to the multimedia data to be processed;
based on preset time length, aggregating the operation behavior data corresponding to the multimedia data to be processed, and determining the aggregation behavior data corresponding to each preset time length;
dividing the multimedia data to be processed to obtain a plurality of slices, wherein the duration of the slices is longer than the preset duration;
selecting a target slice from the plurality of slices according to the aggregation behavior data and the voice data in the multimedia data to be processed;
generating target multimedia data corresponding to the multimedia data to be processed according to the target slice;
the aggregate behavior data includes a number of live-broadcast-room entrances, and the selecting a target slice from the plurality of slices according to the aggregate behavior data and voice data in the multimedia data to be processed includes:
According to the aggregation behavior data, selecting a slice corresponding to the maximum number of people in the live broadcasting room from the plurality of slices as a first slice;
and selecting a keyword slice from the plurality of slices according to voice data in the multimedia data to be processed, and taking the keyword slice and the first slice as target slices.
2. The method according to claim 1, wherein the aggregate behavior data further includes a virtual item click number, and the selecting a target slice from the plurality of slices according to the aggregate behavior data and the voice data in the multimedia data to be processed further includes:
and selecting a slice corresponding to the maximum value of the virtual object click number from the plurality of slices according to the aggregation behavior data, and taking the keyword slice, the first slice and the second slice as target slices.
3. The method according to claim 2, wherein the aggregate behavior data further includes an interaction data outlier, and the selecting a target slice from the plurality of slices according to the aggregate behavior data and the voice data in the multimedia data to be processed further includes:
And selecting a slice corresponding to the maximum abnormal value of the interaction data from the plurality of slices according to the aggregation behavior data, taking the slice as a third slice, and taking the keyword slice, the first slice, the second slice and the third slice as target slices.
4. The method according to claim 1, wherein generating target multimedia data corresponding to the multimedia data to be processed according to the target slice includes:
identifying voice data in the target slice to obtain subtitle information and key word information;
generating a text special effect according to the subtitle information and the key word information;
and generating target multimedia data corresponding to the multimedia data to be processed based on the text special effect and the target slice.
5. The method according to claim 4, wherein generating a text effect from the subtitle information and the accent word information includes:
if the key word information exists in the preset starting time of the target slice, generating a character special effect matched with the key word information;
and if the key word information does not exist in the preset starting time of the target slice, randomly selecting the character special effects from a preset special effect library.
6. The method according to claim 4, wherein generating a text effect from the subtitle information and the accent word information includes:
determining the display time, display position and display style corresponding to each key word;
the generating the target multimedia data corresponding to the multimedia data to be processed based on the text special effect and the target slice comprises the following steps:
and displaying the key words in the form of the display style at the display time and the display position of the target slice to obtain the target multimedia data.
7. A multimedia data processing apparatus, comprising:
the device comprises an acquisition unit, a processing unit and a processing unit, wherein the acquisition unit is configured to acquire to-be-processed multimedia data and operation behavior data corresponding to the to-be-processed multimedia data;
the aggregation unit is configured to perform aggregation on the operation behavior data corresponding to the multimedia data to be processed based on preset time lengths, and determine aggregation behavior data corresponding to each preset time length;
the slicing unit is configured to divide the multimedia data to be processed to obtain a plurality of slices, and the duration of the slices is longer than the preset duration;
A screening unit configured to perform screening of a target slice from the plurality of slices according to the aggregate behavior data and the voice data in the multimedia data to be processed;
a generating unit configured to perform generating target multimedia data corresponding to the multimedia data to be processed according to the target slice;
the aggregate behavior data includes live room entrances, the screening unit configured to perform:
according to the aggregation behavior data, selecting a slice corresponding to the maximum number of people in the live broadcasting room from the plurality of slices as a first slice;
and selecting a keyword slice from the plurality of slices according to voice data in the multimedia data to be processed, and taking the keyword slice and the first slice as target slices.
8. The multimedia data processing apparatus of claim 7, wherein the aggregate behavior data further comprises a virtual item click number, the filtering unit further configured to perform:
and selecting a slice corresponding to the maximum value of the virtual object click number from the plurality of slices according to the aggregation behavior data, and taking the keyword slice, the first slice and the second slice as target slices.
9. The multimedia data processing apparatus of claim 8, wherein the aggregate behavior data further comprises an interaction data outlier, the filtering unit further configured to perform:
and selecting a slice corresponding to the maximum abnormal value of the interaction data from the plurality of slices according to the aggregation behavior data, taking the slice as a third slice, and taking the keyword slice, the first slice, the second slice and the third slice as target slices.
10. The multimedia data processing apparatus of claim 7, wherein the generating unit is further configured to perform:
identifying voice data in the target slice to obtain subtitle information and key word information;
generating a text special effect according to the subtitle information and the key word information;
and generating target multimedia data corresponding to the multimedia data to be processed based on the text special effect and the target slice.
11. The multimedia data processing apparatus according to claim 10, wherein the generating unit is configured to perform:
if the key word information exists in the preset starting time of the target slice, generating a character special effect matched with the key word information;
And if the key word information does not exist in the preset starting time of the target slice, randomly selecting the character special effects from a preset special effect library.
12. The multimedia data processing apparatus according to claim 10, wherein the generating unit is configured to perform:
determining the display time, display position and display style corresponding to each key word;
and displaying the key words in the form of the display style at the display time and the display position of the target slice to obtain the target multimedia data.
13. An electronic device, comprising:
a processor;
a memory for storing the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the multimedia data processing method of any one of claims 1 to 6.
14. A computer readable storage medium, characterized in that instructions in the computer readable storage medium, when executed by a processor of a multimedia data processing electronic device, enable the multimedia data processing electronic device to perform the multimedia data processing method of any one of claims 1 to 6.
CN202210654770.2A 2022-06-10 2022-06-10 Multimedia data processing method and device, electronic equipment and storage medium Active CN115190356B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210654770.2A CN115190356B (en) 2022-06-10 2022-06-10 Multimedia data processing method and device, electronic equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210654770.2A CN115190356B (en) 2022-06-10 2022-06-10 Multimedia data processing method and device, electronic equipment and storage medium

Publications (2)

Publication Number Publication Date
CN115190356A CN115190356A (en) 2022-10-14
CN115190356B true CN115190356B (en) 2023-12-19

Family

ID=83512672

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210654770.2A Active CN115190356B (en) 2022-06-10 2022-06-10 Multimedia data processing method and device, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN115190356B (en)

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108540854A (en) * 2018-03-29 2018-09-14 努比亚技术有限公司 Live video clipping method, terminal and computer readable storage medium
CN109756751A (en) * 2017-11-07 2019-05-14 腾讯科技(深圳)有限公司 Multimedia data processing method and device, electronic equipment, storage medium
CN111447505A (en) * 2020-03-09 2020-07-24 咪咕文化科技有限公司 Video clipping method, network device, and computer-readable storage medium
CN111695505A (en) * 2020-06-11 2020-09-22 北京市商汤科技开发有限公司 Video processing method and device, electronic equipment and storage medium
CN112328816A (en) * 2020-11-03 2021-02-05 北京百度网讯科技有限公司 Media information display method and device, electronic equipment and storage medium
CN113094522A (en) * 2021-06-09 2021-07-09 北京达佳互联信息技术有限公司 Multimedia resource processing method and device, electronic equipment and storage medium
CN113473182A (en) * 2021-09-06 2021-10-01 腾讯科技(深圳)有限公司 Video generation method and device, computer equipment and storage medium
CN114143575A (en) * 2021-12-31 2022-03-04 上海爱奇艺新媒体科技有限公司 Video editing method and device, computing equipment and storage medium
CN114428867A (en) * 2022-01-20 2022-05-03 北京有竹居网络技术有限公司 Data mining method and device, storage medium and electronic equipment
CN114554271A (en) * 2022-01-26 2022-05-27 北京达佳互联信息技术有限公司 Information pushing and displaying method and device, electronic equipment and storage medium

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109756751A (en) * 2017-11-07 2019-05-14 腾讯科技(深圳)有限公司 Multimedia data processing method and device, electronic equipment, storage medium
CN108540854A (en) * 2018-03-29 2018-09-14 努比亚技术有限公司 Live video clipping method, terminal and computer readable storage medium
CN111447505A (en) * 2020-03-09 2020-07-24 咪咕文化科技有限公司 Video clipping method, network device, and computer-readable storage medium
CN111695505A (en) * 2020-06-11 2020-09-22 北京市商汤科技开发有限公司 Video processing method and device, electronic equipment and storage medium
CN112328816A (en) * 2020-11-03 2021-02-05 北京百度网讯科技有限公司 Media information display method and device, electronic equipment and storage medium
CN113094522A (en) * 2021-06-09 2021-07-09 北京达佳互联信息技术有限公司 Multimedia resource processing method and device, electronic equipment and storage medium
CN113473182A (en) * 2021-09-06 2021-10-01 腾讯科技(深圳)有限公司 Video generation method and device, computer equipment and storage medium
CN114143575A (en) * 2021-12-31 2022-03-04 上海爱奇艺新媒体科技有限公司 Video editing method and device, computing equipment and storage medium
CN114428867A (en) * 2022-01-20 2022-05-03 北京有竹居网络技术有限公司 Data mining method and device, storage medium and electronic equipment
CN114554271A (en) * 2022-01-26 2022-05-27 北京达佳互联信息技术有限公司 Information pushing and displaying method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN115190356A (en) 2022-10-14

Similar Documents

Publication Publication Date Title
CN105843615B (en) Notification message processing method and device
CN108932253B (en) Multimedia search result display method and device
CN111405302B (en) Live broadcast room attention reminding method and device, electronic equipment and storage medium
CN111783001A (en) Page display method and device, electronic equipment and storage medium
CN110688527A (en) Video recommendation method and device, storage medium and electronic equipment
CN111369271B (en) Advertisement ordering method and device, electronic equipment and storage medium
US20200012701A1 (en) Method and apparatus for recommending associated user based on interactions with multimedia processes
CN111556352B (en) Multimedia resource sharing method and device, electronic equipment and storage medium
CN113676671B (en) Video editing method, device, electronic equipment and storage medium
CN113259226B (en) Information synchronization method and device, electronic equipment and storage medium
CN112464031A (en) Interaction method, interaction device, electronic equipment and storage medium
CN106331328B (en) Information prompting method and device
CN110719530A (en) Video playing method and device, electronic equipment and storage medium
CN108062364A (en) Information displaying method and device
CN112291631A (en) Information acquisition method, device, terminal and storage medium
CN107402767B (en) Method and device for displaying push message
CN113901241B (en) Page display method and device, electronic equipment and storage medium
CN113032627A (en) Video classification method and device, storage medium and terminal equipment
CN104636476B (en) The method and device of commending friends
CN112131466A (en) Group display method, device, system and storage medium
CN114827709A (en) Video display method and device, electronic equipment and storage medium
CN111629270A (en) Candidate item determination method and device and machine-readable medium
CN113259754B (en) Video generation method, device, electronic equipment and storage medium
CN110175293B (en) Method and device for determining news venation and electronic equipment
CN115190356B (en) Multimedia data processing method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant