CN111405371A - Data processing method and related device - Google Patents

Data processing method and related device Download PDF

Info

Publication number
CN111405371A
CN111405371A CN202010158329.6A CN202010158329A CN111405371A CN 111405371 A CN111405371 A CN 111405371A CN 202010158329 A CN202010158329 A CN 202010158329A CN 111405371 A CN111405371 A CN 111405371A
Authority
CN
China
Prior art keywords
target
video playing
playing progress
voice message
target video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010158329.6A
Other languages
Chinese (zh)
Inventor
杨广煜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN202010158329.6A priority Critical patent/CN111405371A/en
Publication of CN111405371A publication Critical patent/CN111405371A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47217End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for controlling playback functions for recorded or on-demand content, e.g. using progress bars, mode or play-point indicators or bookmarks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/239Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests
    • H04N21/2393Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests involving handling client requests
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42203Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] sound input device, e.g. microphone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/488Data services, e.g. news ticker
    • H04N21/4882Data services, e.g. news ticker for displaying messages, e.g. warnings, reminders

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Information Transfer Between Computers (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The embodiment of the application discloses a data processing method and a related device, which are used for enhancing the interactivity with a user and displaying and popularizing scenes such as view-finding places except clothes, articles and the like without depending on hard implanted advertisements. The data processing method provided by the embodiment of the application comprises the following steps: acquiring a voice message and a target video playing progress corresponding to the voice message; determining a target object corresponding to the voice message according to the target video playing progress so as to obtain a target result corresponding to the target object; and sending a first prompt message to the client, wherein the first prompt message carries the target result so as to prompt the client to display the target result.

Description

Data processing method and related device
Technical Field
The embodiment of the application relates to the technical field of computers, in particular to a data processing method and a related device.
Background
At present, users usually use a television or a mobile terminal device such as a mobile phone or a tablet to watch videos. When watching a video, a user always watches a familiar scenario scene or a character garment and an article in a current scenario, and the user can hope to know more related information about the scenario scene or the character garment and the article in the current scenario. However, the conventional way of understanding is to use a patch advertisement or an embedded advertisement to promote and display articles, clothes, etc. used by a few of the drama characters of the crown name.
However, the current display method is only directed at a small amount of articles and clothes, and not only the episode scene, the view-finding place and the like are not displayed, but also the user is conflicted by the harsh embedded advertisements and the patch advertisements, and the interactivity with the user is lacked.
Disclosure of Invention
The embodiment of the application provides a data processing method and a related device, which are used for enhancing the interactivity with a user and displaying and popularizing scenes such as view-finding places except clothes, articles and the like without depending on hard implanted advertisements.
In a first aspect, an embodiment of the present application provides a data processing method, including:
acquiring a voice message and a target video playing progress corresponding to the voice message;
determining a target object corresponding to the voice message according to the target video playing progress so as to obtain a target result corresponding to the target object;
and sending a first prompt message to the client, wherein the first prompt message carries the target result so as to prompt the client to display the target result.
In a second aspect, an embodiment of the present application provides a data processing method, including:
receiving a voice message and a target video playing progress corresponding to the voice message;
sending the voice message and the target video playing progress to a server, so that the server determines a target object corresponding to the voice message according to the target video playing progress to obtain a target result corresponding to the target object;
receiving the first prompt message sent by the server, wherein the first prompt message carries the target result;
and displaying the target result.
In a third aspect, an embodiment of the present application provides a server, including:
the acquisition module is used for acquiring the voice message and the target video playing progress corresponding to the voice message;
the determining module is used for determining a target object corresponding to the voice message according to the target video playing progress so as to obtain a target result corresponding to the target object;
and the sending module is used for sending a first prompt message to the client, wherein the first prompt message carries the target result determined by the determining module so as to prompt the client to display the target result.
With reference to the third aspect, in a first possible implementation manner, the determining module includes:
the analysis unit is used for analyzing the voice message to obtain a first field;
the first determining unit is used for determining target information corresponding to the first field under the target video playing progress according to a preset corresponding relation when the first field is matched with the candidate field of the target video playing progress;
a first determining unit, configured to determine that the target information is a target object corresponding to the voice message.
With reference to the first possible implementation manner of the third aspect, in a second possible implementation manner,
the sending unit is further configured to send a second prompt message to the client when the first field is not matched with the candidate field of the target video playing progress, where the second prompt message is used to prompt that the first field has no corresponding target information in the target video playing progress.
With reference to the third aspect, in a third possible implementation manner, the determining module includes:
the second determining unit is used for determining the picture frame corresponding to the target video playing progress;
the second determining unit is configured to determine a similar picture, where the similar picture has a similar relationship with a picture corresponding to the picture frame;
the second determining unit is configured to determine the picture information corresponding to the similar picture as a target object corresponding to the voice message.
With reference to the first or second possible implementation manner of the third aspect, in a fourth possible implementation manner, the target information includes a target scene or a target commodity.
In a fourth aspect, an embodiment of the present application provides a client, including:
the receiving unit is used for receiving a voice message and a target video playing progress corresponding to the voice message;
the sending unit is used for sending the voice message received by the receiving unit and the target video playing progress to a server so that the server determines a target object corresponding to the voice message according to the target video playing progress to obtain a target result corresponding to the target object;
the receiving unit is configured to receive the first prompt message sent by the server, where the first prompt message carries the target result;
and the display unit is used for displaying the target result.
With reference to the fourth aspect, in a first possible implementation manner, the client further includes:
the receiving unit is configured to receive a second prompt message sent by the server, where the second prompt message is sent by the server when it is determined that a first field is not matched with a candidate field of the target video playing progress, and the first field is obtained by analyzing the voice message by the server;
and the prompting unit is used for prompting that the first field has no corresponding target information in the target video playing progress according to the second prompting message.
In a fifth aspect, an embodiment of the present application provides a server, including: a processor and a memory; the memory is configured to store program instructions, and when the server is running, the processor executes the program instructions stored in the memory to cause the server to perform the method of data processing according to the first aspect or any one of the possible implementations of the first aspect.
In a sixth aspect, an embodiment of the present application provides a client, including: a processor and a memory; the memory is used for storing program instructions, and when the client runs, the processor executes the program instructions stored in the memory, so that the client executes the method for data processing according to the second aspect or any one of the possible implementation manners of the second aspect.
In a seventh aspect, an embodiment of the present application provides a computer-readable storage medium, which includes instructions that, when executed on a computer, cause the computer to perform a method according to any one of the first aspect, the second aspect, or any one of the possible implementation manners of the second aspect.
In an eighth aspect, embodiments of the present application provide a computer program product containing instructions that, when executed on a computer, cause the computer to perform a method according to any one of the first aspect, the second aspect, or any one of the second aspect.
According to the technical scheme, the embodiment of the application has the following advantages:
in the embodiment of the application, the voice message and the target video playing progress corresponding to the voice message are obtained, so that the target object corresponding to the voice message can be determined according to the target video playing progress, the target result corresponding to the target object is obtained, the target result can be sent to the client side, and the client side displays the target result to prompt a user. In the embodiment, the target result is determined based on the voice message and the corresponding target video playing progress, so that the interaction with the user is enhanced, and scenes such as view-finding places except clothes, articles and the like can be displayed and popularized without depending on hard implanted advertisements.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings used in the description of the embodiments will be briefly introduced below, and it is apparent that the drawings in the following description are only some embodiments of the present application.
FIG. 1 is a schematic diagram of an architecture of a system for data processing in an embodiment of the present application;
FIG. 2 is a schematic diagram of an embodiment of a method for processing data provided in the present embodiment;
FIG. 3 is a schematic diagram of another embodiment of the data processing method provided in the present embodiment;
FIG. 4 is a schematic diagram of a preset mapping relationship provided in an embodiment of the present application;
FIG. 5 is a diagram illustrating a target result displayed in a pop-up frame manner in an embodiment of the present application;
FIG. 6 is a schematic diagram of one embodiment of a server provided in an embodiment of the present application;
FIG. 7 is a schematic diagram of one embodiment of a client provided in an embodiment of the present application;
fig. 8 is a schematic diagram of a hardware configuration of a communication apparatus in the embodiment of the present application.
Detailed Description
The embodiment of the application provides a data processing method and a related device, which are used for enhancing the interactivity with a user and displaying and popularizing scenes such as view-finding places except clothes, articles and the like without depending on hard implanted advertisements.
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The terms "first," "second," "third," "fourth," and the like in the description and in the claims of the present application and in the drawings described above, if any, are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the application described herein are capable of operation in sequences other than those illustrated or described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus. The naming or numbering of the steps appearing in the present application does not mean that the steps in the method flow have to be executed in the chronological/logical order indicated by the naming or numbering, and the named or numbered process steps may be executed in a modified order depending on the technical purpose to be achieved, as long as the same or similar technical effects are achieved.
Artificial Intelligence (AI) is a theory, method, technique and application system that uses a digital computer or a machine controlled by a digital computer to simulate, extend and expand human intelligence, perceive the environment, acquire knowledge and use the knowledge to obtain the best results. In other words, artificial intelligence is a comprehensive technique of computer science that attempts to understand the essence of intelligence and produce a new intelligent machine that can react in a manner similar to human intelligence. Artificial intelligence is the research of the design principle and the realization method of various intelligent machines, so that the machines have the functions of perception, reasoning and decision making.
The artificial intelligence software technology mainly comprises a computer vision technology, a voice processing technology, a natural language processing technology, a machine learning/deep learning (D L) and the like.
Among the key technologies of speech technology (speech technology) are automatic speech recognition technology (ASR) and speech synthesis technology (TTS), as well as voiceprint recognition technology. The computer can listen, see, speak and feel, and the development direction of the future human-computer interaction is provided, wherein the voice becomes one of the best viewed human-computer interaction modes in the future.
Generally, when a user watches videos through a television, a mobile phone, a computer, a tablet, or the like, the user always wants to know more related information about a character garment and articles in a scenario scene or a current scenario, but the conventional understanding method is to popularize and display articles, garments, and the like used by a small number of popular-name-crowned characters in a form of a tile advertisement or an embedded advertisement, not only does not display a scenario scene, a view-finding place, and the like, but also causes a conflicting emotion of the user due to a hard embedded advertisement and a tile advertisement, and lacks interactivity with the user.
Therefore, in the process of the rapid development of the voice technology, when an automatic voice recognition technology, which is a key technology in the voice technology, is adopted, a user inputs a voice instruction aiming at a certain target video playing progress through a voice remote controller or a voice device of a computer or a mobile terminal, so that a client can forward the target video playing progress and the voice instruction to a server at a background, and the server can analyze and process the voice instruction so as to determine a target result of the voice instruction at the target video playing progress.
The present embodiment mainly provides a data processing method, which can be applied to the data processing system shown in fig. 1, please refer to fig. 1, which is an architecture diagram of the data processing system in the present embodiment. As shown in fig. 1, the architecture diagram includes a client and a server, where the client may obtain a voice message and a target video playing progress corresponding to the voice message, so that the server may perform analysis processing on the voice instruction to determine a target result of the voice instruction in the target video playing progress, and not only does not depend on a hard advertisement, but also increases interactivity with a user, and can display scenes such as a viewing place except clothing, articles, and the like.
It should be understood that the above-mentioned client may be a client having a video playing function, and the client may be installed on a tablet Computer, a laptop Computer (L ap Computer), a notebook Computer, a Personal Digital Assistant (PDA) Computer, a mobile terminal, a smart terminal, a handheld terminal, etc., wherein the described mobile terminal includes but is not limited to a tablet, a smart phone, etc., and the described smart terminal includes but is not limited to a smart television, etc., and the embodiments of the present application will not be limited to the description.
The data processing method in this embodiment may be applied to the system architecture shown in fig. 1, and may also be applied to other system architectures, which are not limited herein.
To better understand the proposed solution in the embodiment of the present application, a specific flow in the embodiment will be described below with reference to an interaction between a server and a client, and please refer to fig. 2, which is a schematic diagram of an embodiment of a method for data processing provided in the embodiment, where the method may include:
201. and the client receives the voice message and the target video playing progress corresponding to the voice message.
In this embodiment, the target video playing progress refers to a specific time point when the target video is played by the user while watching the target video, and the voice message is a voice instruction input by the user at the current target video playing progress through a voice device on a terminal such as a mobile terminal or through a voice remote controller. For example: the current target video is 'XXXX movie & television tenth album', a backpack of the A role appears when the target video playing progress is 3 min 32 s-3 min 36 s, then the user can input 'the backpack of the A role' through voice, the client receives the voice message 'the backpack of the A role' at the moment, and the client acquires the current target video playing progress (3 min 32 s-3 min 36 s) corresponding to the backpack of the voice message 'the A role' through the playing component. It should be understood that the voice message at this time is a message obtained by the client while the target video playing is in progress.
202. And the client sends the voice message and the target video playing progress to the server.
In this embodiment, after receiving the corresponding voice message and the target video playing progress, the client may send the corresponding voice message and the target video playing progress to the server, so that the server may determine the target object corresponding to the voice message.
203. And the server determines a target object corresponding to the voice message according to the target video playing progress so as to obtain a target result corresponding to the target object.
In this embodiment, the server may perform corresponding processing on the objects such as scenes, goods, and the like of the video in different playing time periods in advance according to the video data provided by the video source side, so that the server may preset a corresponding relationship between each target video playing progress in the video and the corresponding object such as the scene or the goods.
Therefore, after receiving the target video playing progress sent by the client, the server can determine the target object corresponding to the voice message according to the target video playing progress, so that the corresponding target result is obtained according to the target object. For example, when the target object includes a target scene, the target result corresponding to the target scene may be a name of a viewing place, a route, a brief introduction, or the like of the target scene, or when the target object includes a target commodity, the target result corresponding to the target commodity may be a commodity name, a commodity purchase link, a commodity component, or the like of the target commodity. It should be understood that, in practical applications, the target result may also be other information besides the above-mentioned framing place name, route commodity name, commodity purchase link, etc., and will not be specifically limited in the embodiments of the present application.
Optionally, in other embodiments, determining the target object corresponding to the voice message may be performed as follows: the server determines a picture frame corresponding to the target video playing progress; the server determines similar pictures, wherein the similar pictures are pictures with similar relation with pictures corresponding to the picture frames; and the server determines the picture information corresponding to the similar picture as a target object corresponding to the voice message.
In this embodiment, each video is composed of one continuous picture, and each picture composing the video is a frame. Therefore, the server can determine the picture frame corresponding to the target video playing progress, and then query the pictures having the similar relation with the pictures corresponding to the picture frame in the database or through the Web page, so as to obtain the similar pictures. Thus, the server can determine the picture information corresponding to the similar picture as the target object corresponding to the voice message. That is, it is understood that other pictures having similarity with the picture of the current picture frame may be determined, and picture information of the other similar pictures is fed back to the client as a target object result, so that the client may feed back the target result corresponding to the target object to the user, and when the server does not record all objects such as goods or scenes in advance in the target video playing progress, the server may feed back the target result through the picture information of the similar pictures, thereby further enhancing interactivity with the user, and improving the popularization of the objects such as goods and scenes.
204. The server sends a first prompt message to the client, wherein the first prompt message carries the target result.
In this embodiment, in order to enhance interactivity with the user and further show and promote the commodity information or the scene information in the target video, the server may carry the target result in the first prompt message, so as to send the first prompt message to the client. Therefore, the client can display the target result carried by the first prompt message after receiving the first prompt message.
205. The client displays the target result.
In this embodiment, there are various ways to display the target result, for example: the client can display the target result on any area of the screen in a frame popping mode, or display the target result on any area of the screen in a suspension bubbling mode, or display the target result in a voice insertion broadcasting mode. It should be understood that any region of the screen described in the foregoing will not be limited in the embodiments of the present application. In addition, in practical applications, the display modes may include other modes besides the three modes described above, and the embodiments of the present application are not limited to the description.
In the embodiment of the application, the target object corresponding to the voice message can be determined according to the target video playing progress by acquiring the voice message and the target video playing progress corresponding to the voice message, so that the target result corresponding to the target object is acquired, and the target result can be sent to the client side, so that the client side displays the target result to prompt the user, the interactivity with the user is enhanced, and scenes such as a viewing place except clothes, articles and the like can be displayed and popularized without depending on hard implanted advertisements.
On the basis of the embodiment described in fig. 2, in order to better understand the scheme proposed in the embodiment of the present application, a specific flow in the embodiment is introduced below, please refer to fig. 3, which is a schematic diagram of another embodiment of the method for data processing provided in the embodiment, and the method may include:
301. the client receives the voice message and the target video playing progress corresponding to the voice message.
302. And the client sends the voice message and the target video playing progress to the server.
In this embodiment, steps 301 to 302 are similar to steps 201 to 202 described in fig. 2, and are not described herein again.
303. The server parses the voice message to obtain a first field.
In this embodiment, the first field may be a key field used to express the voice message. For example, assuming that the voice message is "knapsack of A role", the first field may include "A role", "knapsack". It should be understood that, in practical applications, the partitioning of the key fields may also be included but not limited, and will not be specifically described herein.
304. And when the first field is matched with the candidate field of the target video playing progress, the server determines that the first field has corresponding target information under the target video playing progress according to the preset corresponding relation.
In this embodiment, the server may perform corresponding processing on scenes, commodities, and the like of the video in different playing time periods in advance according to video data provided by the video source side, so as to obtain a candidate field of the target video playing progress. For example: when the target video playing progress is 3 minutes 32 seconds to 3 minutes 36 seconds, "backpack of a character a", "blue coat of a character B", and "pink hat of a character a" appear, then classification is performed according to different characters or different object objects, so as to obtain candidate fields of the target video playing progress, namely, "a character" a "," backpack "," pink hat "," B character ", and" blue coat ".
Obviously, the first field is matched with the candidate field of the target video playing progress, so that the server can determine that the first field has corresponding target information under the target video playing progress through a preset corresponding relation. That is to say, if the server does not record information of all characters and articles appearing in the target video playing progress in advance, even when the first field is matched with the candidate field, the server cannot determine the target object corresponding to the voice message in the target video playing progress, and cannot obtain the target result corresponding to the target object, so that only when the server determines that the first field has corresponding target information in the target video playing progress according to the preset corresponding relationship, the server can determine the target information as the target object, and thus the server can further obtain the target result corresponding to the target object based on the preset relationship.
It should be noted that the preset corresponding relationship mentioned above is a corresponding relationship between the target video playing progress and the target information. In addition, the target information may include a scene or a commodity. Specifically, fig. 4 may be referred to, which is a schematic diagram illustrating a corresponding relationship between a playing progress and target information provided in this embodiment of the application. As can be seen from fig. 4, the target video playing progress is 3 minutes 32 seconds to 3 minutes 36 seconds, which corresponds to the "backpack of the role a" and the "blue jacket of the role B", and the target information is the "backpack of the role a" and corresponds to the existence of the "model information of the backpack of the role a" and the target information is the "blue jacket of the role B" and corresponds to the existence of the link information of the blue jacket of the role B "; the target information is the mobile phone American image of the role A and corresponds to the mobile phone American image T8 information of the role A when the time is 8 min 57 s-9 min 06 s; the "ground of view" occurs between 14 minutes 35 seconds and 14 minutes 39 seconds: scene information of a certain abandoned railway of Harbin.
It can be seen that, in the first field, the corresponding target object is "the backpack of the a role" and the corresponding target result is "the model information of the backpack of the a role" exist in the target video playing progress. It should be understood that, in practical applications, the preset corresponding relationship may also be a corresponding relationship between other target video playing progress and other target information, and will not be specifically limited in this embodiment of the application.
Optionally, in other embodiments, the method may further include: and when the first field is not matched with the candidate field of the target video playing progress, the server sends a second prompt message to the client, wherein the second prompt message is used for prompting that the first field has no corresponding target information under the target video playing progress.
Optionally, in other embodiments, the client receives a second prompt message sent by the server, where the second prompt message is sent by the server when it is determined that the first field is not matched with the candidate field of the target video playing progress, and the first field is obtained by analyzing the voice message by the server; and the client prompts that the first field has no corresponding target information in the target video playing progress according to the second prompt message.
In this embodiment, if the server determines that the first field has no corresponding target information in the target video playing progress, a second prompt message is sent, so that the client can prompt the user according to the second prompt message, for example: some words such as "no content" may be displayed to indicate that the first field has no corresponding target information at the target video playing progress. In practical applications, other modes may also be used for displaying, and the embodiments of the present application are not limited specifically.
305. And when the target information exists, the server determines the target information as a target object so as to obtain a target result corresponding to the target object.
In this embodiment, when the first field has target information in the target video playing progress, the server may directly determine the target information as the target object. For example: based on the example described above for step 304, the "viewfinding: harbin certain abandoned railways "are determined as target results.
306. The server sends a first prompt message to the client, wherein the first prompt message carries the target result.
In this embodiment, step 306 is similar to step 204 described in fig. 2, and details thereof are not repeated herein.
307. The client displays the target result.
In this embodiment, reference may be made to fig. 5, which is a flowchart illustrating a target result displayed in a pop-up frame manner in the embodiment of the present application. As can be seen from fig. 5, when the target video playing progress is displayed in the range from 14 minutes 35 seconds to 14 minutes 39 seconds, "finding the scene: harbin certain abandoned railways ". It should be understood that there are various ways to display the target result, which can be understood by referring to step 205 described in fig. 2, and will not be described herein in detail.
The scheme provided by the embodiment of the application is mainly introduced from the perspective of a method. It is to be understood that the hardware structure and/or software modules for performing the respective functions are included to realize the above functions. Those of skill in the art will readily appreciate that the various illustrative modules and algorithm steps described in connection with the embodiments disclosed herein may be implemented as hardware or combinations of hardware and computer software. Whether a function is performed as hardware or computer software drives hardware depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
In the embodiment of the present application, functional modules of the apparatus may be divided according to the above method example, for example, each functional module may be divided corresponding to each function, or two or more functions may be integrated into one processing module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode. It should be noted that, in the embodiment of the present application, the division of the module is schematic, and is only one logic function division, and there may be another division manner in actual implementation.
Referring to fig. 6, please refer to fig. 6 for a detailed description of the server 40 in the embodiment of the present application, where fig. 6 is a schematic diagram of an embodiment of the server 40 provided in the embodiment of the present application, and the server 40 includes:
an obtaining module 401, configured to obtain a voice message and a target video playing progress corresponding to the voice message;
a determining module 402, configured to determine, according to the target video playing progress, a target object corresponding to the voice message, so as to obtain a target result corresponding to the target object;
a sending module 403, configured to send a first prompt message to the client, where the first prompt message carries the target result determined by the determining module 402, so as to prompt the client to display the target result.
Optionally, on the basis of the embodiment corresponding to fig. 6, in another embodiment of the server 40 provided in this embodiment of the present application, the determining module 402 includes:
the analysis unit is used for analyzing the voice message to obtain a first field;
the first determining unit is used for determining target information corresponding to the first field under the target video playing progress according to a preset corresponding relation when the first field is matched with the candidate field of the target video playing progress;
and the first determining unit is used for determining the target information as a target object.
Alternatively, on the basis of the above-mentioned alternative embodiment of fig. 6, in another embodiment of the server 40 provided in the embodiment of the present application,
the sending module 403 is further configured to send a second prompt message to the client when the first field is not matched with the candidate field of the target video playing progress, where the second prompt message is used to prompt that the first field has no corresponding target information in the target video playing progress.
Optionally, on the basis of the embodiment corresponding to fig. 6, in another embodiment of the server 40 provided in the embodiment of the present application, the determining module 402 includes:
the second determining unit is used for determining the picture frame corresponding to the target video playing progress;
the second determining unit is configured to determine a similar picture, where the similar picture has a similar relationship with a picture corresponding to the picture frame;
the second determining unit is configured to determine the picture information corresponding to the similar picture as a target object corresponding to the voice message.
Optionally, on the basis of the optional embodiment corresponding to fig. 6, in another embodiment of the server 40 provided in the embodiment of the present application, the target information includes a target scene or a target commodity.
The server 40 in the embodiment of the present application is described above from the perspective of a modular functional entity, and the client 50 in the embodiment of the present application is described below from the perspective of modularity. Referring to fig. 7, fig. 7 is a schematic diagram of an embodiment of a client 50 provided in an embodiment of the present application, where the client 50 includes:
a receiving unit 501, configured to receive a voice message and a target video playing progress corresponding to the voice message;
a sending unit 502, configured to send the voice message and the target video playing progress received by the receiving unit 501 to a server, so that the server determines a target object corresponding to the voice message according to the target video playing progress, so as to obtain a target result corresponding to the target object;
the receiving unit 501 is configured to receive the first prompt message sent by the server, where the first prompt message carries the target result;
a display unit 503, configured to display the target result.
Optionally, on the basis of the optional embodiment corresponding to fig. 7, in another embodiment of the client 50 provided in the embodiment of the present application, the client 50 further includes:
the receiving unit 501 is configured to receive a second prompt message sent by the server, where the second prompt message is sent by the server when it is determined that a first field is not matched with the candidate field of the target video playing progress, and the first field is obtained by analyzing the voice message by the server;
and the prompting unit is used for prompting that the first field has no corresponding target information in the target video playing progress according to the second prompting message.
The server 40 and the client 50 in the embodiment of the present application are described above from the perspective of a modular functional entity, and the server 40 and the client 50 in the embodiment of the present application are described below from the perspective of hardware processing. Fig. 8 is a schematic diagram of a hardware configuration of a communication apparatus in the embodiment of the present application. As shown in fig. 8, the communication apparatus may include:
the communication device includes at least one processor 601, communication lines 607, memory 603 and at least one communication interface 604.
The processor 601 may be a general processing unit (CPU), a microprocessor, an application-specific integrated circuit (server IC), or one or more ICs for controlling the execution of programs in accordance with the present invention.
The communication link 607 may include a path for transmitting information between the aforementioned components.
The communication interface 604 may be any device, such as a transceiver, for communicating with other devices or communication networks, such as AN ethernet, a Radio Access Network (RAN), a wireless local area network (W L AN), etc.
The memory 603 may be a read-only memory (ROM) or other type of static storage device that may store static information and instructions, a Random Access Memory (RAM) or other type of dynamic storage device that may store information and instructions, and may be separate and coupled to the processor via a communication link 607. The memory may also be integral to the processor.
The memory 603 is used for storing computer-executable instructions for executing the present application, and is controlled by the processor 601 to execute the instructions. The processor 601 is configured to execute computer-executable instructions stored in the memory 603, so as to implement the data processing method provided by the above-mentioned embodiment of the present application.
Optionally, the computer-executable instructions in the embodiments of the present application may also be referred to as application program codes, which are not specifically limited in the embodiments of the present application.
In particular implementations, the communication device may include multiple processors, such as processor 601 and processor 602 in fig. 8, as an example. Each of these processors may be a single-core (single-CPU) processor or a multi-core (multi-CPU) processor. A processor herein may refer to one or more devices, circuits, and/or processing cores for processing data (e.g., computer program instructions).
In one implementation, the communication apparatus may further include an output device 605 and an input device 606. Output device 605 is in communication with processor 601 and may display information in a variety of ways. The input device 606 is in communication with the processor 601 and may receive user input in a variety of ways. For example, the input device 606 may be a mouse, a touch screen device, or a sensing device, among others.
The communication device may be a general-purpose device or a dedicated device. In particular implementations, the communication device may be a desktop, laptop, web server, wireless terminal device, embedded device, or a device having a similar structure as in fig. 8. The embodiment of the present application does not limit the type of the communication device.
The acquiring module 401 and the receiving unit 501 may be implemented by an input device 606, the transmitting module 403 and the transmitting unit 502 may be implemented by an output device 605, and the determining module 402 and the displaying unit 503 may be implemented by the processor 501 or the processor 602.
In the above embodiments, the implementation may be wholly or partially realized by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product.
It is clear to those skilled in the art that, for convenience and brevity of description, the specific working processes of the above-described systems, apparatuses and units may refer to the corresponding processes in the foregoing method embodiments, and are not described herein again.
In the several embodiments provided in the present application, it should be understood that the disclosed system, apparatus and method may be implemented in other manners. For example, the above-described apparatus embodiments are merely illustrative, and for example, a division of a unit is merely a logical division, and an actual implementation may have another division, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.
Units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
The integrated unit, if implemented in the form of a software functional unit and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present application may be substantially implemented or contributed to by the prior art, or all or part of the technical solution may be embodied in a software product, which is stored in a storage medium and includes instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method of the embodiments of the present application. And the aforementioned storage medium includes: various media capable of storing program codes, such as a usb disk, a removable hard disk, a read-only memory (ROM), a Random Access Memory (RAM), a magnetic disk, or an optical disk.
The above embodiments are only used to illustrate the technical solutions of the present application, and not to limit the same; although the present application has been described in detail with reference to the foregoing embodiments, it should be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions in the embodiments of the present application.

Claims (10)

1. A method of data processing, comprising:
acquiring a voice message and a target video playing progress corresponding to the voice message;
determining a target object corresponding to the voice message according to the target video playing progress so as to obtain a target result corresponding to the target object;
and sending a first prompt message to a client, wherein the first prompt message carries the target result so as to prompt the client to display the target result.
2. The method of claim 1, wherein determining the target object corresponding to the voice message according to the target video playing progress comprises:
analyzing the voice message to obtain a first field;
when the first field is matched with the candidate field of the target video playing progress, determining target information corresponding to the first field under the target video playing progress according to a preset corresponding relation;
and determining the target information as a target object corresponding to the voice message.
3. The method of claim 2, further comprising:
and when the first field is not matched with the candidate field of the target video playing progress, sending a second prompt message to the client, wherein the second prompt message is used for prompting that the first field has no corresponding target information in the target video playing progress.
4. The method of claim 1, wherein determining the target object corresponding to the voice message according to the target video playing progress comprises:
determining a picture frame corresponding to the target video playing progress;
determining similar pictures, wherein the similar pictures are pictures with similar relation with the pictures corresponding to the picture frames;
and determining the picture information corresponding to the similar picture as a target object corresponding to the voice message.
5. The method of claim 2 or 3, wherein the target information comprises a target scene or a target commodity.
6. A method of data processing, comprising:
receiving a voice message and a target video playing progress corresponding to the voice message;
sending the voice message and the target video playing progress to a server, so that the server determines a target object corresponding to the voice message according to the target video playing progress to obtain a target result corresponding to the target object;
receiving the first prompt message sent by the server, wherein the first prompt message carries the target result;
and displaying the target result.
7. The method of claim 6, further comprising:
receiving a second prompt message sent by the server, wherein the second prompt message is sent by the server when a first field is determined to be not matched with the candidate field of the target video playing progress, and the first field is obtained by analyzing the voice message by the server;
and prompting that the first field has no corresponding target information under the target video playing progress according to the second prompt message.
8. A server, characterized in that the server comprises:
an input/output (I/O) interface, a processor and a memory,
the memory has stored therein program instructions;
the processor is configured to execute program instructions stored in the memory to perform the method of any of claims 1 to 5.
9. A client, characterized in that the client comprises:
an input/output (I/O) interface, a processor and a memory,
the memory has stored therein program instructions;
the processor is configured to execute program instructions stored in the memory to perform the method of any of claims 1-5 or 6-7.
10. A computer-readable storage medium comprising instructions that, when executed on a computer device, cause the computer device to perform the method of any of claims 1-5 or 6-7.
CN202010158329.6A 2020-03-09 2020-03-09 Data processing method and related device Pending CN111405371A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010158329.6A CN111405371A (en) 2020-03-09 2020-03-09 Data processing method and related device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010158329.6A CN111405371A (en) 2020-03-09 2020-03-09 Data processing method and related device

Publications (1)

Publication Number Publication Date
CN111405371A true CN111405371A (en) 2020-07-10

Family

ID=71413242

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010158329.6A Pending CN111405371A (en) 2020-03-09 2020-03-09 Data processing method and related device

Country Status (1)

Country Link
CN (1) CN111405371A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111932322A (en) * 2020-09-24 2020-11-13 腾讯科技(深圳)有限公司 Advertisement display method, related device, equipment and storage medium
CN112035086A (en) * 2020-08-19 2020-12-04 海尔优家智能科技(北京)有限公司 Audio playing method and device

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102509235A (en) * 2011-10-31 2012-06-20 福州博远无线网络科技有限公司 Purchase interactive method based on mobile video of intelligent handheld device
CN103297840A (en) * 2012-03-01 2013-09-11 阿里巴巴集团控股有限公司 Additional information display method and system based on video moving focus
CN106254925A (en) * 2016-08-01 2016-12-21 乐视控股(北京)有限公司 Destination object extracting method based on video identification, equipment and system
CN106488332A (en) * 2016-10-25 2017-03-08 华为软件技术有限公司 TV shopping method and apparatus
CN108833948A (en) * 2018-06-14 2018-11-16 广州视源电子科技股份有限公司 Commodity method for pushing, system, readable storage medium storing program for executing and terminal
CN109936773A (en) * 2017-12-19 2019-06-25 展讯通信(上海)有限公司 Implementation method, device and the user terminal of video shopping
CN110458640A (en) * 2019-06-27 2019-11-15 拉扎斯网络科技(上海)有限公司 A kind of merchandise display method, apparatus, server and can storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102509235A (en) * 2011-10-31 2012-06-20 福州博远无线网络科技有限公司 Purchase interactive method based on mobile video of intelligent handheld device
CN103297840A (en) * 2012-03-01 2013-09-11 阿里巴巴集团控股有限公司 Additional information display method and system based on video moving focus
CN106254925A (en) * 2016-08-01 2016-12-21 乐视控股(北京)有限公司 Destination object extracting method based on video identification, equipment and system
CN106488332A (en) * 2016-10-25 2017-03-08 华为软件技术有限公司 TV shopping method and apparatus
CN109936773A (en) * 2017-12-19 2019-06-25 展讯通信(上海)有限公司 Implementation method, device and the user terminal of video shopping
CN108833948A (en) * 2018-06-14 2018-11-16 广州视源电子科技股份有限公司 Commodity method for pushing, system, readable storage medium storing program for executing and terminal
CN110458640A (en) * 2019-06-27 2019-11-15 拉扎斯网络科技(上海)有限公司 A kind of merchandise display method, apparatus, server and can storage medium

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112035086A (en) * 2020-08-19 2020-12-04 海尔优家智能科技(北京)有限公司 Audio playing method and device
CN112035086B (en) * 2020-08-19 2024-03-22 海尔优家智能科技(北京)有限公司 Audio playing method and device
CN111932322A (en) * 2020-09-24 2020-11-13 腾讯科技(深圳)有限公司 Advertisement display method, related device, equipment and storage medium

Similar Documents

Publication Publication Date Title
CN110286976B (en) Interface display method, device, terminal and storage medium
CN113473182B (en) Video generation method and device, computer equipment and storage medium
CN111476871B (en) Method and device for generating video
CN111241340B (en) Video tag determining method, device, terminal and storage medium
CN110162670B (en) Method and device for generating expression package
CN113395542B (en) Video generation method and device based on artificial intelligence, computer equipment and medium
CN107169135B (en) Image processing method and device and electronic equipment
CN110868635A (en) Video processing method and device, electronic equipment and storage medium
CN112839223B (en) Image compression method, image compression device, storage medium and electronic equipment
CN113870133B (en) Multimedia display and matching method, device, equipment and medium
CN110458130A (en) Character recognition method, device, electronic equipment and storage medium
US20170262991A1 (en) Browsing interface for item counterparts having different scales and lengths
CN111176533A (en) Wallpaper switching method, device, storage medium and terminal
CN111405371A (en) Data processing method and related device
CN112866577B (en) Image processing method and device, computer readable medium and electronic equipment
CN112822539B (en) Information display method, device, server and storage medium
CN110517096A (en) Content method for implantation, device, electronic equipment and storage medium
CN108833964A (en) A kind of real-time successive frame Information Embedding identifying system
CN110784762B (en) Video data processing method, device, equipment and storage medium
CN114780181B (en) Resource display method, device, computer equipment and medium
CN112149599A (en) Expression tracking method and device, storage medium and electronic equipment
CN116703797A (en) Image fusion method, image fusion system, computer device and storage medium
CN114846812A (en) Abstract video generation method and device and server
CN112235516B (en) Video generation method, device, server and storage medium
CN113761281B (en) Virtual resource processing method, device, medium and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20200710

RJ01 Rejection of invention patent application after publication