CN115037959A - Search result display method, device, equipment and non-transitory computer storage medium - Google Patents

Search result display method, device, equipment and non-transitory computer storage medium Download PDF

Info

Publication number
CN115037959A
CN115037959A CN202210622097.4A CN202210622097A CN115037959A CN 115037959 A CN115037959 A CN 115037959A CN 202210622097 A CN202210622097 A CN 202210622097A CN 115037959 A CN115037959 A CN 115037959A
Authority
CN
China
Prior art keywords
video
key information
information
target
target video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210622097.4A
Other languages
Chinese (zh)
Inventor
林涤生
张宁静
冯杨兰
王琳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing ByteDance Network Technology Co Ltd
Original Assignee
Beijing ByteDance Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing ByteDance Network Technology Co Ltd filed Critical Beijing ByteDance Network Technology Co Ltd
Priority to CN202210622097.4A priority Critical patent/CN115037959A/en
Publication of CN115037959A publication Critical patent/CN115037959A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/232Content retrieval operation locally within server, e.g. reading video streams from disk arrays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/432Content retrieval operation from a local storage medium, e.g. hard-disk

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention provides a search result display method, a device, equipment and a non-transitory computer storage medium, and relates to the technical field of video search, wherein the method comprises the steps of receiving search information, and acquiring a target video and resource information matched with the search information, wherein the target video comprises a plurality of video clips, and the resource information comprises key information for describing core contents of all the video clips; and the display length proportion of each piece of key information is matched with the duration proportion of each video clip, the target video and the resource information are displayed on the search result page, and the video clip corresponding to the target key information is played on the search result page in response to the triggering operation of the target key information. The method can quickly find the interesting information from the video and improve the user experience.

Description

Search result display method, device, equipment and non-transitory computer storage medium
Technical Field
The present disclosure relates to the field of video search technologies, and in particular, to a method, an apparatus, a device, and a non-transitory computer storage medium for displaying search results.
Background
At present, various videos can be recommended to a user by a video playing platform, and the user can share the videos with other users, so that the user can passively acquire the videos, and the consumption and entertainment requirements of the user are met. When a user actively searches videos, a target video meeting the requirement needs to be selected from a search result list.
When the user selects the target video from the search result list, the user needs to browse each video content and determine whether the video content is the target video according to the video content. However, the content of the target video is more, so that the speed of browsing the video content by the user is slower, and the user experience is affected because the information of interest cannot be quickly found from the video.
Disclosure of Invention
According to an aspect of the present disclosure, there is provided a search result presentation method, the method including:
receiving search information, and acquiring a target video matched with the search information and resource information of the target video; the target video comprises a plurality of video clips, and the resource information comprises key information describing core content of each video clip; the display length proportion of each piece of key information is correspondingly matched with the duration proportion of each video clip;
displaying the target video and the resource information on a search result page;
and responding to the triggering operation of the target key information, and playing a video clip corresponding to the target key information on the search result page.
According to another aspect of the present disclosure, there is provided a search result presentation apparatus including:
the receiving module is used for receiving search information, acquiring a target video matched with the search information and resource information of the target video; the target video comprises a plurality of video clips, and the resource information comprises key information describing core content of each video clip; the display length proportion of each piece of key information is correspondingly matched with the duration proportion of each video clip;
and the display module is used for displaying the target video and the resource information on a search result page, responding to the trigger operation of the target key information, and playing a video clip corresponding to the target key information on the search result page.
According to another aspect of the present disclosure, there is provided an electronic device including:
a processor; and the number of the first and second groups,
a memory for storing a program;
wherein the program comprises instructions which, when executed by the processor, cause the processor to perform the method according to an exemplary embodiment of the present disclosure.
According to another aspect of the present disclosure, there is provided a non-transitory computer readable storage medium characterized by storing computer instructions for causing the computer to perform the method according to exemplary embodiments of the present disclosure.
In one or more technical solutions provided in the embodiments of the present disclosure, the resource information of the target video includes key information describing core content of each video clip, and therefore, when the target video and the resource information are displayed on the search result page, the video clip of interest may be quickly determined based on the core content of the video clip described by the key information, then the target key information is determined based on the video clip of interest, and in response to a trigger operation on the target key information, the video clip corresponding to the target key information is played on the search result page, thereby achieving a purpose of quickly browsing the video clip corresponding to the target key information. Meanwhile, the display length proportion of each key information is matched with the duration proportion of each video clip, so that the display length of each key information can relatively reflect the duration of the corresponding video clip, and when the target key information is determined, the target key information can be determined through the core content of the video clip described by the key information, and whether the duration of the corresponding video clip meets the requirement of a user for fast browsing the video clip can be determined through the display length of the target key information. When the duration of the video clip corresponding to the target key information meets the requirement of a user for quickly browsing the video clip, the video clip corresponding to the target key information can meet the requirement of the user for the content of the video clip, and can ensure that the user quickly finishes browsing the video content within a limited time range, thereby achieving the effect of ensuring both browsing speed and browsing quality.
Drawings
Further details, features and advantages of the disclosure are disclosed in the following description of exemplary embodiments, taken in conjunction with the accompanying drawings, in which:
FIG. 1 shows a schematic diagram of an example search network architecture in which various methods described herein may be implemented, according to an example embodiment of the present disclosure;
FIG. 2 is a flowchart illustrating a search result presentation method according to an exemplary embodiment of the present disclosure;
FIG. 3 illustrates an interface diagram of a search results page of an exemplary embodiment of the present disclosure;
FIG. 4 shows another interface diagram of a search results page of an exemplary embodiment of the present disclosure;
FIG. 5A shows an interface diagram of a search results page without key information presented in an exemplary embodiment of the present disclosure;
FIG. 5B is an interface diagram illustrating a search results page showing key information according to an exemplary embodiment of the present disclosure;
FIG. 5C is a diagram illustrating an operation of an exemplary embodiment of the present disclosure in recalling 4# key information;
FIG. 5D is an interface diagram illustrating the presentation of # 4 key information on a search results page in accordance with an exemplary embodiment of the present disclosure;
FIG. 6A shows another interface diagram of a search results page presenting key information according to an exemplary embodiment of the present disclosure;
FIG. 6B illustrates an operational diagram of an exemplary embodiment of the present disclosure to call out 3# critical information;
FIG. 6C is an interface diagram illustrating the presentation of # 3 key information on a search results page in accordance with an exemplary embodiment of the present disclosure;
FIG. 7A is a schematic diagram of an interface for displaying key information in a full screen mode according to an exemplary embodiment of the disclosure;
FIG. 7B is a diagram illustrating an operation of an exemplary embodiment of the present disclosure to screen out 4# key information;
FIG. 7C is a schematic diagram of an interface for displaying the 4# key information in a full screen manner according to an exemplary embodiment of the disclosure;
FIG. 8A is another schematic interface diagram of a full screen presentation of key information of an exemplary embodiment of the present disclosure;
FIG. 8B illustrates another operational diagram of an exemplary embodiment of the present disclosure calling out 4# critical information;
FIG. 8C shows another interface diagram of a full screen presentation of 4# key information according to an exemplary embodiment of the disclosure;
FIG. 8D shows a schematic diagram of a full screen presentation review interface of an exemplary embodiment of the present disclosure;
FIG. 9A illustrates a video clip playback operation of an exemplary embodiment of the present disclosure;
FIG. 9B shows an interface schematic of the 2# segment of an exemplary embodiment of the present disclosure;
FIG. 10 illustrates another flowchart of a search result presentation method of an exemplary embodiment of the present disclosure;
FIG. 11 is a schematic diagram illustrating a triggered operation of a target video according to an exemplary embodiment of the present disclosure;
FIG. 12A is a schematic diagram illustrating a triggering operation of a target video on a search result page according to an exemplary embodiment of the disclosure;
FIG. 12B is a diagram illustrating another triggering operation of a target video on a search results page according to an exemplary embodiment of the present disclosure;
FIG. 12C is a diagram illustrating a trigger result of a target video on a search results page according to an exemplary embodiment of the present disclosure;
FIG. 12D is a diagram illustrating another trigger result of a target video on a search results page according to an exemplary embodiment of the present disclosure;
FIG. 13 is a diagram illustrating a trigger result of a full screen presentation of a target video according to an exemplary embodiment of the present disclosure;
FIG. 14 is a schematic diagram illustrating another trigger result of a full screen presentation of a target video according to an exemplary embodiment of the present disclosure;
FIG. 15 shows an interface diagram of a search results page with a teletext entry revealed in an exemplary embodiment of the disclosure;
fig. 16 shows a presentation interface schematic of the teletext description information of an exemplary embodiment of the disclosure;
FIG. 17 shows a schematic block diagram of a search result presentation apparatus according to an example embodiment of the present disclosure;
FIG. 18 shows a schematic block diagram of a chip according to an example embodiment of the present disclosure;
FIG. 19 illustrates a block diagram of an exemplary electronic device that can be used to implement embodiments of the present disclosure.
Detailed Description
Embodiments of the present disclosure will be described in more detail below with reference to the accompanying drawings. While certain embodiments of the present disclosure are shown in the drawings, it is to be understood that the present disclosure may be embodied in various forms and should not be construed as limited to the embodiments set forth herein, but rather are provided for a more thorough and complete understanding of the present disclosure. It should be understood that the drawings and embodiments of the disclosure are for illustration purposes only and are not intended to limit the scope of the disclosure.
It should be understood that the various steps recited in the method embodiments of the present disclosure may be performed in a different order, and/or performed in parallel. Moreover, method embodiments may include additional steps and/or omit performing the illustrated steps. The scope of the present disclosure is not limited in this respect.
The term "including" and variations thereof as used herein is intended to be open-ended, i.e., "including but not limited to". The term "based on" is "based, at least in part, on". The term "one embodiment" means "at least one embodiment"; the term "another embodiment" means "at least one additional embodiment"; the term "some embodiments" means "at least some embodiments". Relevant definitions for other terms will be given in the following description. It should be noted that the terms "first", "second", and the like in the present disclosure are only used for distinguishing different devices, modules or units, and are not used for limiting the order or interdependence relationship of the functions performed by the devices, modules or units.
It is noted that references to "a", "an", and "the" modifications in this disclosure are intended to be illustrative rather than limiting, and that those skilled in the art will recognize that "one or more" may be used unless the context clearly dictates otherwise.
Before describing exemplary embodiments of the present disclosure, terms related to exemplary embodiments of the present disclosure are first explained as follows:
the video structuralization is a technology for structuralizing and extracting important information in video content, and the important information is organized into text information or visual graphic information which can be understood by a computer and people by processing means such as space-time segmentation, feature extraction, object identification, deep learning and the like according to semantic relation.
The search detection word point ratio is called the ratio that the search word has clicks, also called query point ratio. After searching a certain search term, whether the searched result has click behavior or not is judged. The action of clicking does not include: corresponding flow guide modules for error correction/map, related search and the like; multiple results click the same click, where the click behavior has only the distinguished results of 0 and 1.
Video View Volume (VV) is one of the important parameters for measuring the Video playing effect, which can be interpreted as one play when the Video is played for more than 5 seconds.
The segmentation chapter function refers to segmenting a video according to video content to form a plurality of video segments, and adding a title describing the video segments to each video segment.
A User Generated Content (UGC) mode is a mode for mobilizing the enthusiasm of netizens to participate in video creation, and generally, an enterprise collects video works related to the enterprise through activities. In the UGC model, the net friends are no longer just viewers, but rather become producers and providers of internet content.
Optical Character Recognition (OCR) is the meaning of recognizing Optical characters by image processing and pattern Recognition technology, and is an important aspect in the field of research and application of automatic Recognition technology.
Automatic Speech Recognition (ASR) is a technology that converts human Speech into text.
The recall rate is also called recall rate, which is the ratio of the number of the searched related documents to the number of all the related documents in the document library, and the balance is the recall rate of the retrieval system; the precision is the ratio of the number of the searched relevant documents to the total number of the searched documents, and the precision is measured by the precision ratio of the searching system.
The embodiment of the disclosure provides a search result display method, which can be applied to a video search scene, and a structured video mode is adopted to ensure that a user can quickly acquire core content of a video clip from a video, so that the user can quickly find interesting information to meet the requirement of active exploration of the user.
Fig. 1 shows a schematic diagram of an example search network architecture in which various methods described herein may be implemented, according to an example embodiment of the present disclosure. As shown in fig. 1, a search network architecture 100 of an exemplary embodiment of the present disclosure includes a client system 102 and a search server system 103 connected to each other through a network 101. It should be understood that the search network architecture shown in fig. 1 is applicable to all examples of the exemplary embodiments of the present disclosure, and the illustrated network architecture is convenient for understanding the aspects of the exemplary embodiments of the present disclosure, and should not be construed as limiting the application scenarios, the presentation manners, and various communication relationships of the exemplary embodiments of the present disclosure.
In an alternative, as shown in FIG. 1, the first resource server 104 bypasses the network 101 and the search server system 103. The second resource server 105 is connected to the search server system 103 via the network 101. Additionally, the first resource server 104 may communicate with the search server system 103 via the network 101. The second resource server 105 may be connected to the search server system 103 bypassing the network 101.
Furthermore, although fig. 1 shows that the number of client systems 102, search server systems 103, first resource servers 104, second resource servers 105, and networks 101 is one, it does not constitute a limitation on the number of client systems 102, search server systems 103, first resource servers 104, second resource servers 105, and networks 101. Based on this, the search network architecture 100 of the exemplary embodiments of the present disclosure may include a plurality of client systems 102, a plurality of search server systems 103, a plurality of first resource servers 104, a plurality of second resource servers 105, and a plurality of networks 101.
As shown in fig. 1, the first resource server 104 and the second resource server 105 of the exemplary embodiment of the present disclosure may provide services and resources to the search server system 103 through different network communication methods. Where resources may include, but are not limited to, multimedia resources such as video, text, audio, etc., and services may include, but are not limited to, services such as computer vision recognition, speech recognition, natural language processing, etc. The first resource server 104 and the second resource server 105 may be configured as cloud servers such as a private cloud, a public cloud, a hybrid cloud, a private cloud, and the like, and may be deployed with training capabilities and deployment capabilities capable of executing neural network models of the foregoing and the services.
As shown in fig. 1, the network 101 of the exemplary embodiments of the present disclosure may include one or more networks 101, and any suitable network 101 is contemplated. By way of example and not limitation, one or more portions of Network 101 may include an ad hoc Network, an intranet, an extranet, a Virtual Private Network (VPN), a Local Area Network (LAN), a Wireless Local Area Network (WLAN), a Wide Area Network (WAN), a Wireless Wide Area Network (WWAN), a Metropolitan Area Network (metropolar Area Network), a MAN), a portion of the internet, a portion of a Public Switched Telephone Network (PSTN), a cellular Telephone Network, or a combination of two or more of these.
As shown in FIG. 1, the client system 102, the search server system 103, the first resource server 104, and the second resource server 105 of the exemplary embodiments of the present disclosure may be connected to the communication network 101 or to each other by a network communication link 106. The disclosed exemplary embodiments contemplate any suitable network communication link 106.
Illustratively, the one or more Network communication links 106 may include one or more wireline (e.g., Digital Subscriber Line (abbreviated DSL) or wireline Cable Service Interface specification (abbreviated DOCSIS)), wireless (e.g., wireless fidelity (abbreviated WiFi) or Worldwide Interoperability for Microwave Access (abbreviated WiMAX)) or Optical (e.g., Synchronous Optical Network (abbreviated SONET) or Synchronous Digital Hierarchy (abbreviated SDH)) links.
Illustratively, as shown in fig. 1, the one or more network communication links 106 each comprise an ad hoc network, an intranet, an extranet, a VPN, a LAN, a WLAN, a WAN, a WWAN, a MAN, a portion of the internet, a portion of the PSTN, a cellular technology-based network, a satellite communication technology-based network, another network communication link 106, or a combination of two or more such network communication links 106. The network communication links 106 are not necessarily the same throughout the search network architecture 100. The one or more first network communication links may differ in one or more respects from the one or more second network communication links.
In an alternative, as shown in FIG. 1, client system 102 of exemplary embodiments of the present disclosure may access search server system 103 directly or via network 101 using web browser/search application 1021 or a local application associated with search server system 103 (e.g., a mobile search application, a messaging application, another suitable application, or any combination thereof).
Illustratively, as shown in FIG. 1, a client system 102 of exemplary embodiments of the present disclosure may be an electronic device that includes hardware, software, or embedded logic components or a combination of two or more such components and is capable of performing the appropriate functions implemented or supported by the client system 102.
Illustratively, as shown in fig. 1, the client System 102 of exemplary embodiments of the present disclosure may include a computer System, such as a desktop computer, a notebook or laptop computer, a netbook, a tablet computer, an e-book reader, a Global Positioning System (GPS) device, a camera, a Personal Digital Assistant (PDA), a handheld electronic device, a cellular telephone, a smart phone, other suitable electronic device, or any suitable combination thereof. It should be appreciated that any suitable client system 102 is contemplated by embodiments of the present invention. Client system 102 may enable a network user at client system 102 to access network 101. Client system 102 may enable its user to communicate with other users at other client systems 102.
In an alternative, as shown in FIG. 1, the search server system 103 of the exemplary embodiments of the present disclosure may be a network addressable computing system that may carry online searches. Search server system 103 may generate, store, receive, and transmit search information. The search server system 103 may be accessed directly by other components of the search network architecture 100 or via the network 101.
Illustratively, as shown in fig. 1, the search server system 103 of the exemplary embodiments of the present disclosure may include one or more servers 1031. Each server 1031 may be a single server or a distributed server spanning multiple computers or multiple data centers. Server 1031 may be of various types, such as, but not limited to, a web server, a news server, a mail server, a messaging server, an advertising server, a file server, an application server, an exchange server, a database server, a proxy server, another server adapted to perform the functions or processes described herein, or any combination thereof. In a particular implementation, each server 1031 may include hardware, software, or embedded logic components or a combination of two or more such components for performing the appropriate functions implemented or supported by server 1031.
Illustratively, as shown in FIG. 1, the search server system 103 of exemplary embodiments of the present disclosure may include one or more data stores 1032. Data storage 1032 may be used to store various types of information. In one particular implementation, the information stored in data stores 1032 may be organized according to particular data structures. In one particular implementation, each data store 1032 may be a relational database, a column database, a relational database, or other suitable database. Although embodiments of the present invention describe or illustrate a particular type of database, embodiments of the present invention contemplate any suitable type of database. One particular implementation may provide an interface that enables client system 102, search server system 103, or first resource server 104, second resource server 105 to manage, retrieve, modify, add, or delete information stored in data store 1032.
Illustratively, as shown in fig. 1, the search server system 103 of exemplary embodiments of the present disclosure may store a plurality of reading objects or a plurality of video objects in one or more data stores 1032. Search server system 103 may provide users with the ability to take actions on various types of items or objects that search server system 103 supports.
Illustratively, as shown in FIG. 1, the search server system 103 of the exemplary embodiments of the present disclosure is capable of linking various entities. By way of example and not limitation, search server system 103 may enable users to interact with each other and receive content from first resource server 104, second resource server 105, or other entities, or allow users to interact with these entities through an Application programming Interface (abbreviated API) or other communication channel.
In one alternative, as shown in FIG. 1, the first resource server 104, the second resource server 105 of the exemplary embodiments of the present disclosure may include one or more types of servers, one or more data stores, one or more interfaces (including but not limited to APIs), one or more web services, one or more content sources, one or more networks, or any other suitable component with which, for example, a server may communicate. The first resource server 104, the second resource server 105 may be operated by an entity different from the entity operating the search server system 103.
In another alternative, as shown in fig. 1, the search server system 103 and the first and second resource servers 104, 105 of the exemplary embodiment of the present disclosure may cooperate with each other to provide search services to users of the search server system 103 or the first and second resource servers 104, 105. In this sense, the search server system 103 may provide a platform or backbone that other systems, such as the first resource server 104, the second resource server 105, may use to provide search services and functionality to users over the internet.
The method provided by the embodiment of the disclosure can be executed by a terminal with a display function. The terminal may be a mobile phone, a tablet computer, a wearable device, a vehicle-mounted device, a notebook computer, an ultra-mobile personal computer (abbreviated as UMPC), a netbook, a PDA, and a wearable device based on augmented reality (abbreviated as AR) and/or virtual reality (abbreviated as VR) technology, and the like.
For example, when the terminal is a wearable device, the wearable device may also be a generic term for intelligently designing daily wearing by applying a wearing technology, and developing wearable devices, such as glasses, gloves, watches, clothes, shoes, and the like. A wearable device is a portable device that is worn directly on the body or integrated into the clothing or accessories of the user. The wearable device is not only a hardware device, but also realizes powerful functions through software support, data interaction and cloud interaction. The generalized wearable intelligent device has the advantages that the generalized wearable intelligent device is complete in function and large in size, can realize complete or partial functions without depending on a smart phone, such as a smart watch or smart glasses, and only is concentrated on a certain application function, and needs to be matched with other devices such as the smart phone for use, such as various smart bracelets for monitoring physical signs, smart jewelry and the like.
The search result presentation method of the exemplary embodiment of the present disclosure may be executed by a terminal, and may also be executed by a chip applied to the terminal. The following embodiments take terminals as the execution subjects, respectively.
Fig. 2 shows a flowchart of a search result presentation method according to an exemplary embodiment of the present disclosure. As shown in fig. 2, a search result presentation method according to an exemplary embodiment of the present disclosure includes:
step 201: search information is received. The search information of the exemplary embodiments of the present disclosure may refer to content input by a user in a search box of an interactive interface, and requires a search engine to perform a search operation and feed back a search result. It should be understood that, for example, the search engine of the exemplary embodiment of the present disclosure may include a web search engine, a vertical search engine, a knowledge search engine, a question and answer search engine, a cloud search engine, etc. in terms of product classification, and may be constructed based on one or more combinations of gold, Python, elastic search, etc. in terms of technology classification.
The search information of the exemplary embodiments of the present disclosure may include one or more keywords. When the keyword is plural, the search information may further include logic between the keywords, and the like. If the search information received by the terminal is the search information input by the user in a voice mode, the voice information can be converted into text information based on a voice recognition technology, keywords in the text information are extracted based on semantic analysis, and the keywords can be displayed in a search box. If the search information is picture information, the terminal may identify a picture input by the user based on an existing intelligent image recognition technology (such as an end intelligent image recognition technology or an online image recognition technology), so as to obtain a search keyword, and may display the search keyword in the search box.
In practical application, a user inputs search information such as search words in an interactive interface displayed by a terminal, for example, a search box of a search page, and triggers a search button, the terminal can send a search request containing the search information to a search server system, the search server system can crawl various search results related to the search information from a network through a search engine, and the search results are aggregated and sorted to be displayed on a search result page of the terminal. For example: the search server system can crawl and call various search results meeting the requirements related to the search information from one or more of a data storage, a first resource server and a second resource server in the search server system, and the search results are displayed on a search result page of the terminal. The search result page of the exemplary embodiment of the present disclosure may include various videos, articles, audios, etc., but is not limited thereto. It should be understood that the video of the search result page may be displayed in a landscape mode or a portrait mode, and whether displayed in a landscape mode or a portrait mode, the method of the exemplary embodiment of the present disclosure is consistent with the method of the exemplary embodiment of the present disclosure.
FIG. 3 shows an interface diagram of a search results page of an exemplary embodiment of the present disclosure. As shown in fig. 3, the related keyword "oven roast" is displayed in a search box 301 of a search results page 300, which may illustratively include video-type search results 303 and teletext-type search results 304 under a general option 302. The video copy 3031 such as a title and a summary included in the video search result 303 and the comment field 3032 are displayed in an area other than the video 3033. For example: in fig. 3, the video copy 3031 is located above the video 3033, and the comment area 3032 is located below the video 3033. It should be understood that the number of videos shown in fig. 3 is one, but in practice other larger numbers of videos, articles, audios, etc. may be included, and fig. 3 should not be understood as a limitation to the search results of the exemplary embodiment of the present disclosure.
FIG. 4 shows another interface diagram of a search results page of an exemplary embodiment of the present disclosure. As shown in fig. 4, the relevant key word "oven roasting" is displayed in a search box 401 of a search results page 400, which may illustratively include four videos 403 under a video option 402, with each video title, summary, etc. video copy 404 displayed in a floating layer on the respective video 403. It should be understood that the number of videos shown in fig. 4 is 4, but other larger or smaller numbers of videos may also be included in practice, and fig. 4 should not be construed as limiting the search results of the exemplary embodiment of the present disclosure.
Step 202: and acquiring a target video matched with the search information and resource information of the target video, wherein the target video can comprise a plurality of video clips, and the resource information comprises key information describing core content of each video clip. It should be understood that the duration of the target video of the exemplary embodiments of the present disclosure may be defined according to actual situations. Illustratively, the duration of the target video may be greater than or equal to 1 min. For example: the duration of the target video is 1min to 10min, for example: the duration of the target video is 1min, 3min, 5min or 10min and the like. The number of the video clips can be more than 2, and can be set according to the time length of the target video. For example: for the current short video, the number of video segments can be set to be not more than 11.
The resource information of the exemplary embodiment of the present disclosure includes key information describing core content of each video clip, and thus, the key information may have a relationship corresponding to each video clip. When the key information describes the core content of a video clip, the key information and the video clip of the described core content have a corresponding relation; when the key information describes core content of two or more video clips, the key information may have a correspondence relationship with the two or more video clips of the described core content.
In the exemplary embodiment of the disclosure, the making videos of the cabbage and meat dumplings of the food making category are taken as examples to describe the key information corresponding to different video segments, and table 1 shows a key information labeling list of the making videos of the cabbage and meat dumplings of the exemplary embodiment of the disclosure.
TABLE 1 Key information annotation list of making video of Chinese cabbage meat dumplings
Figure RE-RE-GDA0003742630140000091
Figure RE-RE-GDA0003742630140000101
As can be seen from table 1, the production video of the cabbage dumpling of the exemplary embodiment of the present disclosure is divided into 7 video segments, which are respectively the 1# segment, the 2# segment, the 3# segment, the 4# segment, the 5# segment, the 6# segment, and the 7# segment, and two adjacent video segments may be divided by the segment end point of the video segment with the front timestamp. When the user needs to know how to season streaky pork, the position of the video clip can be known from the key information of the 4# clip, and when the user needs to know the amount of the Chinese cabbage added to the minced meat, the position of the video clip can be known from the key information of the 5# clip.
It can be seen that, since the key information of the exemplary embodiment of the present disclosure can describe the core content of the corresponding video clip, it can be indirectly known whether the content of the corresponding video clip is the content that is of interest to the user through the key information (from the perspective of the user, when the content of the video clip is the content of interest, the content of the video clip is useful content for the user). If the content which the user is interested in exists, the user can not only play the corresponding video clip content, but also browse the whole video, therefore, the exemplary embodiment of the disclosure can not only quickly screen out the content which the user is interested in from the target video based on the key information, but also improve the video playing amount, the video watching completion degree, the staying time, the query dot ratio and the like.
In an alternative manner, the search information of the exemplary embodiment of the present disclosure may include information associated with the operation step semantics, information associated with the classification introduction semantics, and information associated with both the operation step and the classification introduction semantics.
When the search information contains information semantically related to the operation steps, the search result fed back to the terminal by the search server system may include various contents having time course elements expressing an operation method, an execution method, a detection method, and the like.
For example, when the keyword included in the search information is "food preparation", the search result fed back to the terminal by the search server system may include various food preparation methods, including but not limited to a method for preparing a cabbage-meat dumpling, a method for preparing a roast meat, and other videos, articles, graphic data, and the like. For another example, when the keyword included in the search information is "travel strategy", the search result fed back to the terminal by the search server system may include travel strategies of various regions, including but not limited to videos, articles, and graphic data such as travel schedules.
When the search information contains information associated with both the classified introduction semantics, the search result fed back to the terminal by the search server system may include various contents belonging to different categories of the same topic. For example: when the keyword contained in the search information is 'skin care product recommendation', the search result fed back to the terminal by the search server system can comprise videos, articles, image-text data and the like which recommend skin care products of the same attribute and different types.
When the search information contains information related to both the operation steps and the classification introduction semantics, the search result fed back to the terminal by the search server system can comprise various contents with time process elements belonging to the same subject and can also comprise detailed classification contents of one or more of the contents with the time process elements. For example: when the keyword contained in the search information is "food exploration shop", the search result fed back to the terminal by the search server system may include videos, articles, graphic and text data and the like of various food exploration shops. The content of the food exploration store can comprise the trial eating experience of going to various types of food stores, wherein the trial eating experience of a certain food store can also comprise the cooking method of food, so that the content contained in the search result can embody not only different types of content belonging to the same theme, but also the content with time process elements.
Step 203: and displaying the target video and the resource information on the search result page. When the target videos and the resource information are displayed on the search result page, the display length proportion of each piece of key information is matched with the duration proportion of each video clip. Taking table 1 as an example, when the length ratio of the 1# segment, the 2# segment, the 3# segment, the 4# segment, the 5# segment, the 6# segment, and the 7# segment is 16:22:6:17:171:25:10, the duration ratio of the 1# key information, the 2# key information, the 3# key information, the 4# key information, the 5# key information, the 6# key information and the 7# key information is 16:22:6:17:171:25: 10. Therefore, the display length of each piece of key information can relatively reflect the duration of the corresponding video clip, and a user can determine whether the duration of the corresponding video clip meets the requirement of the user for quickly browsing the video clip according to the display length of the key information.
In practical application, when the display length proportion of each piece of key information is matched with the duration proportion of each video clip, the size of the display area of each piece of key information in the target direction is matched with the size of the corresponding video clip in the target direction. That is, after the length of the target video in the target direction is segmented according to the time length of each video segment, the length of the corresponding video segment in the target direction can be obtained. On the basis, the length of the display area of each key message in the target direction can be equal to the length of the corresponding video clip, and can also be proportional to the length of the corresponding video frequency band. The ratio may be less than 1 or greater than 1. Taking the video of the cooking method of oven grilled meat as an example, the detailed contents of the key information of each video clip can be referred to table 2.
TABLE 2 Key information labeling list of cooking method video of oven barbecue
Figure RE-RE-GDA0003742630140000111
As can be seen from table 2, the target video includes 4 video clips, which are respectively a 1# clip, a 2# clip, a 3# clip, and a 4# clip, and there are 4 pieces of key information, which are respectively 1# key information, 2# key information, 3# key information, and 4# key information. The time length ratio of the 1# segment, the 2# segment, the 3# segment and the 4# segment is 2:6:3: 1.
The target video is divided into four regions in the width direction according to the time length proportion of the 1# segment, the 2# segment, the 3# segment and the 4# segment, and the length of each region in the width direction is the length of the corresponding video segment in the width direction. For example: when the length L of the target video in the width direction is 60mm, the target video is area-divided in the width direction in accordance with 2:6:3:1, and the obtained length L1 of the first area in the width direction is equal to 10mm, the length of the corresponding 1# clip in the width direction is 10mm, the length L2 of the second area in the width direction is equal to 30mm, the length of the corresponding 2# clip in the width direction is 30mm, the length L3 of the third area in the width direction is equal to 30mm, the length of the corresponding 3# clip in the width direction is 15mm, the length L4 of the fourth area in the width direction is equal to 5mm, and the length of the corresponding 4# clip in the width direction is 5 mm.
When the length of the display area of each piece of key information in the width direction is equal to the length of the corresponding video clip in the width direction, the length of the display area of the 1# key information in the width direction is 10mm, the length of the display area of the 2# key information in the width direction is 30mm, the length of the display area of the 3# key information in the width direction is 15mm, and the length of the display area of the 4# key information in the width direction is 5 mm.
When the length of the display area of each piece of key information in the width direction is proportional to the length of the corresponding video clip in the width direction, the length of the display area of the 1# key information is 15mm, the length of the display area of the 2# key information in the width direction is 45mm, the length of the display area of the 3# key information in the width direction is 22.5mm, and the length of the display area of the 4# key information in the width direction is 7.5 mm. It can be seen that, when the length of the display area of each key information in the width direction is equal to the length of the corresponding video clip in the width direction, the ratio of the length of the display area of each key information in the width direction to the length of the corresponding video clip in the width direction is 1.5.
The target video and the resource information of the exemplary embodiment of the present disclosure may be displayed on the search result page at the same time, or the target video may be displayed first, and then it is determined whether the user is interested in the target video, or whether there is an incentive to browse the target video. And if the target video is motivated to browse, displaying key information contained in the resource information. It should be understood that whether the playing time of the target video reaches the preset time or not may be detected, and if the playing time reaches the preset time, it indicates that the user is motivated to browse the target video. At this time, the playing time of the target video meets the display requirement, and the terminal can display the resource information on the search result page. The preset time duration can be set according to actual conditions, for example: 2s to 10s or one of 5s, 8s and 10 s.
When the key information is displayed, the exemplary embodiment of the disclosure may divide the key information into multiple lines for display, and may also display the key information in a single line along the target direction. When the display areas of the key information are distributed along the target direction, the display areas of the key information can be distributed according to the sequence of the timestamps of the corresponding video clips, and can also be distributed randomly.
For example, when the key information is displayed in an area outside the search result page, the position of the display area of the key information may be designed according to the layout of the display interface. The display area of each piece of key information may be located in at least one of an upper area of the target video, a lower area of the target video, a left area of the target video, and a right area of the target video.
Illustratively, when the display areas of the key information are distributed according to the sequence of the timestamps of the corresponding video clips, the user can regularly see the core content of different video clips described by the key information along the same direction according to the sequence distribution of the timestamps of the video clips, so that the user experience is improved. For example: the display area of each key information can be displayed along the width direction of the target video or along the length direction of the target video.
All the key information of the exemplary embodiment of the present disclosure may be presented on the search result page at one time, or only a part of the key information may be presented. When part of the key information is partially displayed on the search result page and the other part of the key information is not displayed on the search result page, if the display areas of the key information are distributed along the target direction, the key information can slide on the search result page, the key information which is not displayed can be called out in a sliding mode, and the core content of the corresponding video clip is identified from the called key information. The display position of the key information can be displayed in an area of the search result page outside the target video, or can be displayed on the target video in a floating layer manner. In order to avoid mutual influence, the display area of the key information is not overlapped with other texts, components and the like, so that the problem that the display area of the key information interferes with the areas to influence the use experience is avoided.
When each piece of key information is displayed in a lower area, such as a functional area, of the search result page, wherein the lower area is located in the target video, the display area of each piece of key information is distributed along the width direction of the target video, and the display areas of the piece of key information are located above the interactive components such as approval, comment viewing and sharing, so that interactive controls such as approval, comment viewing and sharing are avoided. The following describes a display manner of the key information on the search result page, taking the preset time length as 5s as an example. It should be understood that the following illustrates the displaying process of the key information on the search result page, and the distribution position of the display area of the key information may also be in other ways, and is not limited to the following illustrated one way.
Fig. 5A illustrates an interface diagram of a search results page without key information in an exemplary embodiment of the present disclosure. As shown in fig. 5A, in the search result page 500, the playing time of the target video 501 is 3s, which is less than 5s, and therefore, the search result page 500 does not show the resource information of the target video 501. Fig. 5B is a schematic diagram of an interface for displaying key information on a search result page according to an exemplary embodiment of the disclosure. As shown in fig. 5B, in the search result page 500, the playing time of the target video 501 is 5s, which is equal to 5s, and the display requirement is met, so that the key information of each video clip is displayed below the target video 500 along the width direction of the target video. The display area of the key information is positioned below the target video.
Due to the limitation of the length of the display screen of the terminal in the width direction of the target video, as can be seen from fig. 5B and table 2, the search result page 500 shown in fig. 5B completely shows the key information 502 # and 2# and partially shows the key information 504 # and does not show the key information # 4. To view the non-presented 4# key information, one or more of the already presented 1# key information 502, 2# key information 503 and 3# key information 504 may be slid to bring up the 4# key information.
Fig. 5C shows an operation diagram of calling out the 4# key information according to the exemplary embodiment of the present disclosure. Fig. 5D shows an interface diagram of presenting the 4# key information on the search result page according to the exemplary embodiment of the present disclosure. As shown in fig. 5C, in the search result page 500, the 2# key information 503 may be slid in the target video 501 in a direction indicated by a dotted arrow, so that the complete 3# key information 504 and 4# key information 505 are called out in the search result page 500 shown in fig. 5D, and the 4# key information 505 is presented below the target video 501 in the manner shown in fig. 5D.
For example, when each piece of key information is displayed on the target video in a floating layer, the display area of each piece of key information may be located above the progress bar or below the progress bar. Meanwhile, according to the display state of the target video, various anchor points, such as interactive components like praise, share, comment and play switches, video files like titles and abstracts, existing on the target video can be avoided in the display area of the key information.
When each piece of key information is displayed on the target video in a floating layer, the target video may be displayed in a manner as shown in fig. 5A. At this time, each piece of key information can be directly displayed on the target video in a floating layer manner by ignoring the playing time of the target video, and can also be displayed under the condition that the playing time of the target video is longer than the preset time.
Fig. 6A illustrates another interface diagram of a search results page showing key information according to an exemplary embodiment of the present disclosure. As shown in fig. 6A, the display area of each key information in the search result page 600 is located in a floating layer at the bottom functional area of the target video 601 and above the progress bar 602, limited by the width of the target video 601, and as can be seen from fig. 6A and table 2, the search result page 600 shown in fig. 6A displays the key information 1# 603 and the key information 2# but does not display the key information 3# and the key information 4# 604. To view the non-presented 3# key information, one or both of the already presented 1# key information 603 and 2# key information 604 may be slid to bring up the 3# key information.
Fig. 6B shows an operation diagram of calling out 3# key information according to an exemplary embodiment of the present disclosure. As shown in fig. 6B, the 2# key information 604 may be slid in the direction indicated by the dotted arrow in the search result page 600, so that the 3# key information 605 on the target video of the search result page 600 in fig. 6C is displayed in a floating layer above the progress bar 602 of the target video 601 in the manner shown in fig. 6C.
The target video of the search result page of the exemplary embodiment of the present disclosure may also be displayed in a full screen manner, for example: when the search result based on the search information query is obtained, the target video included in the search result list may be displayed in a non-full screen manner as shown in fig. 3 or fig. 4, or may be displayed in a full screen manner.
When the target video is displayed in a full screen mode, and the target video is not played or the playing time is shorter than the preset time, all key information contained in the resource information can be directly displayed, so that a user can conveniently obtain information interesting to the target video in time. It should be understood that under the condition of full-screen display of the video, interactive components such as titles, abstracts and other video files, praise, comment checking, sharing and the like can be avoided, interference on users is avoided, and use experience is reduced.
When the target video is displayed in a non-full screen mode in the manner shown in fig. 3 or fig. 4, if the user needs to display the target video in a full screen mode, the terminal device may respond to the user to click, double-click or click a full screen button on the target video, so that the target video enters full screen display. It should be understood that the operation modes of the target video entering the full screen display are various, and are not limited to three modes of clicking, double clicking and full screen button clicking.
In this process, since the operation of entering the full screen presentation is a related operation of the terminal in response to the user for the target video, in this case, the user has a strong motivation to browse the target video. At this time, whether the target video which is displayed in a full screen mode is in a playing state or not, or whether the time length for playing the target video is greater than or equal to the preset time length or not, the key information contained in the resource information can be directly displayed.
Fig. 7A shows an interface diagram of displaying key information in a full screen manner according to an exemplary embodiment of the present disclosure. As shown in fig. 7A, in a full screen interface 700, a display area of each piece of key information is located in a floating layer manner in a bottom functional area of a target video 701, above a progress bar 702, and simultaneously below interactive components 704 such as a video file 703 of a title, an abstract and the like, and a comment 7041, a comment view 7042, a share 7043 and the like.
Due to the limitation of the length of the display screen in the width direction of the target video 701, as can be seen from fig. 7A and table 2, the 1# key information 705, the 2# key information 706, and the 3# key information 707 are displayed in the overall interface 700 shown in fig. 7A, but the 4# key information is not displayed. In order to view the non-presented 4# key information, the key information of the exemplary embodiment of the present disclosure may slide one or more of the already presented 1# key information 705, 2# key information 706, and 3# key information 707, and call out the 4# key information.
Fig. 7B shows an operation diagram of the exemplary embodiment of the present disclosure for calling out 4# key information on full screen. As shown in fig. 7B, the 2# key information 706 can be slid along the direction indicated by the dotted arrow in the comprehensive interface 700, so as to call up the 4# key information 708 shown in fig. 7C, so that the 4# key information 708 is shown in a floating layer manner above the progress bar 702 of the target video 701 and below the video files 703 such as titles and summaries and the interactive components 704 such as praise 7041, review 7042, share 7043, and the like, according to the display manner of the full-screen interface 700 shown in fig. 7C.
Fig. 8A shows another interface diagram of a full screen presentation of key information according to an exemplary embodiment of the present disclosure. As shown in fig. 8A, in a full screen interface 800, a display area of each key information is located below a progress bar 802 of a target video 801 and above a pull-down control 803. Due to the limitation of the length of the display screen in the width direction of the target video 801, as can be seen from fig. 8A and table 2, the 1# key information 804, the 2# key information 805, and the 3# key information 806 are displayed on the full screen interface shown in fig. 8A, but the 4# key information is not displayed. In order to view the non-presented 4# key information, the key information of the exemplary embodiment of the present disclosure may slide one or more of the already presented 1# key information 804, 2# key information 805, and 3# key information 806, and call out the 4# key information.
Fig. 8B shows another operation diagram of the exemplary embodiment of the present disclosure to call out 4# key information in full screen. As shown in fig. 8B, in the target video 801 of the full screen interface 800, the 2# key information 805 may be slid along the direction indicated by the dotted arrow, thereby calling up the 4# key information 807 shown in the full screen interface 800 of fig. 8C, so that the 4# key information 807 is displayed below the progress bar 802 of the target video 801 and above the pull-down control 803 in the manner shown in fig. 8C.
As shown in fig. 8C, when the user needs to call up the comment interface, the user can click on the pull-down control 803, and the comment interface shown in fig. 8D is displayed. For example: a comment interface 808 may be presented in full screen interface 800 shown in fig. 8D in response to operation of a drop-down control 803 shown in fig. 8C. The comment interface 808 may be presented in a floating layer on the target video 801, or in other forms.
Step 204: and responding to the triggering operation of the target key information, and playing a video clip corresponding to the target key information on the search result page. The target key information may be defined herein as one in which a plurality of key information is triggered.
In practical application, the key information may be displayed on the search result page in the form of a control, a label, and the like, and an association relationship or a link is established with the corresponding video clip, so that when the user triggers the target key information, the video clip corresponding to the target key information can be played.
When the target video and the resource information are displayed on the search result page, the interested video segment can be quickly determined based on the core content of the video segment described by the key information, and then the target key information is determined based on the interested video segment. Based on the method, when the triggering operation of the target key information is responded, the video clip corresponding to the target key information can be played on the search result page, and the purpose of quickly browsing the video clip corresponding to the target key information is achieved. Meanwhile, when the target key information is determined, the target key information can be determined through the core content of the video clip described by the key information, and whether the duration of the corresponding video clip meets the requirement of a user for fast browsing the video clip can be determined through the display length of the target key information. When the duration of the video clip corresponding to the target key information meets the requirement of a user for quickly browsing the video clip, the video clip corresponding to the target key information can meet the requirement of the user for the content of the video clip, and can ensure that the user quickly finishes browsing the video content within a limited time range, thereby achieving the effect of ensuring both browsing speed and browsing quality.
Fig. 9A illustrates a video clip playing operation diagram according to an exemplary embodiment of the present disclosure, and fig. 9B illustrates an interface diagram of a 2# clip according to an exemplary embodiment of the present disclosure. As can be seen from fig. 9A and table 2, in the search result page 900, if the 2# key information is determined to be the target key information, when the display area 901 of the 2# key information is clicked, the 2# clip corresponding to the 2# key information can be played on the search result page 900 in the manner shown in fig. 9B. As can be seen, the exemplary embodiments of the present disclosure may play the corresponding video clip by clicking the display area of the key information. It should be understood that the key information display in the exemplary embodiment of the present disclosure is displayed on the target video in a floating layer manner, and the video clip playing may also be performed in a manner referring to fig. 9A, which is not described in detail below.
Illustratively, when the display area of the key information is located in the search result page, if the background color of the display area of the key information cannot be distinguished from the color of the key information display area corresponding to the search result page, the display area of the key information identified by the user may be affected, and therefore, the background color of the display area of the key information may have a high contrast with the color of the key information display area corresponding to the search result page, so that the user may conveniently identify the display area of the key information. At the moment, the user can easily trigger the target key information without carefully identifying the key information display area or triggering the key information text, so that the user experience is improved.
For example: in a full-screen playing state, if the color of the comment area below the progress bar is black, the background color of the display area of the key information can be highlighted, so that the position of the display area of the key information can be accurately judged, and therefore it is guaranteed that a user can trigger the target key information through the display area of the key information.
Similarly, when the user identifies that the key information is displayed on the target video in a floating layer manner, the background color of the key information display area and the color of the key information display area corresponding to the target video can be adjusted, so that the key information and the target video have obvious contrast, and the user can conveniently trigger the target key information. In order to reduce the shielding of the target video picture, the transparency of the floating layer can be adjusted, and the influence on the user experience of watching the video due to overhigh contrast is avoided.
For example, when the target video is played, if the played video segments are different, a jump may also occur on the corresponding key information, for example: the display area of the key information can be provided with a state identifier of the key information, and when a video clip corresponding to the key information is played, the state identifier of the key information is displayed as a playing state; and when the video clip corresponding to the key information is played or not played, displaying the state identifier of the key information as a closed playing state. That is, when the video clip is switched, the key information of the playing status is also switched.
As can be seen from fig. 9A and table 2, when the 1# clip is played, the status flag 902 of the 1# key information is in the playing status, and the status flag 903 of the 2# key information is in the closed playing status. Similarly, as can be seen from fig. 9B and table 2, when the 2# clip is played, the status flag 903 of the 2# key information is in the playing status, and the status flag 902 of the 1# key information is in the closed playing status. Therefore, when the 1# clip is switched to the 2# clip for playing, the key information in the playing state is switched from the 1# key information to the 2# key information.
For example, when the exemplary embodiment of the present disclosure responds to a trigger operation on the target key information, the corresponding target video may be played from the start position of the target key information, so that the user may play from the start position of the video clip corresponding to the target key information, and information of the video clip is prevented from being missed. For example; when fig. 9A clicks on the presentation area 901 of the 2# key information, referring to fig. 9B, the 2# clip corresponding to the 2# key information is played from the end point of the 1# clip, i.e., the start point of the 2# clip.
When the target key information is triggered, if the video clip corresponding to the triggered target key information is being played, the corresponding target video can be directly played at the initial position of the target key information, or the playing progress of the target video corresponding to the target key information is not changed.
When the user accidentally touches the display area of the target key information corresponding to the video clip being played, the playing progress of the target video corresponding to the target key information is not changed, and the problem that the video clip is watched again accidentally is avoided.
For example: with reference to fig. 9B and table 2, if the user accidentally touches the display area of the 2# key information, the playing progress of the 2# clip is maintained, so that on one hand, the time wasted by the user viewing the 2# clip again can be avoided, and on the other hand, the time wasted by the user repositioning the playing progress of the 2# clip (adjusting to the original playing progress) can also be avoided.
In order to increase the diversity of the video clip playing manners and the comprehensiveness of the collected information, fig. 10 shows another flowchart of the search result presentation method according to the exemplary embodiment of the present disclosure. As shown in fig. 10, the method for displaying search results of the present disclosure may further include:
step 1001: and in response to the triggering operation on the target video, determining the target video segment according to the termination position of the triggering operation. The trigger operation of the exemplary embodiment of the present disclosure on the target video may be a sliding operation on the target video, or may be a trigger operation on a progress bar of the target video when the target video has the progress bar. It should be understood that, after the target video segment is determined, the target video segment may be played from the target video segment, or the target video segment may be played from the progress of the target video segment corresponding to the termination position.
For example, when the exemplary embodiment of the present disclosure performs a sliding operation on a target video during a triggering operation on the target video, a position where a sliding action ends, that is, a termination position, may be used, and then based on a correspondence relationship between the termination position falling into an area and a video clip, the target video clip may be determined.
In practical application, the video segments can be sequenced according to the timestamps of the video segments, then the regions of the target video in the target direction are divided according to the time length proportion of the sequenced video segments, the regions corresponding to the video segments can be obtained, and the sequencing of the regions in the target direction is consistent with the sequencing of the video segments.
Fig. 11 is a diagram illustrating a trigger operation of a target video according to an exemplary embodiment of the present disclosure. As shown in fig. 11, with reference to table 2 and the related description thereof, the target video 1101 shown in the search result page 1100 is divided into four regions in the width direction, which are a first region I, a second region II, a third region III, and a fourth region IV, respectively. In the manner shown in fig. 11, sliding the target video from point a to point B, it can be found that point B falls into the second region II. And the second region II corresponds to the 2# clip, so the determined target video clip is the 2# clip.
Step 1002: and displaying key information corresponding to the target video clip and/or video frames (hereinafter referred to as key frames) in the target video clip, wherein the video frames can represent the corresponding key information most. At this time, since the key information corresponding to the target video clip is displayed, in step 1001, the key information corresponding to each video clip may be hidden in response to a trigger operation on the target video, and then the key information corresponding to the target video clip may be displayed when the key information of the target video clip needs to be checked, so as to reduce unnecessary image rendering, and further reduce the hardware requirement on the terminal.
When the key frames in the target video clip are displayed, the key information in one or two adjacent video clips of the target video clip can be displayed, and the key information of the adjacent video clips can be displayed in a weakening way relative to the key information of the target video clip, so that the influence of the key information and the key frames of the non-target video clip is reduced. As can be seen from the key information content of the target video shown in table 1, if the target video clip is the 2# clip, the 1# key information and the 3# key information can be displayed in a weakened manner, if the target video clip is the 1# clip, the 2# key information can be displayed in a weakened manner, and if the target video clip is the 4# clip, the 3# key information can be displayed in a weakened manner.
Therefore, when the user browses the key information and the key frames of the video clip, the key information which is already ahead can be traced back, the key information which is not played is foreseen, the purpose of providing comprehensive information for the user is achieved, and the user can conveniently compare, sort and systematize the information.
When the video frame which can represent the corresponding key information most in the target video clip is displayed, the key frame in the target video clip can be stored. The saving operation may be a saving operation of the terminal in response to the user, or may be an automatic downloading operation of the terminal.
In practical applications, the author of the target video may label the key frames to indicate whether the storage authority of the key frames is for all users or part of users, or only view the right that the key frames are not stored. Meanwhile, the creator of the target video may also provide a manner for saving the key frame, such as saving a screenshot, or long-pressing the key frame, calling a saving control, saving the key frame using the saving control, and the like, but is not limited thereto.
In a scene, a user needs to search for a picture meeting requirements through searching information, and by adding a key frame in a video clip, the user can retrieve not only a picture of a video cover but also an image of video content in the searching process, so that the recall rate is increased, the sequencing signal is strengthened, and the use experience of the user is improved. For example: for some classified videos, the video cover generally does not show the content of the details in the videos, so that the retrieval accuracy and comprehensiveness can be ensured by adding key frames to video clips.
In another scenario, a user needs to search for information that may be contained in the video content but not contained in the video file such as title, abstract, etc., and by adding key frames to the video segment, the video content can be more fully understood, thereby increasing the recall rate and enhancing the ranking signal. For example: for some methods such as strategy and cooking videos, the key frames added to the video clips can reflect key information in the methods, and the key frame image information amount is large, so that the retrieval accuracy and comprehensiveness can be ensured by adding the key frames to the video clips.
For example, when the target video has a progress bar, the target video clip may be determined according to a positioning position of a progress positioning operation in response to the progress positioning operation with respect to the progress bar. It should be understood that the exemplary embodiments of the present disclosure may determine the target video clip in a non-full screen state, and may also determine the target video clip in a full screen state. When the target video clip is determined in a full-screen state, the target video clip can be displayed in a full-screen state, then the target video clip is determined according to the positioning position of the progress positioning operation in response to the progress positioning operation aiming at the progress bar, and further the key information and/or the key frame of the target video clip is displayed.
The progress positioning operation for the progress bar according to the exemplary embodiment of the present disclosure may include a progress selection operation for the progress bar, and may also include a progress drag operation for the progress bar. Regardless of the progress selection operation or the progress drag operation, the determination of the positioning position of the progress positioning operation may determine the target video section with reference to the process of determining the target video section of fig. 11. In addition, the progress bar of the exemplary embodiment of the present disclosure has three states, which are a hidden state, a dragged state, and an activated state, respectively. In the hidden state and the dragged state, the thickness of the progress bar can not be changed, and in the activated state, the thickness of the progress bar can be changed.
In order to facilitate identification of correspondence between different positions of the progress bar and the video segments, the resource information of the exemplary embodiment of the present disclosure further includes time nodes corresponding to the video segments, and the time nodes corresponding to the respective video segments are displayed. These time nodes may be presented on a progress bar to facilitate progress selection or progress dragging operations on the progress bar. It should be understood that, the number of time nodes of the exemplary embodiments of the present disclosure may have a correspondence relationship with the video segments, and the correspondence relationship here is not a correspondence relationship in number, but a correspondence relationship in node attribute. For example: for short videos, the number of video nodes does not exceed 10.
For example, in combination with fig. 11 and table 2, the progress bar of the target video has 3 nodes from left to right, which are respectively a first node 1102, a second node 1103 and a third node 1104, the target video 1101 may be divided into 4 sub-segments in the extension direction of the progress bar, which are respectively a first sub-segment, a second sub-segment, a third sub-segment and a fourth sub-segment, the first sub-segment may correspond to a 1# segment, the second sub-segment may correspond to a 2# segment, the third sub-segment may correspond to a 3# segment, and the fourth sub-segment may correspond to a 4# sub-segment. The first node 1102 may serve as an end point of the 1# segment and as a start point of the 2# segment. The second node 1103 can be used as an end point of the 2# segment and a start point of the 3# segment. The third node 1104 can be the end point of the 3# segment and can be the start point of the 4# segment. It can be seen that 3 nodes can be used as the end point of one video segment and also as the start point of another video segment.
When the progress positioning operation for the progress bar includes a selection operation of the progress bar, the user can directly select the target video clip on the progress bar by clicking and the like. When the progress positioning operation for the progress bar includes a progress dragging operation performed on the progress bar, the user can drag a slider control on the progress bar, and the target video clip is determined according to the position where the slider control finally stops.
Fig. 12A is a schematic diagram illustrating a trigger operation of a target video on a search result page according to an exemplary embodiment of the present disclosure. As shown in fig. 12A, in the case of referring to table 2 and the related description thereof, as shown in fig. 12A, when the target video 1201 is presented on the search result page 1200 in the manner shown in fig. 12A, the position of the second sub-segment 1203 corresponding to the 2# segment in the progress bar 1202 may be clicked to determine that the target video segment is the 2# segment.
Fig. 12B is a schematic diagram illustrating a trigger operation of a target video on a search result page according to an exemplary embodiment of the present disclosure. With reference to table 2 and its related description, as shown in fig. 12B, in the search result page 1200, the slider control on the progress bar 1202 of the target video 1201 may be slid along the direction of the dashed arrow shown in fig. 12B, and the slider control finally stops at the P point position of the second sub-segment shown in fig. 12C, so that the target video segment is determined to be the # 2 segment.
Fig. 12C is a schematic diagram illustrating a trigger result of a target video in a search result page according to an exemplary embodiment of the disclosure. As shown in fig. 12C, in the search result page 1200, when it is determined that the target video 1201 includes the target video clip of the # 2 clip, not only the key information 1203 and the key frame 1204 of the target video clip may be displayed, but also the # 2 clip may be played, the playing progress 1205 of the target video clip may be displayed, the key information 1206 of the # 1 clip may be displayed in a faded manner, and the key information 1207 of the # 3 clip may be displayed.
Fig. 12D is a schematic diagram illustrating another trigger result of the target video on the search result page according to the exemplary embodiment of the disclosure. Fig. 12D is different from fig. 12C in that the key information in fig. 12C is still shown below the search result page 1200 of the target video 1201, whereas in the search result page 1200 of fig. 12D, the key information below the progress bar has disappeared in the search result, the key information 1203 of the target video clip, the key frame 1204, the play progress 1205 of the target time clip, the key information 1206 of the faded-display 1# clip, the key information 1207 of the 3# clip, and the like are retained. In addition, as can be seen from a comparison of fig. 12A, 12B, 12C and 12D, when there is no progress positioning operation for the progress bar, the progress bar is in an initial state, and the progress bar is thin, and when there is a progress positioning operation for the progress bar, the progress bar is in an activated state, and the progress bar is thick.
It should be noted that, in the case of non-full-screen display of the target video, the key information may be located below the target video in the search result page as shown in fig. 12A and 12B, or may be located on the target video in a floating layer. The target video of the exemplary embodiment of the present disclosure may be located below the progress bar of the target video in a full-screen display condition, and may also be located above the progress bar of the target video in a floating layer manner. However, whether the target video is full screen or not and whether the target video is displayed on the target video in a floating layer manner or not, the operation can be triggered in two ways shown in fig. 12A and 12B, which will not be described in detail herein. Whether the display area of the original key information is reserved in the triggering result of dragging the progress bar or not can be set according to actual conditions.
For example: when the target video is displayed in a non-full screen mode, the display area of the original key information is located on the target video in a floating layer mode, the display area of the original key information may not be reserved as a trigger result, and the interface schematic diagram may refer to fig. 12D.
Another example is: when the target video is in a full-screen display condition, the display area of the original key information is located on the target video in a floating layer manner, the display area of the original key information may not be reserved as a trigger result, and the interface schematic diagram of the full-screen interface 1300 may refer to fig. 13.
For another example: when the target video is displayed in a full screen, the display area of the original key information is located below the progress bar, the display area of the original key information may be reserved as a trigger result, and reference may be made to fig. 14 for an interface schematic diagram of a full screen interface 1400. Certainly, the display area of the original key information may not be reserved in the trigger result, and the interface schematic diagram of the full-screen interface may refer to fig. 13.
Illustratively, in the process of dragging the slider control of the target video, if a certain time (for example, 1s) can reach the end point node of the currently playing video segment, in order to accurately locate the start position of the next video segment, the dragging control on the progress bar can be automatically adsorbed to the end point node, so that the user can accurately obtain the information of the start time of the next video segment, avoid repositioning the start position of the next video segment, and thus improve the comprehensiveness of information acquisition. In order to remind the user to watch the video clip which is started to be played, the terminal can send out a prompt message, wherein the prompt message can be a voice prompt message, a prompt box prompt message or a vibration prompt message. It should be understood that for voice prompts and vibration prompts, the terminal should satisfy the hardware required for voice prompts and vibration prompts. For example: the terminal can be provided with a vibration switch for prompting the vibration.
In an alternative manner, when the search result page shows a teletext entry associated with the target video, the method of the exemplary embodiment of the present disclosure may further include: and responding to the trigger operation aiming at the teletext entrance, and displaying the teletext description information of the target video.
FIG. 15 illustrates an interface diagram of a search results page that presents a teletext entry according to an exemplary embodiment of the disclosure. As shown in fig. 15, in a target video 1501 shown in a search result page 1500, the target video 1501 has a teletext entry 1502 provided in a floating layer. When the user clicks on the teletext portal 1502, the teletext descriptor 1600 shown in fig. 16 can be displayed.
In practical applications, as shown in fig. 15, a text entry 1502 may be positioned above a target video 1501 in a floating layer, and when a user clicks the text entry 1501, the text description information may be displayed on a separate page. The teletext information comprises teletext information for a plurality of video segments, the teletext information for each video segment comprising a video frame (e.g. a key frame) most representative of the corresponding key information and text describing the corresponding video segment.
The aforementioned key information and key frames of the exemplary embodiments of the present disclosure, and the key frames and the texts describing the corresponding target videos included in the teletext description information may all be from the recognition result of the recognition model on the video pictures of the videos and/or the text description contents of the videos. The key frames contained in the textual description may also be downloaded when the user is saving the key frames. At this time, the exemplary embodiment of the present disclosure may store not only the video frame corresponding to the target video clip, but also the video frames corresponding to other video clips, and it can be seen that the exemplary embodiment of the present disclosure may store the video frame that can represent the most corresponding key information in at least one video clip.
In connection with the example of table 2, fig. 16 shows a schematic diagram of a presentation interface of the teletext description information according to the exemplary embodiment of the disclosure. As shown in fig. 16, the cooking method of the barbeque grill is divided into four steps, each step is provided with a key frame 1602 of a video clip corresponding to key information, while a core content 1601 described by the corresponding key information is used as a title. The user may secure key frames for one or more of the steps. The detailed introduction text 1603 under each heading comes from the recognition result of the recognition model on the video picture of the video and/or the text description content of the video.
In an alternative mode, when the publishing platform publishes various videos including the target video, the videos can be sliced first based on the segmentation chapter function, and then the videos are identified in a segmentation mode, so that powerful support is provided for video structuring. At this time, the videos may include a plurality of video clips, each of which is added with key information to describe core content of the corresponding video clip.
In practical applications, various videos of the exemplary embodiments of the present disclosure may also be videos with obvious structured video features, such as diet-related videos of exploration shops, menu courses, and the like in the cate, and videos related to travel strategies of hotel visitors and lodging visits, N-day strategies, N-station strategies, route strategies, and the like in the travel; and (4) collection recommendation in fashion, sectional and step-by-step method explanation, putting-on and putting-on presentation and the like. It should be understood that the videos of the exemplary embodiments of the present disclosure may also not exclude videos without obvious structures, such as journal, narrative, experiential videos, videos of tour guides and guild-talks in tourist categories, historical classics, personal experience record type vlog, travel photography beauty photos, and the like.
When the target video of the exemplary embodiment of the present disclosure has a video with obviously structured video features, the target video includes a plurality of video segments belonging to the same topic. For example: the contents of the video clips have a sequential logic sequence according to the sequential sequence of the timestamps, and the contents of the target video belong to videos of methods, such as a menu course, a route strategy, an N-day strategy, an N-station strategy segment and the like. Another example is: whether each video clip is in the sequence of the time stamps or not, the content of each video clip is parallel and belongs to the same theme, and the content of the target video belongs to classified introduction video content, such as store exploration, collection recommendation, wearing and taking presentation, hotel people and lodging exploration and the like.
As can be seen from the above, when the target video has a video with obvious structural video features, if the content of the target video belongs to a video in a method class, the target video can be segmented more conveniently by using the steps of the method as the segmentation basis, and if the target video includes different content that belongs to the same topic and is parallel, the target video can be segmented more conveniently by using the different content that is parallel as the segmentation basis.
After the video is segmented into a plurality of video segments by adopting the video slicing technology, the key information of each video can be marked in a manual mode and/or an identification model.
The first method comprises the following steps: and marking the key information of the video clip in a manual mode for standard. For example: the publishing platform may provide tools for the author to segment the video into chapters. Another example is: the publisher publishes the video through the interactive interface publishing platform displayed by the terminal, and after the video is audited, other people such as staff of the publishing platform can manually perform subsection and chaptering on the video.
The second method comprises the following steps: and segmenting the video by adopting the recognition model. For example: and after the video is published by the publisher through the publishing platform and is audited, the key information and the key frames corresponding to each video clip can be extracted from the video picture and/or the text description content through the identification model. Based on this, if the target video of the exemplary embodiment of the present disclosure employs the recognition model to extract the key information and the key frames, the key information is from the recognition result of the recognition model on the video picture of the video and/or the text description content of the video.
The recognition model of the exemplary embodiment of the present disclosure may include one or more combinations of an image recognition model and a natural language model. The image recognition model and the voice recognition model can be trained in advance, and then the trained image recognition model and the trained voice recognition model are used for intelligently extracting the key information corresponding to each video clip. The image recognition model may include an OCR recognition model and may also include other models such as ResNet, etc., and the speech recognition model may include an ASR model such as a hidden markov model, a gaussian mixture model, etc., but is not limited thereto.
Illustratively, when the videos are segmented and chaptered by adopting the artificial identification key information and the artificial intelligence technology, the videos of the artificially marked key information can be used as training data to train the recognition model, so that the recognition model can automatically add the key information to the video clips. Tests prove that after the artificial intelligence technology is introduced, the efficiency of video annotation key information can be obviously submitted, and more than half of labor cost can be saved.
For example, the text description of the exemplary embodiment of the present disclosure may include a title, a summary, and other video documents published by the creator. At this time, the recognition model can be combined with the video file on the basis of understanding the video content, so that the key information of each video clip can be determined more accurately and comprehensively.
For example, for some videos, their video copy may reflect the video content in a full scale, in which case the video has no significant gain. For example: it is common to belong to the categorical introduction class of video. For example: when the video file is 'knocking the blackboard, the points like sisters who like nail care are necessarily paid attention to cheering- # nail care # health', the recognition model understands that the core content of the video is the attention about nail care, and therefore the recognition model is assisted to determine the key information of each video fragment.
For example, the text description content of the exemplary embodiment of the present disclosure may further include a high-quality comment meeting the requirement, and when the comment is a high-quality comment, the identification model may add key information to the video clip in combination with the high-quality comment auxiliary identification model on the basis of understanding the video content.
For example, for some videos, video files of the videos cannot comprehensively reflect video contents, and in this case, the videos have obvious gain information, so that under the condition that manual labeling and identification model labeling reach a certain user coverage, after the videos are subjected to cold start, various comments appear below the videos, and high-quality comments with strong summarization capability can be used as part of input information of the identification model, so that the identification model can be combined with the high-quality comments on the basis of understanding the video contents, and thus the key information of each video segment can be determined more accurately and comprehensively. For example: some classes of review zones that are frequently present in the review zone represent reviews or summarize reviews.
The high-quality comments of the exemplary embodiments of the present disclosure can be measured by one or more combinations of the number of praise, the number of forwarding times, the number of sub-comments, the length of comment text, and the like.
Taking the number of praise as an example, when the video has comments of which the number of praise exceeds the preset number, the comments can be marked as high-quality comments. Taking the forwarding times as an example, when the forwarding times of the comments exceed the preset times, the comments can be marked as high-quality comments, taking the number of the sub-comments as an example, and when the number of the sub-comments of a certain comment reaches the preset number, the comment can be marked as a high-quality comment. Taking the length of the comment text as an example, when the length of the comment text exceeds a preset length, the comment can be marked as a high-quality comment. It should be understood that the parameters such as the preset times, the forwarding times, the preset number, the preset length, etc. can be adjusted according to the actual situation. For example, on the premise that the comment is marked as a high-quality comment for 10 times, since the number of comment points below the video shown in fig. 3 is 12, the comment is marked as a high-quality comment and can be used for identifying the model. It should be appreciated that more high-quality reviews may appear over time, and thus, as the high-quality reviews change over time as part of the recognition model input, the key information for the video clip may also dynamically change, making the key information for the video clip more and more accurate and comprehensive.
Therefore, after the creator publishes the video, the high-quality comments meeting the requirements are selected from the video comments, the recognition model is independently used for recognition, or the key information is extracted by combining the video file and the video published by the creator, so that other users can participate in the labeling of the key information, the mode is called as UGC mode, the participation of other users can be increased, the exposure of other users is improved, and the video coverage is enlarged.
After at least one of the above manners is adopted to label the key information of the video segment, the video can be divided into a plurality of video segments, and each video segment also corresponds to the key information describing the core content of the video segment, so that when a search engine searches for the content related to the search information from a large number of videos, the search engine can search for the videos by retrieving the video file, and can also use the reference information as a reference, thereby improving the recall rate of the video information. And when the search engine utilizes the sorting function to sort the search results, the reference information is also considered in the sorting condition, so that the search engine can reasonably organize the information, the optimized sorting is realized, and the sorting accuracy and effectiveness of the search results are increased. At this time, the video with higher matching degree with the search information is arranged at the position near the search result, and when the user browses the search result, the user can quickly find the target video from a large number of search results.
In summary, in one or more technical solutions provided in the embodiments of the present disclosure, the resource information of the target video includes key information describing core content of each video clip, and therefore, when the target video and the resource information are displayed on the search result page, the video clip of interest can be quickly determined based on the core content of the video clip described by the key information, then the target key information is determined based on the video clip of interest, and in response to a trigger operation on the target key information, the video clip corresponding to the target key information is played on the search result page, so as to achieve a purpose of quickly browsing the video clip corresponding to the target key information. Meanwhile, the display length proportion of each key information is matched with the duration proportion of each video clip, so that the display length of each key information can relatively reflect the duration of the corresponding video clip, and when the target key information is determined, the target key information can be determined through the core content of the video clip described by the key information, and whether the duration of the corresponding video clip meets the requirement of a user for fast browsing the video clip can be determined through the display length of the target key information. When the duration of the video clip corresponding to the target key information meets the requirement of a user for quickly browsing the video clip, the video clip corresponding to the target key information can meet the requirement of the user for the content of the video clip, and can ensure that the user quickly finishes browsing the video content within a limited time range, thereby achieving the dual guarantee effects of browsing speed and browsing quality.
Therefore, the method of the exemplary embodiment of the disclosure can improve the video VV, the video watching completion degree, the dwell time, the query bit ratio and the like in a video searching scene, so that the user experience is relatively high. Moreover, the segmentation chapter function combines manual and artificial intelligence, and a UGC mode can be adopted to improve the user participation rate, so that the use penetration rate of the segmentation chapter function of the exemplary embodiment of the present disclosure is high.
The foregoing mainly introduces the solution provided by the embodiment of the present disclosure from the perspective of the terminal. It is understood that the terminal includes corresponding hardware structures and/or software modules for performing the respective functions in order to implement the above-described functions. Those of skill in the art will readily appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as hardware or combinations of hardware and computer software. Whether a function is performed as hardware or computer software drives hardware depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present disclosure.
The embodiment of the present disclosure may perform division of the functional units on the terminal according to the method example, for example, each functional module may be divided corresponding to each function, or two or more functions may be integrated into one processing module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode. It should be noted that, the division of the modules in the embodiments of the present disclosure is illustrative, and is only one division of logic functions, and there may be another division in actual implementation.
In the case of adopting a method of dividing each function module corresponding to each function, the exemplary embodiments of the present disclosure provide a search result presentation apparatus, which may be a terminal or a chip applied to the terminal. Fig. 17 shows a schematic block diagram of a search result presentation apparatus according to an exemplary embodiment of the present disclosure. As shown in fig. 17, the search result presentation apparatus 1700 includes:
a receiving module 1701, configured to receive search information, and obtain a target video matched with the search information and resource information of the target video; the target video comprises a plurality of video clips, and the resource information comprises key information describing core content of each video clip; the display length proportion of each piece of key information is correspondingly matched with the duration proportion of each video clip;
a displaying module 1702, configured to display the target video and the resource information on a search result page, and in response to a trigger operation on the target key information, play a video clip corresponding to the target key information on the search result page.
In one possible implementation manner, the display areas of the key information are distributed along the target direction, and the key information is slidable on the search result page.
In a possible implementation manner, the size of the display area of each piece of key information in the target direction is matched with the size of the corresponding video clip in the target direction, and each piece of key information is displayed on the target video in a floating layer manner.
In one possible implementation, the key information is displayed in an area of the search result page outside the target video.
In a possible implementation manner, the displaying module 1702 is further configured to determine, in response to the triggering operation on the target video, the target video segment according to the termination position of the triggering operation, and display the key information corresponding to the target video segment and/or a video frame in the target video segment that can most represent the corresponding key information.
In a possible implementation manner, the target video has a progress bar, and the presentation module 1702 is configured to present the target video in a full screen manner, respond to a progress positioning operation on the progress bar, and determine a target video segment according to a positioning position of the progress positioning operation.
In a possible implementation manner, the resource information further includes time nodes corresponding to the video segments, and the showing module 1702 is further configured to show the time nodes corresponding to the video segments.
In a possible implementation manner, the search result presentation apparatus 1700 of the exemplary embodiment of the present disclosure further includes a hiding module 1703, configured to hide key information corresponding to each video segment in response to a trigger operation on a target video.
In one possible implementation, the search information contains information semantically related to the operation steps and/or information semantically related to the classification introduction.
In one possible implementation, the search result page is displayed with a text entry associated with the target video, and the display module 1702 is further configured to display the text description information of the target video in response to a triggering operation for the text entry.
In one possible implementation, the teletext description information comprises teletext information for a plurality of video segments, the teletext information for each video segment comprising a video frame which best characterizes the respective key information and text describing the respective video segment.
In one possible implementation, the key information and/or the video frames that best characterize the respective key information are derived from the recognition result of the recognition model on the video pictures of the video and/or the textual description of the video.
In a possible implementation manner, the search result presentation apparatus 1700 of the exemplary embodiment of the present disclosure further includes a saving module 1704, configured to save a video frame, which is most representative of the corresponding key information, in at least one video clip.
Fig. 18 shows a schematic block diagram of a chip according to an example embodiment of the present disclosure. As shown in fig. 18, the chip 1800 includes one or more (including two) processors 1801 and a communication interface 1802. The communication interface 1802 may support a server to perform a data transceiving step in the search result presentation method, and the processor 1801 may support the server to perform a data processing step in the search result presentation method.
Optionally, as shown in fig. 18, the chip 1800 further includes a memory 1803, and the memory 1803 may include a read-only memory and a random access memory, and provides the processor with operating instructions and data. The portion of memory may also include non-volatile random access memory (NVRAM).
In some embodiments, as shown in fig. 18, the processor 1801 performs the corresponding operation by calling an operation instruction stored in the memory (the operation instruction may be stored in an operating system). The processor 1801 controls processing operations of any of the terminal devices, and may also be referred to as a Central Processing Unit (CPU). The memory 1803 may include read-only memory and random access memory, and provides instructions and data to the processor 1801. A portion of the memory 1803 may also include NVRAM. For example, in applications where the memory, communication interface, and memory are coupled together by a bus system that may include a power bus, a control bus, a status signal bus, etc., in addition to a data bus. For clarity of illustration, however, the various busses are labeled in fig. 18 as the bus system 1804.
The method disclosed by the embodiment of the present disclosure can be applied to a processor, or can be implemented by the processor. The processor may be an integrated circuit chip having signal processing capabilities. In implementation, the steps of the above method may be performed by integrated logic circuits of hardware in a processor or by instructions in the form of software. The processor may be a general purpose processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a field-programmable gate array (FPGA) or other programmable logic device, discrete gate or transistor logic device, or discrete hardware components. The various methods, steps, and logic blocks disclosed in the embodiments of the present disclosure may be implemented or performed. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like. The steps of the method disclosed in connection with the embodiments of the present disclosure may be directly implemented by a hardware decoding processor, or implemented by a combination of hardware and software modules in the decoding processor. The software modules may be located in ram, flash, rom, prom, or eprom, registers, etc. as is well known in the art. The storage medium is located in a memory, and a processor reads information in the memory and completes the steps of the method in combination with hardware of the processor.
An exemplary embodiment of the present disclosure also provides an electronic device including: at least one processor; and a memory communicatively coupled to the at least one processor. The memory stores a computer program executable by the at least one processor, the computer program, when executed by the at least one processor, is for causing the electronic device to perform a method according to an embodiment of the disclosure.
The disclosed exemplary embodiments also provide a non-transitory computer readable storage medium storing a computer program, wherein the computer program, when executed by a processor of a computer, is adapted to cause the computer to perform a method according to an embodiment of the present disclosure.
The exemplary embodiments of the present disclosure also provide a computer program product comprising a computer program, wherein the computer program, when executed by a processor of a computer, is adapted to cause the computer to perform a method according to an embodiment of the present disclosure.
FIG. 19 shows a block diagram of an exemplary electronic device that can be used to implement embodiments of the present disclosure. As shown in fig. 19, a block diagram of a structure of an electronic device 1900, which is a server or a terminal of the present disclosure, is an example of a hardware device that can be applied to aspects of the present disclosure. Electronic device is intended to represent various forms of digital electronic computer devices, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other suitable computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular phones, smart phones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be examples only, and are not meant to limit implementations of the disclosure described and/or claimed herein.
As shown in fig. 19, the electronic apparatus 1900 includes a computing unit 1901, which can perform various appropriate actions and processes in accordance with a computer program stored in a Read Only Memory (ROM) 1902 or a computer program loaded from a storage unit 1908 into a Random Access Memory (RAM) 1903. In the RAM 1903, various programs and data necessary for the operation of the electronic apparatus 1900 can also be stored. The computing unit 1901, ROM 1902, and RAM 1903 are connected to each other via a bus 1904. An Input/Output (I/O) interface 1905 is also connected to the bus 1904.
A number of components in electronic device 1900 are connected to I/O interface 1905, including: an input unit 1906, an output unit 1907, a storage unit 1908, and a communication unit 1909. The input unit 1906 may be any type of device capable of inputting information to the electronic device 1900, and the input unit 1906 may receive input numeric or character information and generate key signal inputs related to user settings and/or function control of the electronic device. Output unit 1907 can be any type of device capable of presenting information and can include, but is not limited to, a display, speakers, a video/audio output terminal, a vibrator, and/or a printer. Storage unit 1904 may include, but is not limited to, a magnetic disk, an optical disk. The communication unit 1909 allows the electronic device 1900 to exchange information/data with other devices via a computer network, such as the internet, and/or various telecommunications networks, and may include, but is not limited to, modems, network cards, infrared communication devices, wireless communication transceivers and/or chipsets, such as bluetooth devices, WiFi devices, WiMax devices, cellular communication devices, and/or the like.
As shown in FIG. 19, computing unit 1901 may be a variety of general and/or special purpose processing components with processing and computing capabilities. Some examples of the computing Unit 1901 include, but are not limited to, a Central Processing Unit (CPU), a Graphics Processing Unit (GPU), various dedicated Artificial Intelligence (AI) computing chips, various computing units running machine learning model algorithms, a Digital Signal Processor (DSP), and any suitable processor, controller, microcontroller, and the like. The calculation unit 1901 performs the respective methods and processes described above. For example, in some embodiments, the methods of the exemplary embodiments of the present disclosure may be implemented as a computer software program tangibly embodied in a machine-readable medium, such as the storage unit 1908. In some embodiments, part or all of the computer program can be loaded and/or installed onto the electronic device 1900 via the ROM 1902 and/or the communication unit 1909. In some embodiments, the computing unit 1901 may be configured by any other suitable means (e.g., by means of firmware) to perform the methods of the exemplary embodiments of the present disclosure.
In the above embodiments, all or part of the implementation may be realized by software, hardware, firmware, or any combination thereof. When implemented in software, it may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer programs or instructions. When the computer program or instructions are loaded and executed on a computer, the procedures or functions described in the exemplary embodiments of the present disclosure are performed in whole or in part. The computer may be a general purpose computer, a special purpose computer, a computer network, a terminal, a user device, or other programmable apparatus. The computer program or instructions may be stored in a computer readable storage medium or transmitted from one computer readable storage medium to another computer readable storage medium, for example, the computer program or instructions may be transmitted from one website, computer, server or data center to another website, computer, server or data center by wire or wirelessly. The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device such as a server, data center, etc. that integrates one or more available media. The usable medium may be a magnetic medium, such as a floppy disk, a hard disk, a magnetic tape; or optical media such as Digital Video Disks (DVDs); it may also be a semiconductor medium, such as a Solid State Drive (SSD).
While the present disclosure has been described herein with reference to various embodiments, other variations to the disclosed embodiments can be understood and effected by those skilled in the art in practicing the claimed disclosure, from a review of the drawings, the disclosure, and the appended claims. In the claims, the word "comprising" does not exclude other elements or steps, and the word "a" or "an" does not exclude a plurality. A single processor or other unit may fulfill the functions of several items recited in the claims. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage.
While the disclosure has been described in conjunction with specific features and embodiments thereof, it will be evident that various modifications and combinations can be made thereto without departing from the spirit and scope of the disclosure. Accordingly, the specification and drawings are merely illustrative of the present disclosure as defined in the appended claims and are intended to cover any and all modifications, variations, combinations, or equivalents within the scope of the disclosure. It will be apparent to those skilled in the art that various changes and modifications may be made to the present disclosure without departing from the spirit and scope of the disclosure. Thus, it is intended that the present disclosure also include such modifications and variations insofar as they come within the scope of the appended claims and their equivalents.

Claims (16)

1. A search result presentation method is characterized by comprising the following steps:
receiving search information, and acquiring a target video matched with the search information and resource information of the target video; the target video comprises a plurality of video clips, and the resource information comprises key information describing core content of each video clip; the display length proportion of each piece of key information is correspondingly matched with the duration proportion of each video clip;
displaying the target video and the resource information on a search result page;
and responding to the triggering operation of the target key information, and playing a video clip corresponding to the target key information on the search result page.
2. The method according to claim 1, wherein the display area of each piece of key information is distributed along a target direction, and each piece of key information is slidable on the search result page.
3. The method according to claim 2, wherein a size of a display area of each of the key information in the target direction matches a size of the corresponding video clip in the target direction, and each of the key information is displayed on the target video in a floating layer.
4. The method of claim 2, wherein each of the key information is presented in an area of the search results page outside of the target video.
5. The method of claim 1, further comprising:
responding to the trigger operation of the target video, and determining a target video segment according to the termination position of the trigger operation;
and displaying key information corresponding to the target video clip and/or video frames which can represent the corresponding key information most in the target video clip.
6. The method according to claim 5, wherein the target video has a progress bar, and the determining a target video segment according to the termination position of a trigger operation in response to the trigger operation on the target video comprises:
displaying the target video in a full screen mode, responding to progress positioning operation of the progress bar, and determining a target video clip according to the positioning position of the progress positioning operation.
7. The method according to any one of claims 1 to 6, wherein the resource information further includes a time node corresponding to a video clip, the method further comprising:
and displaying the time node corresponding to each video clip.
8. The method according to any one of claims 1 to 6, further comprising:
and hiding the key information corresponding to each video clip in response to the triggering operation of the target video.
9. The method according to any one of claims 1 to 6, characterized in that the search information contains information semantically related to the operation steps and/or information semantically related to the classification introduction.
10. The method according to any one of claims 1 to 6, wherein the search result page presents a teletext entry associated with the target video, the method further comprising:
and responding to the trigger operation aiming at the teletext entrance, and displaying the teletext description information of the target video.
11. The method of claim 10, wherein the teletext description information comprises teletext information for a plurality of video segments, the teletext information for each video segment comprising a video frame that best characterizes the corresponding key information and text describing the corresponding video segment.
12. The method according to claim 11, wherein the key information and/or the video frames that most characterize the respective key information are derived from recognition results of a recognition model on video pictures of the video and/or textual description of the video.
13. The method of claim 11, further comprising:
and storing the video frame which can most represent the corresponding key information in at least one video clip.
14. A search result presentation apparatus, comprising:
the receiving module is used for receiving search information, and acquiring a target video matched with the search information and resource information of the target video; the target video comprises a plurality of video clips, and the resource information comprises key information describing core content of each video clip; the display length proportion of each piece of key information is correspondingly matched with the duration proportion of each video clip;
and the display module is used for displaying the target video and the resource information on a search result page, responding to the trigger operation of the target key information, and playing a video clip corresponding to the target key information on the search result page.
15. An electronic device, comprising:
a processor; and (c) a second step of,
a memory for storing a program;
wherein the program comprises instructions which, when executed by the processor, cause the processor to carry out the method according to any one of claims 1 to 13.
16. A non-transitory computer readable storage medium storing computer instructions for causing a computer to perform the method according to any one of claims 1 to 13.
CN202210622097.4A 2022-06-01 2022-06-01 Search result display method, device, equipment and non-transitory computer storage medium Pending CN115037959A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210622097.4A CN115037959A (en) 2022-06-01 2022-06-01 Search result display method, device, equipment and non-transitory computer storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210622097.4A CN115037959A (en) 2022-06-01 2022-06-01 Search result display method, device, equipment and non-transitory computer storage medium

Publications (1)

Publication Number Publication Date
CN115037959A true CN115037959A (en) 2022-09-09

Family

ID=83123396

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210622097.4A Pending CN115037959A (en) 2022-06-01 2022-06-01 Search result display method, device, equipment and non-transitory computer storage medium

Country Status (1)

Country Link
CN (1) CN115037959A (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106658199A (en) * 2016-12-28 2017-05-10 网易传媒科技(北京)有限公司 Video content display method and apparatus
CN113111220A (en) * 2021-03-26 2021-07-13 北京达佳互联信息技术有限公司 Video processing method, device, equipment, server and storage medium
CN113709571A (en) * 2021-07-30 2021-11-26 北京搜狗科技发展有限公司 Video display method and device, electronic equipment and readable storage medium
CN114329254A (en) * 2021-12-31 2022-04-12 北京字节跳动网络技术有限公司 Search result display method and device, computer equipment and storage medium
US20220129679A1 (en) * 2020-10-27 2022-04-28 International Business Machines Corporation Summarizing Videos Via Side Information

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106658199A (en) * 2016-12-28 2017-05-10 网易传媒科技(北京)有限公司 Video content display method and apparatus
US20220129679A1 (en) * 2020-10-27 2022-04-28 International Business Machines Corporation Summarizing Videos Via Side Information
CN113111220A (en) * 2021-03-26 2021-07-13 北京达佳互联信息技术有限公司 Video processing method, device, equipment, server and storage medium
CN113709571A (en) * 2021-07-30 2021-11-26 北京搜狗科技发展有限公司 Video display method and device, electronic equipment and readable storage medium
CN114329254A (en) * 2021-12-31 2022-04-12 北京字节跳动网络技术有限公司 Search result display method and device, computer equipment and storage medium

Similar Documents

Publication Publication Date Title
CN112104899B (en) Information recommendation method and device in live broadcast, electronic equipment and storage medium
CN110134931B (en) Medium title generation method, medium title generation device, electronic equipment and readable medium
JP7356206B2 (en) Content recommendation and display
WO2019062044A1 (en) Method for interaction between electronic book and electronic book topic computing device, and storage medium
US20210097105A1 (en) Because you reacted to content
WO2016091044A1 (en) Method, apparatus, system, device and computer storage medium for recommending hot words
US20210099405A1 (en) Content item module arrangements
CN109791680A (en) Key frame of video on online social networks is shown
CN111638832A (en) Information display method, device, system, electronic equipment and storage medium
CN107241622A (en) video location processing method, terminal device and cloud server
CN106874467A (en) Method and apparatus for providing Search Results
CN109690529A (en) Document is compiled into timeline by event
CN107656787B (en) Method for generating topics based on electronic book, computing device and computer storage medium
CN103988202A (en) Image attractiveness based indexing and searching
CN111818370B (en) Information recommendation method and device, electronic equipment and computer-readable storage medium
CN109308334B (en) Information recommendation method and device and search engine system
US20120177297A1 (en) Image Analysis System and Method Using Image Recognition and Text Search
US11477143B2 (en) Trending content view count
CN113886707B (en) Encyclopedia information determining method, encyclopedia information displaying method, encyclopedia information determining device, encyclopedia information displaying device, encyclopedia information determining equipment and encyclopedia information medium
CN107622074A (en) A kind of data processing method, device and computing device
CN114219586A (en) Shopping recommendation method, device, equipment and storage medium based on video
CN107515870A (en) A kind of searching method and device, a kind of device for being used to search for
CN109189899A (en) Content interest acquisition and content-data recommended method, device, equipment/terminal/server and storage medium
CN113407775A (en) Video searching method and device and electronic equipment
CN110827063A (en) Multi-strategy fused commodity recommendation method, device, terminal and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination