WO2022000983A1 - 视频处理方法、装置、电子设备和存储介质 - Google Patents

视频处理方法、装置、电子设备和存储介质 Download PDF

Info

Publication number
WO2022000983A1
WO2022000983A1 PCT/CN2020/133172 CN2020133172W WO2022000983A1 WO 2022000983 A1 WO2022000983 A1 WO 2022000983A1 CN 2020133172 W CN2020133172 W CN 2020133172W WO 2022000983 A1 WO2022000983 A1 WO 2022000983A1
Authority
WO
WIPO (PCT)
Prior art keywords
vibration
video file
target
video
time point
Prior art date
Application number
PCT/CN2020/133172
Other languages
English (en)
French (fr)
Inventor
李子朝
宋堃
Original Assignee
百度在线网络技术(北京)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 百度在线网络技术(北京)有限公司 filed Critical 百度在线网络技术(北京)有限公司
Priority to JP2021553359A priority Critical patent/JP2022541968A/ja
Priority to EP20920756.2A priority patent/EP3955585A4/en
Priority to KR1020217027630A priority patent/KR20210114059A/ko
Priority to US17/461,016 priority patent/US20210392394A1/en
Publication of WO2022000983A1 publication Critical patent/WO2022000983A1/zh

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8455Structuring of content, e.g. decomposing content into time segments involving pointers to the content, e.g. pointers to the I-frames of the video stream

Definitions

  • the present application relates to the field of data processing, in particular to the field of artificial intelligence and the field of video production and editing, and in particular to a video processing method, apparatus, electronic device and storage medium.
  • the present application provides a video processing method, apparatus, electronic device and storage medium.
  • a video processing method comprising:
  • the target video file is sent to the target server, so that the user can trigger the sensor of the terminal device to vibrate in response to the vibration event during the process of playing the target video file on the terminal device.
  • another video processing method comprising:
  • vibration event includes: time point and corresponding vibration parameter
  • the target video file is played, and when the time point is reached, the sensor of the terminal device is triggered to vibrate according to the vibration parameter corresponding to the time point in response to the vibration event.
  • a video processing apparatus comprising:
  • Add module for editing original video file by video editor, in described original video file, add vibration event and generate target video file, wherein, described vibration event comprises: time point and corresponding vibration parameter;
  • the sending module is configured to send the target video file to the target server, so that the user can trigger the sensor of the terminal device to vibrate in response to the vibration event when the terminal device plays the target video file.
  • another video processing apparatus comprising:
  • an acquisition module for acquiring a target video file containing a vibration event sent by the server, wherein the vibration event includes: a time point and a corresponding vibration parameter;
  • a playing module for playing the target video file
  • a vibration response module configured to respond to the vibration event when the target video file is played to the time point, and trigger the sensor of the terminal device to vibrate according to the vibration parameter corresponding to the time point.
  • an electronic device comprising:
  • the memory stores instructions executable by the at least one processor, the instructions being executed by the at least one processor to enable the at least one processor to perform the video processing method of the first aspect above.
  • a non-transitory computer-readable storage medium storing computer instructions, wherein the computer instructions are used to cause the computer to execute the video processing method described in the first aspect.
  • the original video file can be edited by a video editor, a vibration event is added to the original video file to generate a target video file, and the target video file is sent to the target server, so that the user can play the target video file on the terminal device.
  • the sensor of the terminal device is triggered to vibrate in response to the vibration event, which allows the viewer to use the client to watch the video more immersive, and greatly improves the user's experience of watching the video.
  • FIG. 1 is a flowchart of a video processing method according to an embodiment of the present application.
  • FIG. 2 is a flowchart of a video processing method according to another embodiment of the present application.
  • FIG. 3 is a flowchart of a video processing method according to yet another embodiment of the present application.
  • FIG. 4 is a flowchart of a video processing method according to another embodiment of the present application.
  • FIG. 5 is a structural block diagram of a video processing apparatus according to an embodiment of the present application.
  • FIG. 6 is a structural block diagram of a video processing apparatus according to another embodiment of the present application.
  • FIG. 7 is a structural block diagram of a video processing apparatus according to still another embodiment of the present application.
  • FIG. 8 is a block diagram of an electronic device for implementing the video processing method according to the embodiment of the present application.
  • FIG. 1 is a flowchart of a video processing method according to an embodiment of the present application. It should be noted that the video processing method in the embodiment of the present application can be applied to the video processing apparatus in the embodiment of the present application, and the video processing apparatus can be configured on an electronic device.
  • the electronic device may be a mobile terminal, and the mobile terminal may be a hardware device with various operating systems, such as a mobile phone, a tablet computer, and a personal digital assistant.
  • the video processing method may include:
  • Step 101 Edit the original video file through a video editor, and add a vibration event to the original video file to generate a target video file, where the vibration event includes a time point and corresponding vibration parameters.
  • the video processing method in the embodiment of the present application can be applied to an electronic device, and the electronic device can have a video editor, wherein the video editor can provide a user with a video acquisition interface.
  • the video editor can obtain the original video file to be processed provided by the user through this interface. That is, the original video file may be provided by the user, and a video editor is used to add a vibration event to the original video file, so that the generated video has a vibration effect.
  • a vibration event may be added to the original video file to generate a target video file.
  • the way of adding the vibration event may be that the video editor on the electronic device uses artificial intelligence technology to analyze the content of the original video file, obtains the special element information in the original video file, and then adds the corresponding vibration event based on the special element information , so that the processed video file has a vibration effect.
  • the way of adding the vibration event may also be manually added by the user, that is, the user adds the vibration event at a certain position or certain positions of the original video file according to requirements. It can be seen that the generated target video file includes both the video content of the original video file and the vibration event.
  • Step 102 Send the target video file to the target server, so that the user can trigger the sensor of the terminal device to vibrate in response to the vibration event during the process of playing the target video file on the terminal device.
  • the target video file may be uploaded to the target server.
  • the target video file may be uploaded to the target server.
  • another user downloads the target video file from the target server through the terminal device, and plays the target video file through the terminal device, when the time included in the vibration event is played, respond to the vibration event, according to the The vibration parameter corresponding to the time point triggers the sensor of the terminal device to vibrate.
  • the specific implementation process of adding a vibration event to an original video file to generate a target video file may include:
  • Step 201 parsing the original video file to extract key pictures.
  • extract key frames from the original video file and use an image recognition model to perform image recognition on the key frames in the original video file to obtain key pictures in the original video file.
  • an image recognition model For example, in a horror video, a sudden vibration can instantly increase the level of horror.
  • key frames can be extracted from the original video file, and image recognition of the key frames can be performed to extract the horror pictures existing in the original video file. .
  • vibrations are used to remind the viewer to concentrate.
  • the viewer needs to concentrate on listening to the lecture.
  • the voice or video display content in the original video file is used to determine which or which places in the original video file are key pictures, so that vibration events can be added at these key pictures to remind the viewer to vibrate when watching the video. Remind to focus on watching this video content.
  • Step 202 query a preset template database to obtain a target vibration template corresponding to the key picture, wherein the target vibration template includes vibration parameters.
  • a preset template database can be queried to obtain a target vibration template corresponding to the key picture.
  • the template database may be on an electronic device.
  • a template database may be integrated in the video editor, wherein the template database includes: an inherent vibration template provided by a device provider, and/or an automatic vibration template set by a user by selecting a control parameter.
  • the vibration template may include vibration parameters, which may include, but are not limited to, one or more of the number of vibrations, the time interval between each vibration, and the duration of each vibration.
  • each type of key picture may correspond to at least one vibration template, and the vibration parameters in each vibration template are different.
  • the template database set up to obtain the target vibration template corresponding to the key picture. For example, taking adding a vibration event to a horror video as an example, suppose the horror level is divided into 5 levels, and each level corresponds to a vibration template. In this way, after obtaining the key picture of the horror video, the level of the key picture can be determined. , and find out the vibration template corresponding to the level of the key picture from the template database according to the level of the key picture. In this way, different vibration events can be added to the original video file based on the difference in the level of key pictures, so as to realize adding corresponding vibration effects based on the video content, which can greatly improve the user's experience of watching videos.
  • vibration templates can also be obtained according to the key level of the explanation content, and then corresponding vibration effects can be added to the corresponding positions in the original video file, which can remind the viewer to concentrate.
  • Step 203 Acquire the time point corresponding to the key picture in the original video file.
  • the time point corresponding to the key picture in the original video file can be obtained.
  • the playback time point corresponding to the horror picture in the horror video can be obtained, and the playback time point can be used as Where the vibration event was added in the original video file.
  • Step 204 splicing the target vibration template to the position corresponding to the time point in the original video file to generate the target video file.
  • the target vibration template corresponding to the key picture can be fused with the original video file, for example, in the The triggering mark of the vibration event is added at the time point, and the original video file to which the triggering mark of the vibration event is added and the target vibration template are synthesized into one file, so as to obtain the target video file with the vibration effect.
  • the special element information in the original video file is obtained, and then vibration events are automatically added to the original video file based on the special element information, which can simplify user operation steps and improve Efficiency, realize the intelligence of video processing.
  • the user can select the location where the vibration event is added in the video based on his own needs.
  • the specific implementation process of adding vibration events to the file to generate the target video file may include:
  • Step 301 Obtain content points of interest in the original video file.
  • each frame of image in the original video file is extracted by a video editor and displayed to the user, and the user can choose which frame content in the original video file is interested in according to their own needs, so as to obtain the content interest in the original video file. point.
  • Step 302 calling the vibration setting interface, selecting vibration parameters according to the content interest points, and setting the vibration parameters in association with the time points corresponding to the content interest points in the original video file.
  • the vibration setting interface can be improved for the user, and a vibration parameter can be selected according to the content interest point, and after determining that the user selects the vibration parameter, the user-selected vibration parameter can be selected.
  • the vibration parameter is set in association with the time point corresponding to the content interest point in the original video file, so that the target video file with the vibration effect can be generated.
  • the user selects the location where the vibration event is added in the video based on their own needs, and then adds the vibration event at the location, providing the user with the editing function of adding a vibration effect to the video, so that the user has the ability to edit the vibration mode for the video. , which can meet the personalized needs of users and further improve the user experience.
  • the original video in this embodiment of the present application may be a short video or other video, and may also be a real-time video stream uploaded to the live server in real time during the live broadcast. That is to say, the video processing method of the embodiment of the present application can also be applied to a live video scene.
  • a video editor is embedded in the live broadcast software, so that when the host performs live broadcast through the live broadcast software, the host wants to push vibration to the user to complete the new
  • the anchor can add a vibration template at a certain moment of the live broadcast through the video editor in the live broadcast software, and the electronic device can obtain the real-time video stream of the anchor during the live broadcast, and record the moment in the corresponding real-time video stream.
  • the time point is determined as the adding position of the vibration event, and the vibration template is associated with the adding position, and the real-time video stream associated with the vibration template is uploaded to the target server.
  • the user watching the live broadcast obtains the real-time video stream of the live broadcast from the target server through the terminal device, and plays the real-time video stream of the live broadcast. Trigger the sensor of the terminal device to vibrate, so as to achieve the effect of pushing the vibration to the audience during the live broadcast.
  • the terminal device can pull the stream from the live broadcast server and play the real-time video stream corresponding to the live broadcast obtained from the server.
  • the sensor in the terminal device is triggered to vibrate, so as to realize the effect of pushing the vibration to the audience during the live broadcast.
  • an original video file can be edited by a video editor, a vibration event is added to the original video file to generate a target video file, and the target video file is sent to the target server, so that the user can play the target video file on the terminal device.
  • the sensor of the terminal device is triggered to vibrate in response to the vibration event.
  • Vibration can be added to better express the editor's own video. Vibration can be added to the live broadcast to better interact with the audience. For video viewers, a new watch has been added.
  • FIG. 4 is a flowchart of a video processing method according to another embodiment of the present application. As shown in Figure 4, the video processing method may include:
  • Step 401 Acquire a target video file containing a vibration event sent by the server, where the vibration event includes: a time point and a corresponding vibration parameter.
  • the viewer can obtain the target video file containing the vibration event from the server through the terminal device.
  • the target video file may be a short video file, and may also be a real-time video stream file of live broadcast.
  • the above-mentioned target video file may be a video file obtained by editing the original video file by using the video processing methods shown in FIG. 1 to FIG. 3 .
  • the target video file contains a vibration event
  • the vibration event includes: a time point and a corresponding vibration parameter.
  • the vibration parameters may include, but are not limited to, one or more of the number of vibrations, the time interval between each vibration, the duration of each vibration, and the like.
  • Step 402 play the target video file, and when the time point in the vibration event is played, in response to the vibration event, trigger the sensor of the terminal device to vibrate according to the vibration parameter corresponding to the time point.
  • the target video file when the target video file is obtained, the target video file can be played.
  • the original content of the video file is played while responding to the vibration event, according to the position.
  • the vibration parameter corresponding to the above vibration event triggers the sensor on the terminal device to vibrate, so as to display the vibration behavior corresponding to the content to the user while playing the original video content.
  • the sensor may be a vibration sensor.
  • a vibration event can be added to the video, so that when the video is played and played to the position where the vibration event is added, the original content of the video is played and the video is displayed to the user. Vibration behavior corresponding to this content.
  • a new editing method has been added. Vibration can be added to better express the editor's own video. Vibration can be added to the live broadcast to better interact with the audience. For video viewers, a new watch has been added. Feeling, when watching videos and live broadcasts, by feeling the vibration of the terminal device to simulate the four-dimensional effect, it allows the viewer to use the client to watch the video more immersive, greatly improving the user's experience of watching videos.
  • the present application also proposes a video processing apparatus.
  • FIG. 5 is a structural block diagram of a video processing apparatus according to an embodiment of the present application.
  • the video processing apparatus 500 may include: an adding module 510 and a sending module 520 .
  • the adding module 510 is configured to edit the original video file through a video editor, and add a vibration event to the original video file to generate a target video file, where the vibration event includes a time point and corresponding vibration parameters.
  • the sending module 520 is configured to send the target video file to the target server, so that the user can trigger the sensor of the terminal device to vibrate in response to a vibration event during the process of playing the target video file on the terminal device.
  • the adding module 510 is specifically used for: parsing the original video file and extracting key pictures; querying a preset template database to obtain a target vibration template corresponding to the key pictures, wherein the target vibration template includes vibration parameters; Obtain the time point corresponding to the key picture in the original video file; splicing the target vibration template to the position corresponding to the time point in the original video file to generate the target video file.
  • the video processing apparatus 500 may further include: an integration module 530, where the integration module 530 is used to integrate a template database in the video editor, wherein the template database includes: a device provided The inherent vibration template provided by the manufacturer, and/or the custom vibration template set by the user by selecting the control parameters.
  • the adding module 510 is specifically used for: acquiring the content interest points in the original video file; calling the vibration setting interface, selecting vibration parameters according to the content interest points, and comparing the vibration parameters with the content interest points in the original video file The time point corresponding to the point is associated with the setting.
  • an original video file can be edited by a video editor, a vibration event is added to the original video file to generate a target video file, and the target video file is sent to the target server, so that the user can play the target video file on the terminal device.
  • the sensor of the terminal device is triggered to vibrate in response to the vibration event.
  • Vibration can be added to better express the editor's own video. Vibration can be added to the live broadcast to better interact with the audience. For video viewers, a new watch has been added.
  • the present application further provides another video processing apparatus.
  • FIG. 7 is a structural block diagram of a video processing apparatus according to still another embodiment of the present application.
  • the video processing apparatus 700 may include: an acquisition module 710 , a playback module 720 and a vibration response module 730 .
  • the obtaining module 710 is configured to obtain a target video file containing a vibration event sent by the server, wherein the vibration event includes: a time point and a corresponding vibration parameter.
  • the playing module 720 is used for playing the target video file.
  • the vibration response module 730 is configured to respond to the vibration event when the target video file is played to the time point, and trigger the sensor of the terminal device to vibrate according to the vibration parameter corresponding to the time point.
  • a vibration event can be added to the video, so that when the video is played and is played to the position where the vibration event is added, the original content of the video is played and the video is displayed to the user. Vibration behavior corresponding to this content.
  • Vibration behavior corresponding to this content.
  • For video editors a new editing method has been added. Vibration can be added to better express the editor's own video. Vibration can be added to the live broadcast to better interact with the audience. For video viewers, a new watch has been added. Feeling, when watching videos and live broadcasts, by feeling the vibration of the terminal device to simulate the four-dimensional effect, it allows the viewer to use the client to watch the video more immersive, greatly improving the user's experience of watching videos.
  • the present application further provides an electronic device and a readable storage medium.
  • FIG. 8 it is a block diagram of an electronic device for implementing a video processing method according to an embodiment of the present application.
  • Electronic devices are intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframe computers, and other suitable computers.
  • Electronic devices may also represent various forms of mobile devices, such as personal digital processors, cellular phones, smart phones, wearable devices, and other similar computing devices.
  • the components shown herein, their connections and relationships, and their functions are by way of example only, and are not intended to limit implementations of the application described and/or claimed herein.
  • the electronic device includes: one or more processors 801, a memory 802, and interfaces for connecting various components, including a high-speed interface and a low-speed interface.
  • the various components are interconnected using different buses and may be mounted on a common motherboard or otherwise as desired.
  • the processor may process instructions executed within the electronic device, including instructions stored in or on memory to display graphical information of the GUI on an external input/output device, such as a display device coupled to the interface.
  • multiple processors and/or multiple buses may be used with multiple memories and multiple memories, if desired.
  • multiple electronic devices may be connected, each providing some of the necessary operations (eg, as a server array, a group of blade servers, or a multiprocessor system).
  • a processor 801 is used as an example.
  • the memory 802 is the non-transitory computer-readable storage medium provided by the present application.
  • the memory stores instructions executable by at least one processor, so that the at least one processor executes the video processing method provided by the present application.
  • the non-transitory computer-readable storage medium of the present application stores computer instructions for causing a computer to execute the video processing method provided by the present application.
  • the memory 802 can be used to store non-transitory software programs, non-transitory computer-executable programs, and modules, such as program instructions/modules corresponding to the video processing method in the embodiments of the present application (for example, appendix).
  • the processor 801 executes various functional applications and data processing of the server by running the non-transitory software programs, instructions and modules stored in the memory 802, ie, implements the video processing method in the above method embodiments.
  • the memory 802 may include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required for at least one function; the storage data area may store a program created according to the use of an electronic device used to implement the video processing method data etc. Additionally, memory 802 may include high-speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid state storage device. In some embodiments, the memory 802 may optionally include memory located remotely relative to the processor 801, and these remote memories may be connected to the electronic device for implementing the video processing method through a network. Examples of such networks include, but are not limited to, the Internet, an intranet, a local area network, a mobile communication network, and combinations thereof.
  • the electronic device for implementing the video processing method may further include: an input device 803 and an output device 804 .
  • the processor 801 , the memory 802 , the input device 803 and the output device 804 may be connected by a bus or in other ways, and the connection by a bus is taken as an example in FIG. 8 .
  • the input device 803 can receive input numerical or character information, and generate key signal input related to user settings and function control of electronic equipment used to implement the video processing method, such as a touch screen, keypad, mouse, trackpad, touchpad, An input device such as a pointing stick, one or more mouse buttons, trackball, joystick, etc.
  • Output devices 804 may include display devices, auxiliary lighting devices (eg, LEDs), haptic feedback devices (eg, vibration motors), and the like.
  • the display device may include, but is not limited to, a liquid crystal display (LCD), a light emitting diode (LED) display, and a plasma display. In some implementations, the display device may be a touch screen.
  • Various implementations of the systems and techniques described herein can be implemented in digital electronic circuitry, integrated circuit systems, application specific ASICs (application specific integrated circuits), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include being implemented in one or more computer programs executable and/or interpretable on a programmable system including at least one programmable processor that The processor, which may be a special purpose or general-purpose programmable processor, may receive data and instructions from a storage system, at least one input device, and at least one output device, and transmit data and instructions to the storage system, the at least one input device, and the at least one output device an output device.
  • the processor which may be a special purpose or general-purpose programmable processor, may receive data and instructions from a storage system, at least one input device, and at least one output device, and transmit data and instructions to the storage system, the at least one input device, and the at least one output device an output device.
  • machine-readable medium and “computer-readable medium” refer to any computer program product, apparatus, and/or apparatus for providing machine instructions and/or data to a programmable processor ( For example, magnetic disks, optical disks, memories, programmable logic devices (PLDs)), including machine-readable media that receive machine instructions as machine-readable signals.
  • machine-readable signal refers to any signal used to provide machine instructions and/or data to a programmable processor.
  • the systems and techniques described herein may be implemented on a computer having a display device (eg, a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to the user ); and a keyboard and pointing device (eg, a mouse or trackball) through which a user can provide input to the computer.
  • a display device eg, a CRT (cathode ray tube) or LCD (liquid crystal display) monitor
  • a keyboard and pointing device eg, a mouse or trackball
  • Other kinds of devices can also be used to provide interaction with the user; for example, the feedback provided to the user can be any form of sensory feedback (eg, visual feedback, auditory feedback, or tactile feedback); and can be in any form (including acoustic input, voice input, or tactile input) to receive input from the user.
  • the systems and techniques described herein may be implemented on a computing system that includes back-end components (eg, as a data server), or a computing system that includes middleware components (eg, an application server), or a computing system that includes front-end components (eg, a user's computer having a graphical user interface or web browser through which a user may interact with implementations of the systems and techniques described herein), or including such backend components, middleware components, Or any combination of front-end components in a computing system.
  • the components of the system may be interconnected by any form or medium of digital data communication (eg, a communication network). Examples of communication networks include: Local Area Networks (LANs), Wide Area Networks (WANs), and the Internet.
  • a computer system can include clients and servers.
  • Clients and servers are generally remote from each other and usually interact through a communication network.
  • the relationship of client and server arises by computer programs running on the respective computers and having a client-server relationship to each other.
  • a vibration event can be added to the video, so that when the video is played and played to the position where the vibration event is added, the original content of the video is played, and the video is displayed to the user.
  • Vibration behavior corresponding to the content.
  • a new editing method has been added.
  • Vibration can be added to better express the editor's own video.
  • Vibration can be added to the live broadcast to better interact with the audience.
  • a new watch has been added.
  • Feeling, when watching videos and live broadcasts by feeling the vibration of the terminal device to simulate the four-dimensional effect, it allows the viewer to use the client to watch the video more immersive, greatly improving the user's experience of watching videos.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Human Computer Interaction (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

本申请公开了一种视频处理方法、装置、电子设备和存储介质,涉及人工智能领域以及视频制作与编辑领域。具体实现方案为:通过视频编辑器编辑原始视频文件,在原始视频文件中添加振动事件生成目标视频文件,其中,振动事件包括:时间点和对应的振动参数;向目标服务器发送目标视频文件,以使用户在终端设备播放目标视频文件的过程中,响应振动事件触发终端设备的传感器进行振动。本申请可以让观看者使用客户端观看视频时更加让其身临其境,大大提升了用户观看视频的体验。

Description

视频处理方法、装置、电子设备和存储介质
相关申请的交叉引用
本申请要求百度在线网络技术(北京)有限公司于2020年6月30日提交的、发明名称为“视频处理方法、装置、电子设备和存储介质”的、中国专利申请号“202010610290.7”的优先权。
技术领域
本申请涉及数据处理领域,具体涉及人工智能领域以及视频制作与编辑领域,特别地涉及一种视频处理方法、装置、电子设备和存储介质。
背景技术
随着互联网的快速发展,短视频和直播已经逐渐走进人们的生活中。相关技术中,针对目前发展火热的短视频和视频直播,其表现形式就是视频与音频的结合。然而,这种视频的表现形式过于单调,对于用户来说,长时间观看这种视频文件,通常会发生走神或不专心等情况,因此,如何让用户使用客户端观看视频时更加让其身临其境,成为亟待解决的问题。
发明内容
本申请提供了一种视频处理方法、装置、电子设备和存储介质。
根据本申请的第一方面,提供了一种视频处理方法,包括:
通过视频编辑器编辑原始视频文件,在所述原始视频文件中添加振动事件生成目标视频文件,其中,所述振动事件包括:时间点和对应的振动参数;
向目标服务器发送所述目标视频文件,以使用户在终端设备播放所述目标视频文件的过程中,响应所述振动事件触发所述终端设备的传感器进行振动。
根据本申请的第二方面,提供了另一种视频处理方法,包括:
获取服务器发送的包含振动事件的目标视频文件,其中,所述振动事件包括:时间点和对应的振动参数;
播放所述目标视频文件,当播放到所述时间点时,响应所述振动事件,根据与所述时间点对应的振动参数触发终端设备的传感器进行振动。
根据本申请的第三方面,提供了一种视频处理装置,包括:
添加模块,用于通过视频编辑器编辑原始视频文件,在所述原始视频文件中添加振动 事件生成目标视频文件,其中,所述振动事件包括:时间点和对应的振动参数;
发送模块,用于向目标服务器发送所述目标视频文件,以使用户在终端设备播放所述目标视频文件的过程中,响应所述振动事件触发所述终端设备的传感器进行振动。
根据本申请的第四方面,提供了另一种视频处理装置,包括:
获取模块,用于获取服务器发送的包含振动事件的目标视频文件,其中,所述振动事件包括:时间点和对应的振动参数;
播放模块,用于播放所述目标视频文件;
振动响应模块,用于在所述目标视频文件播放到所述时间点时,响应所述振动事件,根据与所述时间点对应的振动参数触发终端设备的传感器进行振动。
根据本申请的第五方面,提供了一种电子设备,包括:
至少一个处理器;以及
与所述至少一个处理器通信连接的存储器;其中,
所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行上述第一方面所述的视频处理方法。
根据本申请的第六方面,提出了一种存储有计算机指令的非瞬时计算机可读存储介质,所述计算机指令用于使所述计算机执行上述第一方面所述的视频处理方法。
根据本申请的技术方案,可通过视频编辑器编辑原始视频文件,在原始视频文件中添加振动事件生成目标视频文件,并向目标服务器发送目标视频文件,以使用户在终端设备播放目标视频文件的过程中,响应振动事件触发终端设备的传感器进行振动,可以让观看者使用客户端观看视频时更加让其身临其境,大大提升了用户观看视频的体验。
应当理解,本部分所描述的内容并非旨在标识本申请的实施例的关键或重要特征,也不用于限制本申请的范围。本申请的其它特征将通过以下的说明书而变得容易理解。
附图说明
附图用于更好地理解本方案,不构成对本申请的限定。其中:
图1是根据本申请一个实施例的视频处理方法的流程图;
图2是根据本申请另一个实施例的视频处理方法的流程图;
图3是根据本申请又一个实施例的视频处理方法的流程图;
图4是根据本申请另一个实施例的视频处理方法的流程图;
图5是根据本申请一个实施例的视频处理装置的结构框图;
图6是根据本申请另一个实施例的视频处理装置的结构框图;
图7是根据本申请又一个实施例的视频处理装置的结构框图
图8是用来实现本申请实施例的视频处理方法的电子设备的框图。
具体实施方式
以下结合附图对本申请的示范性实施例做出说明,其中包括本申请实施例的各种细节以助于理解,应当将它们认为仅仅是示范性的。因此,本领域普通技术人员应当认识到,可以对这里描述的实施例做出各种改变和修改,而不会背离本申请的范围和精神。同样,为了清楚和简明,以下的描述中省略了对公知功能和结构的描述。
图1是根据本申请一个实施例的视频处理方法的流程图。需要说明的是,本申请实施例的视频处理方法可应用于本申请实施例的视频处理装置,该视频处理装置可被配置于电子设备上。该电子设备可以是移动终端,该移动终端可以是手机、平板电脑、个人数字助理等具有各种操作***的硬件设备。
如图1所示,该视频处理方法可以包括:
步骤101,通过视频编辑器编辑原始视频文件,在原始视频文件中添加振动事件生成目标视频文件,其中,振动事件包括:时间点和对应的振动参数。
需要说明的是,本申请实施例的视频处理方法可应用于电子设备上,该电子设备上可具有视频编辑器,其中,该视频编辑器可为用户提供视频获取接口,当用户想在某个视频上添加振动效果时,视频编辑器可通过该接口获取到用户所提供的待处理的原始视频文件。也就是说,该原始视频文件可以是用户提供的,利用视频编辑器实现在原始视频文件上添加振动事件,使得生成的视频具有振动效果。
在本申请实施例中,在通过视频编辑器编辑原始视频文件时,可在原始视频文件中添加振动事件生成目标视频文件。其中,该振动事件的添加方式可以是电子设备上视频编辑器采用人工智能技术对原始视频文件的内容进行分析,得到原始视频文件中的特殊元素信息,进而基于该特殊元素信息添加对应的振动事件,从而使得处理后的视频文件带有振动效果。另外,该振动事件的添加方式还可以是用户手动添加的,即用户根据需求在原始视频文件的某个或某些位置处添加振动事件。由此可见,所生成的目标视频文件中既包含有原始视频文件的视频内容还包含振动事件。
步骤102,向目标服务器发送目标视频文件,以使用户在终端设备播放目标视频文件的过程中,响应振动事件触发终端设备的传感器进行振动。
在本申请实施例中,在生成目标视频文件之后,可将该目标视频文件上传到目标服务器上。这样,当有其他用户通过终端设备从目标服务器上下载目标视频文件,并通过终端设备播放该目标视频文件时,当播放到振动事件中所包含的时间的时,响应振动事件,根据与所述时间点对应的振动参数触发终端设备的传感器进行振动。
为了提高效率,简化用户操作步骤,实现视频处理的智能化,可采用人工智能技术对原始视频文件的内容进行分析,得到原始视频文件中的特殊元素信息,进而基于该特殊元素信息自动在原始视频文件中添加振动事件。具体而言,在本申请一些实施例中,如图2所示,所述在原始视频文件中添加振动事件生成目标视频文件的具体实现过程可包括:
步骤201,解析原始视频文件,提取关键画面。
可选地,从原始视频文件中抽取关键帧,并利用图像识别模型对该原始视频文件中的关键帧进行图像识别,以得到该原始视频文件中的关键画面。例如,在恐怖视频中,突然的振动可以瞬间提高恐怖的程度,基于该需求,可从原始视频文件中抽取关键帧,并对关键帧进行图像识别,提取出该原始视频文件中存在的恐怖画面。
又如,在类似讲课或课堂讲解、或培训等类型的视频中,通过振动来提醒观看者精神集中,比如,当某个讲解部分是重点内容,需要观看者集中精神听讲,因此,可通过对原始视频文件中的语音或者视频显示内容,来判断该原始视频文件中哪个或哪些地方为关键画面,以便在该这些关键画面处添加振动事件,以提醒观看者在观看该视频时,通过振动来提醒重点观看该段视频内容。
步骤202,查询预设的模板数据库,获取与关键画面对应的目标振动模板,其中,目标振动模板包括振动参数。
可选地,在从原始视频文件中提取出关键画面之后,可查询预设的模板数据库,以获得与该关键画面对应的目标振动模板。
需要说明的是,该模板数据库可以在电子设备上。可选地,在本申请一些实施例中,可在视频编辑器中集成模板数据库,其中,该模板数据库包括:设备提供商提供的固有振动模板,和/或,用户通过选择控件参数设置的自定义振动模板。在本申请实施例中,振动模板中可包括振动参数,该振动参数可包括但不限于振动次数、每次振动之间的时间间隔、每次振动的时间长短等中的一种或多种。
在本申请的实施例中,每种类型的关键画面可对应至少一种振动模板,每种振动模板中的振动参数不同,这样,在获得原始视频文件之中的关键画面之后,可从查询预设的模板数据库,以获取与该关键画面对应的目标振动模板。例如,以在恐怖视频中添加振动事件为例,假设恐怖等级分为5个等级,每个等级对应有一种振动模板,这样,在获得该恐怖视频的关键画面之后,可确定该关键画面的等级,并根据该关键画面的等级从模板数据库中,找出与该关键画面的等级对应的振动模板。由此,可以为基于关键画面的等级的不同,在原始视频文件中添加不同的振动事件,实现基于视频内容添加相应的振动效果,可以大大提升用户观看视频的体验。
可以理解,对于讲解类的视频而言,也可以根据讲解内容的重点等级,来获取不同的 振动模板,进而在原始视频文件中的对应位置上添加相应振动效果,可以提醒观看者集中精神。
步骤203,获取原始视频文件中关键画面对应的时间点。
也就是说,可获取关键画面在原始视频文件中所对应的时间点。例如,以在恐怖视频中添加振动事件为例,在提取到该恐怖视频中的某个恐怖画面时,可获取该恐怖画面在该恐怖视频中所对应的播放时间点,该播放时间点可作为振动事件在原始视频文件中的添加位置。
步骤204,将目标振动模板拼接到原始视频文件中与时间点对应的位置,生成目标视频文件。
也就是说,在获得原始视频文件中关键画面以及该关键画面在视频文件之中所对应的时间点时,可将该关键画面对应的目标振动模板与原始视频文件进行融合,例如,可在该时间点处添加振动事件的触发标记,将添加了振动事件触发标记的原始视频文件和目标振动模板合成为一个文件,以得到带有振动效果的目标视频文件。
由此,通过采用人工智能技术对原始视频文件的内容进行分析,得到原始视频文件中的特殊元素信息,进而基于该特殊元素信息自动在原始视频文件中添加振动事件,可以简化用户操作步骤,提高效率,实现视频处理的智能化。
为了进一步提升用户体验,满足用户个性化需求,用户可基于自身需求选择振动事件在视频中的添加位置,可选地,在本申请一些实施例中,如图3所示,所述在原始视频文件中添加振动事件生成目标视频文件的具体实现过程可包括:
步骤301,获取原始视频文件中的内容兴趣点。
可选地,通过视频编辑器提取原始视频文件中的各帧图像并显示给用户,用户可根据自身需求选择对原始视频文件中的哪些帧内容感兴趣,从而可以获得原始视频文件中的内容兴趣点。
步骤302,调用振动设置界面,根据内容兴趣点选择振动参数,并将振动参数与原始视频文件中与内容兴趣点对应的时间点进行关联设置。
可选地,在获得用户对原始视频文件中的内容兴趣点之后,可为用户提高振动设置界面,并根据该内容兴趣点选择振动参数,在确定用户选择振动参数之后,可将该用户选择的振动参数与原始视频文件中与内容兴趣点对应的时间点进行关联设置,从而可以生成带有振动效果的目标视频文件。
由此,通过用户基于自身需求选择振动事件在视频中的添加位置,进而在位置添加振动事件,为用户提供在视频中添加振动效果的编辑功能,从而使得用户拥有可以针对视频编辑振动方式的能力,可以满足用户个性化需求,进一步提升了用户体验。
需要说明的是,本申请实施例的原始视频可以是短视频或其他视频,还可以是直播中实时上传到直播服务器上的实时视频流。也就是说,本申请实施例的视频处理方法还可应用于视频直播场景,例如,直播软件中嵌入有视频编辑器,这样在主播通过直播软件进行直播中,主播想向用户推送振动来完成新的互动形式,则主播可通过直播软件中的视频编辑器在直播的某一时刻添加了振动模板,电子设备可获取该主播直播时的实时视频流,并将该时刻在实时视频流所对应的时间点确定为振动事件的添加位置,并将该振动模板与该添加位置进行关联设置,并将关联了振动模板的实时视频流上传到目标服务器上。观看直播的用户通过终端设备从该目标服务器上获取该直播的实时视频流,并播放该直播的实时视频流,当播放到该添加了振动事件的时间点时,可根据振动事件中的振动参数触发终端设备的传感器进行振动,以达到在直播中推送振动给观众的效果。这样,当观看该主播的观众通过终端设备观看其直播时,终端设备可从直播服务器中拉流,并播放从服务器获得的该直播所对应的实时视频流,当播放到对应的振动时间的添加位置(其添加了振动事件的时间点)时,触发终端设备之中的传感器进行振动,以实现在直播中推送振动给观众的效果。
根据本申请实施例的视频处理方法,可通过视频编辑器编辑原始视频文件,在原始视频文件中添加振动事件生成目标视频文件,并向目标服务器发送目标视频文件,以使用户在终端设备播放目标视频文件的过程中,响应振动事件触发终端设备的传感器进行振动。对于视频编辑者,添加了全新的编辑方式,可以添加振动来更好地表现编辑者自己的视频,可以在直播中添加振动来更好地和观众进行互动,对于视频观众,添加了新的观看感受,在观看视频和直播时,通过感受终端设备的振动,来模拟达到四维的效果,可以让观看者使用客户端观看视频时更加让其身临其境,让观看者更好地领会视频的内容,大大提升了用户观看视频的体验。
图4是根据本申请另一个实施例的视频处理方法的流程图。如图4所示,该视频处理方法可以包括:
步骤401,获取服务器发送的包含振动事件的目标视频文件,其中,振动事件包括:时间点和对应的振动参数。
例如,观看者可通过终端设备从服务器上获取该包含振动事件的目标视频文件。其中,该目标视频文件可以是短视频文件,还可以是直播的实时视频流文件。
在本申请实施例中,上述目标视频文件可以是通过使用前述图1至图3所示的视频处理方法对原始视频文件进行编辑处理后得到的视频文件。其中,该目标视频文件包含振动事件,该振动事件包括:时间点和对应的振动参数。其中,该振动参数可包括但不限于振动次数、每次振动之间的时间间隔、每次振动的时间长短等中的一种或多种。
步骤402,播放目标视频文件,当播放到该振动事件中的时间点时,响应振动事件,根据与时间点对应的振动参数触发终端设备的传感器进行振动。
也就是说,在获得目标视频文件时,可播放目标视频文件,在播放过程中,当播放到添加了振动事件的位置时,在播放视频文件原本内容的同时,响应该振动事件,根据该位置上振动事件所对应的振动参数触发终端设备上的传感器进行振动,从而实现在播放视频原始内容的同时,向用户展示该内容所对应的振动行为。其中,在本申请实施例中,该传感器可为振动传感器。
根据本申请实施例的视频处理方法,可通过在视频中添加振动事件,这样,当该视频被播放,且播放到该添加了振动事件的位置时,在播放视频原始内容的同时,向用户展示该内容所对应的振动行为。对于视频编辑者,添加了全新的编辑方式,可以添加振动来更好地表现编辑者自己的视频,可以在直播中添加振动来更好地和观众进行互动,对于视频观众,添加了新的观看感受,在观看视频和直播时,通过感受终端设备的振动,来模拟达到四维的效果,可以让观看者使用客户端观看视频时更加让其身临其境,大大提升了用户观看视频的体验。
为了实现上述实施例,本申请还提出了一种视频处理装置。
图5是根据本申请一个实施例的视频处理装置的结构框图。如图5所示,该视频处理装置500可以包括:添加模块510和发送模块520。
具体地,添加模块510用于通过视频编辑器编辑原始视频文件,在原始视频文件中添加振动事件生成目标视频文件,其中,振动事件包括:时间点和对应的振动参数。
发送模块520用于向目标服务器发送目标视频文件,以使用户在终端设备播放目标视频文件的过程中,响应振动事件触发终端设备的传感器进行振动。
在本申请一些实施例中,添加模块510具体用于:解析原始视频文件,提取关键画面;查询预设的模板数据库,获取与关键画面对应的目标振动模板,其中,目标振动模板包括振动参数;获取原始视频文件中关键画面对应的时间点;将目标振动模板拼接到原始视频文件中与时间点对应的位置,生成目标视频文件。
在本申请实施例中,如图6所示,该视频处理装置500还可包括:集成模块530,其中,集成模块530用于在视频编辑器中集成模板数据库,其中,模板数据库包括:设备提供商提供的固有振动模板,和/或,用户通过选择控件参数设置的自定义振动模板。
在本申请一些实施例中,添加模块510具体用于:获取原始视频文件中的内容兴趣点;调用振动设置界面,根据内容兴趣点选择振动参数,并将振动参数与原始视频文件中与内容兴趣点对应的时间点进行关联设置。
根据本申请实施例的视频处理装置,可通过视频编辑器编辑原始视频文件,在原始视 频文件中添加振动事件生成目标视频文件,并向目标服务器发送目标视频文件,以使用户在终端设备播放目标视频文件的过程中,响应振动事件触发终端设备的传感器进行振动。对于视频编辑者,添加了全新的编辑方式,可以添加振动来更好地表现编辑者自己的视频,可以在直播中添加振动来更好地和观众进行互动,对于视频观众,添加了新的观看感受,在观看视频和直播时,通过感受终端设备的振动,来模拟达到四维的效果,可以让观看者使用客户端观看视频时更加让其身临其境,让观看者更好地领会视频的内容,大大提升了用户观看视频的体验。
为了实现上述实施例,本申请还提供了另一种视频处理装置。
图7是根据本申请又一个实施例的视频处理装置的结构框图。如图7所示,该视频处理装置700可以包括:获取模块710、播放模块720和振动响应模块730。
具体地,获取模块710用于获取服务器发送的包含振动事件的目标视频文件,其中,振动事件包括:时间点和对应的振动参数。
播放模块720用于播放目标视频文件。
振动响应模块730用于在目标视频文件播放到时间点时,响应振动事件,根据与时间点对应的振动参数触发终端设备的传感器进行振动。
根据本申请实施例的视频处理装置,可通过在视频中添加振动事件,这样,当该视频被播放,且播放到该添加了振动事件的位置时,在播放视频原始内容的同时,向用户展示该内容所对应的振动行为。对于视频编辑者,添加了全新的编辑方式,可以添加振动来更好地表现编辑者自己的视频,可以在直播中添加振动来更好地和观众进行互动,对于视频观众,添加了新的观看感受,在观看视频和直播时,通过感受终端设备的振动,来模拟达到四维的效果,可以让观看者使用客户端观看视频时更加让其身临其境,大大提升了用户观看视频的体验。
根据本申请的实施例,本申请还提供了一种电子设备和一种可读存储介质。
如图8所示,是根据本申请实施例的用以实现视频处理方法的电子设备的框图。电子设备旨在表示各种形式的数字计算机,诸如,膝上型计算机、台式计算机、工作台、个人数字助理、服务器、刀片式服务器、大型计算机、和其它适合的计算机。电子设备还可以表示各种形式的移动装置,诸如,个人数字处理、蜂窝电话、智能电话、可穿戴设备和其它类似的计算装置。本文所示的部件、它们的连接和关系、以及它们的功能仅仅作为示例,并且不意在限制本文中描述的和/或者要求的本申请的实现。
如图8所示,该电子设备包括:一个或多个处理器801、存储器802,以及用于连接各部件的接口,包括高速接口和低速接口。各个部件利用不同的总线互相连接,并且可以被安装在公共主板上或者根据需要以其它方式安装。处理器可以对在电子设备内执行的指令 进行处理,包括存储在存储器中或者存储器上以在外部输入/输出装置(诸如,耦合至接口的显示设备)上显示GUI的图形信息的指令。在其它实施方式中,若需要,可以将多个处理器和/或多条总线与多个存储器和多个存储器一起使用。同样,可以连接多个电子设备,各个设备提供部分必要的操作(例如,作为服务器阵列、一组刀片式服务器、或者多处理器***)。图8中以一个处理器801为例。
存储器802即为本申请所提供的非瞬时计算机可读存储介质。其中,所述存储器存储有可由至少一个处理器执行的指令,以使所述至少一个处理器执行本申请所提供的视频处理方法。本申请的非瞬时计算机可读存储介质存储计算机指令,该计算机指令用于使计算机执行本申请所提供的视频处理方法。
存储器802作为一种非瞬时计算机可读存储介质,可用于存储非瞬时软件程序、非瞬时计算机可执行程序以及模块,如本申请实施例中的视频处理方法对应的程序指令/模块(例如,附图5所示的添加模块510和发送模块520,或者,附图7所示的获取模块710、播放模块720和振动响应模块730)。处理器801通过运行存储在存储器802中的非瞬时软件程序、指令以及模块,从而执行服务器的各种功能应用以及数据处理,即实现上述方法实施例中的视频处理方法。
存储器802可以包括存储程序区和存储数据区,其中,存储程序区可存储操作***、至少一个功能所需要的应用程序;存储数据区可存储根据用以实现视频处理方法的电子设备的使用所创建的数据等。此外,存储器802可以包括高速随机存取存储器,还可以包括非瞬时存储器,例如至少一个磁盘存储器件、闪存器件、或其他非瞬时固态存储器件。在一些实施例中,存储器802可选包括相对于处理器801远程设置的存储器,这些远程存储器可以通过网络连接至用以实现视频处理方法的电子设备。上述网络的实例包括但不限于互联网、企业内部网、局域网、移动通信网及其组合。
用以实现视频处理方法的电子设备还可以包括:输入装置803和输出装置804。处理器801、存储器802、输入装置803和输出装置804可以通过总线或者其他方式连接,图8中以通过总线连接为例。
输入装置803可接收输入的数字或字符信息,以及产生与用以实现视频处理方法的电子设备的用户设置以及功能控制有关的键信号输入,例如触摸屏、小键盘、鼠标、轨迹板、触摸板、指示杆、一个或者多个鼠标按钮、轨迹球、操纵杆等输入装置。输出装置804可以包括显示设备、辅助照明装置(例如,LED)和触觉反馈装置(例如,振动电机)等。该显示设备可以包括但不限于,液晶显示器(LCD)、发光二极管(LED)显示器和等离子体显示器。在一些实施方式中,显示设备可以是触摸屏。
此处描述的***和技术的各种实施方式可以在数字电子电路***、集成电路***、专 用ASIC(专用集成电路)、计算机硬件、固件、软件、和/或它们的组合中实现。这些各种实施方式可以包括:实施在一个或者多个计算机程序中,该一个或者多个计算机程序可在包括至少一个可编程处理器的可编程***上执行和/或解释,该可编程处理器可以是专用或者通用可编程处理器,可以从存储***、至少一个输入装置、和至少一个输出装置接收数据和指令,并且将数据和指令传输至该存储***、该至少一个输入装置、和该至少一个输出装置。
这些计算程序(也称作程序、软件、软件应用、或者代码)包括可编程处理器的机器指令,并且可以利用高级过程和/或面向对象的编程语言、和/或汇编/机器语言来实施这些计算程序。如本文使用的,术语“机器可读介质”和“计算机可读介质”指的是用于将机器指令和/或数据提供给可编程处理器的任何计算机程序产品、设备、和/或装置(例如,磁盘、光盘、存储器、可编程逻辑装置(PLD)),包括,接收作为机器可读信号的机器指令的机器可读介质。术语“机器可读信号”指的是用于将机器指令和/或数据提供给可编程处理器的任何信号。
为了提供与用户的交互,可以在计算机上实施此处描述的***和技术,该计算机具有:用于向用户显示信息的显示装置(例如,CRT(阴极射线管)或者LCD(液晶显示器)监视器);以及键盘和指向装置(例如,鼠标或者轨迹球),用户可以通过该键盘和该指向装置来将输入提供给计算机。其它种类的装置还可以用于提供与用户的交互;例如,提供给用户的反馈可以是任何形式的传感反馈(例如,视觉反馈、听觉反馈、或者触觉反馈);并且可以用任何形式(包括声输入、语音输入或者、触觉输入)来接收来自用户的输入。
可以将此处描述的***和技术实施在包括后台部件的计算***(例如,作为数据服务器)、或者包括中间件部件的计算***(例如,应用服务器)、或者包括前端部件的计算***(例如,具有图形用户界面或者网络浏览器的用户计算机,用户可以通过该图形用户界面或者该网络浏览器来与此处描述的***和技术的实施方式交互)、或者包括这种后台部件、中间件部件、或者前端部件的任何组合的计算***中。可以通过任何形式或者介质的数字数据通信(例如,通信网络)来将***的部件相互连接。通信网络的示例包括:局域网(LAN)、广域网(WAN)和互联网。
计算机***可以包括客户端和服务器。客户端和服务器一般远离彼此并且通常通过通信网络进行交互。通过在相应的计算机上运行并且彼此具有客户端-服务器关系的计算机程序来产生客户端和服务器的关系。
根据本申请实施例的技术方案,可通过在视频中添加振动事件,这样,当该视频被播放,且播放到该添加了振动事件的位置时,在播放视频原始内容的同时,向用户展示该内容所对应的振动行为。对于视频编辑者,添加了全新的编辑方式,可以添加振动来更好地 表现编辑者自己的视频,可以在直播中添加振动来更好地和观众进行互动,对于视频观众,添加了新的观看感受,在观看视频和直播时,通过感受终端设备的振动,来模拟达到四维的效果,可以让观看者使用客户端观看视频时更加让其身临其境,大大提升了用户观看视频的体验。
应该理解,可以使用上面所示的各种形式的流程,重新排序、增加或删除步骤。例如,本发申请中记载的各步骤可以并行地执行也可以顺序地执行也可以不同的次序执行,只要能够实现本申请公开的技术方案所期望的结果,本文在此不进行限制。
上述具体实施方式,并不构成对本申请保护范围的限制。本领域技术人员应该明白的是,根据设计要求和其他因素,可以进行各种修改、组合、子组合和替代。任何在本申请的精神和原则之内所作的修改、等同替换和改进等,均应包含在本申请保护范围之内。

Claims (12)

  1. 一种视频处理方法,包括:
    通过视频编辑器编辑原始视频文件,在所述原始视频文件中添加振动事件生成目标视频文件,其中,所述振动事件包括:时间点和对应的振动参数;
    向目标服务器发送所述目标视频文件,以使用户在终端设备播放所述目标视频文件的过程中,响应所述振动事件触发所述终端设备的传感器进行振动。
  2. 如权利要求1所述的方法,其中,在所述原始视频文件中添加振动事件生成目标视频文件,包括:
    解析所述原始视频文件,提取关键画面;
    查询预设的模板数据库,获取与所述关键画面对应的目标振动模板,其中,所述目标振动模板包括振动参数;
    获取所述原始视频文件中所述关键画面对应的时间点;
    将所述目标振动模板拼接到所述原始视频文件中与所述时间点对应的位置,生成所述目标视频文件。
  3. 如权利要求2所述的方法,在所述查询预设的模板数据库之前,还包括:
    在所述视频编辑器中集成所述模板数据库,其中,所述模板数据库包括:设备提供商提供的固有振动模板,和/或,用户通过选择控件参数设置的自定义振动模板。
  4. 如权利要求1所述的方法,其中,在所述原始视频文件中添加振动事件生成目标视频文件,包括:
    获取所述原始视频文件中的内容兴趣点;
    调用振动设置界面,根据所述内容兴趣点选择振动参数,并将所述振动参数与所述原始视频文件中与所述内容兴趣点对应的时间点进行关联设置。
  5. 一种视频处理方法,包括:
    获取服务器发送的包含振动事件的目标视频文件,其中,所述振动事件包括:时间点和对应的振动参数;
    播放所述目标视频文件,当播放到所述时间点时,响应所述振动事件,根据与所述时间点对应的振动参数触发终端设备的传感器进行振动。
  6. 一种视频处理装置,包括:
    添加模块,用于通过视频编辑器编辑原始视频文件,在所述原始视频文件中添加振动事件生成目标视频文件,其中,所述振动事件包括:时间点和对应的振动参数;
    发送模块,用于向目标服务器发送所述目标视频文件,以使用户在终端设备播放所述 目标视频文件的过程中,响应所述振动事件触发所述终端设备的传感器进行振动。
  7. 如权利要求6所述的装置,其中,所述添加模块具体用于:
    解析所述原始视频文件,提取关键画面;
    查询预设的模板数据库,获取与所述关键画面对应的目标振动模板,其中,所述目标振动模板包括振动参数;
    获取所述原始视频文件中所述关键画面对应的时间点;
    将所述目标振动模板拼接到所述原始视频文件中与所述时间点对应的位置,生成所述目标视频文件。
  8. 如权利要求7所述的装置,还包括:
    集成模块,用于在所述视频编辑器中集成所述模板数据库,其中,所述模板数据库包括:设备提供商提供的固有振动模板,和/或,用户通过选择控件参数设置的自定义振动模板。
  9. 如权利要求所述的装置,其中,所述添加模块具体用于:
    获取所述原始视频文件中的内容兴趣点;
    调用振动设置界面,根据所述内容兴趣点选择振动参数,并将所述振动参数与所述原始视频文件中与所述内容兴趣点对应的时间点进行关联设置。
  10. 一种视频处理装置,包括:
    获取模块,用于获取服务器发送的包含振动事件的目标视频文件,其中,所述振动事件包括:时间点和对应的振动参数;
    播放模块,用于播放所述目标视频文件;
    振动响应模块,用于在所述目标视频文件播放到所述时间点时,响应所述振动事件,根据与所述时间点对应的振动参数触发终端设备的传感器进行振动。
  11. 一种电子设备,包括:
    至少一个处理器;以及
    与所述至少一个处理器通信连接的存储器;其中,
    所述存储器存储有可被所述至少一个处理器执行的指令,所述指令被所述至少一个处理器执行,以使所述至少一个处理器能够执行权利要求1至5中任一项所述的视频处理方法。
  12. 一种存储有计算机指令的非瞬时计算机可读存储介质,所述计算机指令用于使所述计算机执行权利要求1至5中任一项所述的视频处理方法。
PCT/CN2020/133172 2020-06-30 2020-12-01 视频处理方法、装置、电子设备和存储介质 WO2022000983A1 (zh)

Priority Applications (4)

Application Number Priority Date Filing Date Title
JP2021553359A JP2022541968A (ja) 2020-06-30 2020-12-01 ビデオ処理方法、装置、電子機器及び記憶媒体
EP20920756.2A EP3955585A4 (en) 2020-06-30 2020-12-01 VIDEO PROCESSING METHOD AND DEVICE, ELECTRONIC DEVICE AND STORAGE MEDIA
KR1020217027630A KR20210114059A (ko) 2020-06-30 2020-12-01 비디오 처리 방법, 장치, 전자 기기 및 저장 매체
US17/461,016 US20210392394A1 (en) 2020-06-30 2021-08-30 Method and apparatus for processing video, electronic device and storage medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010610290.7A CN111935551A (zh) 2020-06-30 2020-06-30 视频处理方法、装置、电子设备和存储介质
CN202010610290.7 2020-06-30

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/461,016 Continuation US20210392394A1 (en) 2020-06-30 2021-08-30 Method and apparatus for processing video, electronic device and storage medium

Publications (1)

Publication Number Publication Date
WO2022000983A1 true WO2022000983A1 (zh) 2022-01-06

Family

ID=73316812

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/133172 WO2022000983A1 (zh) 2020-06-30 2020-12-01 视频处理方法、装置、电子设备和存储介质

Country Status (2)

Country Link
CN (1) CN111935551A (zh)
WO (1) WO2022000983A1 (zh)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115426520A (zh) * 2022-08-30 2022-12-02 无限魔方(深圳)科技有限公司 一种对直播间的直播场景进行布置的方法、装置及设备
WO2023241121A1 (zh) * 2022-06-13 2023-12-21 中兴通讯股份有限公司 数据处理方法、控制装置、电子设备、计算机可读介质

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111935551A (zh) * 2020-06-30 2020-11-13 百度在线网络技术(北京)有限公司 视频处理方法、装置、电子设备和存储介质
CN113411662A (zh) * 2021-06-15 2021-09-17 读书郎教育科技有限公司 一种智慧课堂增加振动提醒的***及方法
CN113411558A (zh) * 2021-06-18 2021-09-17 湖南快乐阳光互动娱乐传媒有限公司 一种多感官观影体验增强方法及装置

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101409797A (zh) * 2008-11-26 2009-04-15 中兴通讯股份有限公司 一种多媒体广播特效播放方法及特效播放终端
CN103458092A (zh) * 2012-05-29 2013-12-18 中国电信股份有限公司 一种移动终端触感播放器以及触感播放方法
US20160342202A1 (en) * 2015-05-18 2016-11-24 Samsung Electronics Co., Ltd. Virtual reality apparatus, feedback apparatus, and system and method for providing realistic service
CN106657975A (zh) * 2016-10-10 2017-05-10 乐视控股(北京)有限公司 视频播放方法及设备
CN106792168A (zh) * 2016-12-09 2017-05-31 北京小米移动软件有限公司 智能设备的控制方法和装置
CN111935551A (zh) * 2020-06-30 2020-11-13 百度在线网络技术(北京)有限公司 视频处理方法、装置、电子设备和存储介质

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012010212A (ja) * 2010-06-28 2012-01-12 Casio Comput Co Ltd 画像表示装置及びプログラム
US9609395B2 (en) * 2012-03-26 2017-03-28 Max Abecassis Second screen subtitles function
CN108319444B (zh) * 2018-01-30 2021-06-29 广州市百果园信息技术有限公司 基于音乐鼓点控制终端震动方法、存储设备及计算机设备

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101409797A (zh) * 2008-11-26 2009-04-15 中兴通讯股份有限公司 一种多媒体广播特效播放方法及特效播放终端
CN103458092A (zh) * 2012-05-29 2013-12-18 中国电信股份有限公司 一种移动终端触感播放器以及触感播放方法
US20160342202A1 (en) * 2015-05-18 2016-11-24 Samsung Electronics Co., Ltd. Virtual reality apparatus, feedback apparatus, and system and method for providing realistic service
CN106657975A (zh) * 2016-10-10 2017-05-10 乐视控股(北京)有限公司 视频播放方法及设备
CN106792168A (zh) * 2016-12-09 2017-05-31 北京小米移动软件有限公司 智能设备的控制方法和装置
CN111935551A (zh) * 2020-06-30 2020-11-13 百度在线网络技术(北京)有限公司 视频处理方法、装置、电子设备和存储介质

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023241121A1 (zh) * 2022-06-13 2023-12-21 中兴通讯股份有限公司 数据处理方法、控制装置、电子设备、计算机可读介质
CN115426520A (zh) * 2022-08-30 2022-12-02 无限魔方(深圳)科技有限公司 一种对直播间的直播场景进行布置的方法、装置及设备

Also Published As

Publication number Publication date
CN111935551A (zh) 2020-11-13

Similar Documents

Publication Publication Date Title
WO2022000983A1 (zh) 视频处理方法、装置、电子设备和存储介质
KR102436734B1 (ko) 비디오 재생 노드 위치 확정 방법, 장치, 전자 장비, 컴퓨터 판독가능 저장 매체 및 컴퓨터 프로그램
US10893307B2 (en) Video subtitle display method and apparatus
WO2019086037A1 (zh) 视频素材的处理方法、视频合成方法、终端设备及存储介质
US10607382B2 (en) Adapting content to augumented reality virtual objects
US9798464B2 (en) Computing device
CN107801096B (zh) 视频播放的控制方法、装置、终端设备及存储介质
WO2020029523A1 (zh) 视频生成方法、装置、电子设备及存储介质
WO2019105467A1 (zh) 信息的分享方法和装置、存储介质、电子装置
CN111541930B (zh) 直播画面的显示方法、装置、终端及存储介质
US20120249575A1 (en) Display device for displaying related digital images
WO2019047878A1 (zh) 语音操控终端的方法、终端、服务器和存储介质
TWI747031B (zh) 視頻播放方法、裝置和多媒體資料播放方法
CN112752121B (zh) 一种视频封面生成方法及装置
CA3001480C (en) Video-production system with dve feature
EP3024223B1 (en) Videoconference terminal, secondary-stream data accessing method, and computer storage medium
CN105637472A (zh) 具有广义屏幕描述的屏幕内容共享***的框架
US20230368461A1 (en) Method and apparatus for processing action of virtual object, and storage medium
CN105808231B (zh) 录制与播放脚本的***及方法
WO2024104333A1 (zh) 演播画面的处理方法、装置、电子设备及存储介质
US20210392394A1 (en) Method and apparatus for processing video, electronic device and storage medium
CN111770388B (zh) 内容处理方法、装置、设备及存储介质
CN114302221A (zh) 一种虚拟现实设备及投屏媒资播放方法
KR102545040B1 (ko) 비디오 재생 방법, 장치, 전자 기기, 저장 매체 및 컴퓨터 프로그램 제품
CN113891108A (zh) 字幕优化方法、装置、电子设备和存储介质

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref document number: 20217027630

Country of ref document: KR

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2021553359

Country of ref document: JP

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2020920756

Country of ref document: EP

Effective date: 20210831

NENP Non-entry into the national phase

Ref country code: DE