CN110659387A - Method and apparatus for providing video - Google Patents

Method and apparatus for providing video Download PDF

Info

Publication number
CN110659387A
CN110659387A CN201910891806.7A CN201910891806A CN110659387A CN 110659387 A CN110659387 A CN 110659387A CN 201910891806 A CN201910891806 A CN 201910891806A CN 110659387 A CN110659387 A CN 110659387A
Authority
CN
China
Prior art keywords
user
video
language type
target
indication information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910891806.7A
Other languages
Chinese (zh)
Inventor
高志超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Zhangmen Science and Technology Co Ltd
Original Assignee
Shanghai Zhangmen Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Zhangmen Science and Technology Co Ltd filed Critical Shanghai Zhangmen Science and Technology Co Ltd
Priority to CN201910891806.7A priority Critical patent/CN110659387A/en
Publication of CN110659387A publication Critical patent/CN110659387A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/73Querying
    • G06F16/735Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9535Search customisation based on user profiles and personalisation

Landscapes

  • Engineering & Computer Science (AREA)
  • Databases & Information Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Multimedia (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

Embodiments of the present disclosure disclose methods and apparatus for providing video. One embodiment of the method comprises: responding to the operation that a user selects a target template video from a preset template video set, and acquiring region indication information of the user; and providing the video obtained after the language type in the target template video is converted into the target language type for the user based on the region indication information and the target template video. The embodiment realizes that videos with different language types and versions are provided for the user according to the regional characteristics of the user, and enriches the diversity of video providing modes.

Description

Method and apparatus for providing video
Technical Field
Embodiments of the present disclosure relate to the field of computer technologies, and in particular, to the field of video processing, and in particular, to a method and an apparatus for providing a video.
Background
At present, there are two main methods for providing videos to users, one is to provide videos with high popularity to users, and the other is to provide videos of interest to users according to the preferences of users.
Along with the continuous improvement of the demand of people on the network entertainment function, the types of videos are more and more abundant, and people are no longer satisfied with the common video providing mode.
Disclosure of Invention
Embodiments of the present disclosure propose methods and apparatuses for providing video.
In a first aspect, an embodiment of the present disclosure provides a method for providing video, the method including: in response to detecting that a user selects a target template video from a preset template video set, acquiring region indication information of the user, wherein the region indication information is used for indicating a region associated with a target language type expected to be used by the user; and providing the video obtained after the language type in the target template video is converted into the target language type for the user based on the region indication information and the target template video.
In some embodiments, providing, to the user, the video that converts the language type in the target template video to the target language type based on the region indication information and the target template video includes: judging whether a finished product video corresponding to the target language type indicated by the region indication information of the user exists in at least one finished product video generated based on the target template video; and in response to detecting that the finished product video corresponding to the target language type indicated by the region indication information of the user exists, providing the finished product video corresponding to the target language type indicated by the region indication information of the user to the user.
In some embodiments, providing, to the user, the video that converts the language type in the target template video to the target language type based on the region indication information and the target template video includes: acquiring audio information of a user; training a first language type conversion model corresponding to the audio information of the user based on the audio information of the user; and converting the language type in the target template video into the target language type by using the first language type conversion model.
In some embodiments, providing, to the user, the video that converts the language type in the target template video to the target language type based on the region indication information and the target template video includes: and converting the language type in the target template video into the target language type by adopting a pre-trained second language type conversion model.
In some embodiments, the second language type conversion model is trained by: acquiring a training sample set, wherein training samples in the training sample set comprise sample videos and videos of target language types corresponding to the sample videos; and taking the sample video as input, taking the video of the target language type corresponding to the sample video as expected output, and training to obtain a second language type conversion model.
In some embodiments, obtaining the regional indication information of the user includes: in response to the fact that the user opens the positioning authority, acquiring position information of the user as region indication information of the user; responding to the fact that the user does not start the positioning authority, and sending the positioning authority to the user to obtain prompt information of the video of the target language type; and responding to the detection that the user starts the positioning authority according to the prompt message, and acquiring the position information of the user as the region indication information of the user.
In some embodiments, obtaining the regional indication information of the user includes: and acquiring the target region type selected by the user from the region type list as the region indication information of the user.
In some embodiments, obtaining the regional indication information of the user includes: acquiring audio information of a user; and identifying a region associated with the language type used by the user based on the audio information as region indication information of the user.
In a second aspect, an embodiment of the present disclosure provides an apparatus for providing video, the apparatus including: an acquisition unit: responding to an operation of selecting a target template video from a preset template video set by a user, and acquiring region indication information of the user, wherein the region indication information is used for indicating a region associated with a target language type expected to be used by the user; and the sending unit is used for providing the video obtained by converting the language type in the target template video into the target language type for the user.
In a third aspect, an embodiment of the present application provides an electronic device, including: one or more processors; a storage device for storing one or more programs which, when executed by one or more processors, cause the one or more processors to implement a method as in any embodiment of the image processing method.
In a fourth aspect, the present application provides a computer-readable storage medium, on which a computer program is stored, which when executed by a processor implements a method as in any one of the embodiments of the image processing method.
According to the method and the device for providing the video, firstly, the region indication information of the user is obtained in response to the fact that the user selects the target template video from the preset template video set, and then the video obtained after the language type in the target template video is converted into the target language type is provided for the user on the basis of the region indication information and the target template video, so that the videos with different language type versions are provided for the user according to the region characteristics of the user, more interesting videos are provided for the user, and the diversity of video providing modes is enriched.
Drawings
Other features, objects and advantages of the disclosure will become more apparent upon reading of the following detailed description of non-limiting embodiments thereof, made with reference to the accompanying drawings in which:
FIG. 1 is an exemplary system architecture diagram in which one embodiment of the present disclosure may be applied;
FIG. 2 is a flow diagram for one embodiment of a method for providing video, according to the present disclosure;
FIG. 3 is a schematic diagram of an application scenario for obtaining positioning rights of a user according to an embodiment of the present disclosure;
FIG. 4 is a schematic diagram of an application scenario showing a user a list of zone types, in accordance with an embodiment of the present disclosure;
FIG. 5 is a flow diagram of yet another embodiment of a method for providing video in accordance with the present disclosure;
FIG. 6 is a schematic structural diagram of an electronic device suitable for use in implementing embodiments of the present disclosure.
Detailed Description
The present disclosure is described in further detail below with reference to the accompanying drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the relevant invention and not restrictive of the invention. It should be noted that, for convenience of description, only the portions related to the related invention are shown in the drawings.
It should be noted that, in the present disclosure, the embodiments and features of the embodiments may be combined with each other without conflict. The present disclosure will be described in detail below with reference to the accompanying drawings in conjunction with embodiments.
Fig. 1 illustrates an exemplary architecture 100 to which the method for providing video or the apparatus for providing video of the present disclosure may be applied.
As shown in fig. 1, the system architecture 100 may include terminal devices 101, 102, 103, a network 104, and a server 105. The network 104 serves as a medium for providing communication links between the terminal devices 101, 102, 103 and the server 105. Network 104 may include various connection types, such as wired, wireless communication links, or fiber optic cables, to name a few.
The terminal devices 101, 102, 103 interact with a server 105 via a network 104 to receive or send messages or the like. The terminal devices 101, 102, 103 may have various communication client applications installed thereon, such as video playing software, a web browser application, a shopping application, a search application, an instant messaging tool, a mailbox client, social platform software, a text editing application, a browser application, a reading application, and the like.
The terminal apparatuses 101, 102, and 103 may be hardware or software. When the terminal devices 101, 102, 103 are hardware, they may be various electronic devices having a display screen and supporting video playing, including but not limited to smart phones, tablet computers, e-book readers, MP3 players (Moving Picture Experts Group Audio Layer III, mpeg compression standard Audio Layer 3), MP4 players (Moving Picture Experts Group Audio Layer IV, mpeg compression standard Audio Layer 4), laptop portable computers, desktop computers, and the like. When the terminal apparatuses 101, 102, 103 are software, they can be installed in the electronic apparatuses listed above. It may be implemented as multiple pieces of software or software modules (e.g., software or software modules used to provide distributed services) or as a single piece of software or software module. And is not particularly limited herein.
The server 105 may be a server providing various services, such as a background application server providing support for applications on the terminal devices 101, 102, 103. The background application server may receive the application data acquisition request sent by the terminal device 101, 102, 103, search for corresponding data, and return the data to the terminal device 101, 102, 103.
It should be noted that the method for providing video provided by the embodiment of the present disclosure may be executed by the terminal devices 101, 102, 103, or may be executed by the server 105. Accordingly, the means for providing video may be provided in the terminal devices 101, 102, 103, or in the server 105. This is not limited in this application.
It should be understood that the number of terminal devices, networks, and servers in fig. 1 is merely illustrative. There may be any number of terminal devices, networks, and servers, as desired for implementation.
With continued reference to fig. 2, a flow 200 of one embodiment of a method for providing video in accordance with the present disclosure is shown. The method for providing video includes the steps of:
step 201, in response to detecting that a user selects a target template video from a preset template video set, obtaining region indication information of the user.
In the present embodiment, an executing subject (e.g., a terminal device or a server shown in fig. 1) of the method for providing a video may detect an operation of a user selecting a target template video from a preset template video set. Here, the template video set may include a plurality of videos, which may be derived from movie fragments, and the like. The target template video may be a video selected by a user in a set of template videos. The execution main body acquires the region indication information of the user after detecting the operation of selecting the target template by the user. Here, the region indication information may be used to indicate a region associated with a target language type that the user desires to use. The execution main body may obtain the region indication information of the user according to the history information of the video watched by the user, for example, most of videos in the history information of the video watched by the user are language type videos corresponding to the region a, and the execution main body uses the region a as the region indication information of the user, that is, the region associated with the target language type that the user desires to use is the region a. The target language type may be a language type used by a certain region, such as japanese, korean, english, shanghai dialect, sichuan dialect, and the like.
In practice, a user selects and clicks to watch an interested video from a video set list which is provided by an application program official platform and can be selected to watch, and the video selected by the user is a target template video. And after the user selects the target template video, acquiring the region indication information of the user.
In some optional implementation manners of this embodiment, the obtaining the region indication information of the user includes: and responding to the fact that the user opens the positioning authority, and obtaining the position information of the user as the region indication information of the user.
In this implementation, in response to determining that the user has opened the Positioning right, the execution main body may directly use the location information of the user as the region indication information of the user, and may obtain the location information of the user through a Global Positioning System (GPS), for example. The execution subject can determine that the user has opened the positioning authority by detecting whether the user allows the application program to access the position of the terminal device.
In some optional implementations, in response to determining that the user does not turn on the positioning right, sending a prompt to the user to turn on the positioning right to obtain the video in the target language type; and responding to the detection that the user starts the positioning authority according to the prompt message, and acquiring the position information of the user as the region indication information of the user.
In this implementation manner, the execution main body detects that the user does not start the positioning right, sends a prompt message for guiding the user to start the positioning right to the user, further initiates a request for accessing the location information to the user, and acquires the location information of the user as the region indication information of the user when the user starts the positioning right according to the content of the prompt message. Here, the reminder may be used to guide the user to turn on the positioning right, e.g. the content of the reminder may "turn on positioning, create a more interesting video".
In practice, when the user refuses to start the positioning right, authorization prompt information can be continuously sent to the user, and the user is further guided to start the positioning right.
Through the implementation mode, the user can be guided to open the positioning authority so as to obtain the position information of the user, and then the language type version video matched with the position information is provided for the user.
In some optional implementation manners of this embodiment, the obtaining the region indication information of the user includes: and acquiring the target region type selected by the user from the region type list as the region indication information of the user.
In this implementation manner, the execution main body may display the region type list to the user, the user may manually select the target region type corresponding to the language that the user desires to use, and the execution main body uses the target region type selected by the user as the region indication information of the user. Here, the region type categories may be exemplarily divided into a country category, a province category, and a city category. The execution main body can load the province list corresponding to the target country after the target country selected by the user in the country list is displayed in a country, province and city step-by-step menu mode, then load the city list corresponding to the target province according to the target province selected by the user, and finally select the target city as the region indication information of the user. The implementation mode can realize that the user quickly inquires the region associated with the used language type expected by the user.
In some optional implementation manners of this embodiment, the obtaining the region indication information of the user includes: acquiring audio information of a user; and identifying a region associated with the language type used by the user based on the audio information as region indication information of the user.
In this implementation manner, the execution main body may obtain audio information of a user, and determine region information corresponding to a language type expected to be used by the user according to an audio feature of the user. Here, the audio information of the user may be derived from video or audio recorded by the user. The execution main body can identify the hometown of the user or the region where the user lives for a long time as the region indication information of the user through the voice recognition model after extracting the audio features of the user. Here, the speech recognition model can be used to recognize the type of language, for example, dialects of various places.
In practice, by identifying the audio information of the user, the hometown information of the user or the region information where the user frequently lives can be obtained. And providing the corresponding language type version video for the user according to the home information of the user or the regional information of frequent residence of the user.
By the implementation mode, the participation degree of the user can be improved, and the video of the language type version matched with the hometown information is provided for the user.
Step 202, providing the video obtained by converting the language type in the target template video into the target language type to the user based on the region indication information and the target template video.
In this embodiment, the executing entity may provide the target template video of the target language type version to the user according to the region associated with the target language type that the user desires to use and indicated by the region indication information of the user after the user selects the target template video in the template video set. Here, the target template video of the target language type version may be a video in which the language type in the target template video is converted into the target language type, that is, only the language type of the audio of the video is converted, and the text content of the audio is not changed. The execution main body can convert the text corresponding to the audio of the target template video into the audio of the target language type through a voice synthesis technology, and then fuse the audio of the target language type with the picture of the target template video according to a time axis to obtain the video obtained after the language type in the target template video is converted into the target language type.
In practice, the user region indication information obtained by the execution main body provides the target template video of the target language type version to the user according to the target language type which the user may expect to use. For example, if the user selects a video, and the region information of the language type desired to be used is the prefecture of the prefecture, a video of the prefecture version of the prefecture is provided to the user.
In the method for providing a video according to the above embodiment of the present disclosure, in response to detecting that a user selects a target template video from a preset template video set, region indication information of the user is obtained, and based on the region indication information and the target template video, a video obtained by converting a language type in the target template video into a target language type is provided to the user, so that videos of different language type versions are provided to the user according to the region characteristics of the user, more interesting videos are provided to the user, and diversity of video providing modes is enriched.
With continued reference to fig. 3, fig. 3 is a schematic diagram of an application scenario for acquiring a positioning right of a user according to an embodiment of the present disclosure. In the application scenario of fig. 3, the execution main body sends a prompt message 301 requesting permission to access the location information of the user to the user, and when the user clicks a button that is not permitted, the execution main body continues to send a prompt message 302 guiding the user to start the positioning authority to the user.
With continued reference to fig. 4, fig. 4 is a schematic diagram of an application scenario showing a user a list of zone types, according to an embodiment of the present disclosure. In the application scenario of fig. 4, the execution main body may send a prompt message 401 for selecting a region to the user, where the prompt message 401 may include options of a country, a province, and a city, and the user may sequentially select a corresponding region type.
With further reference to fig. 5, a flow 500 of yet another embodiment of a method for providing video is shown. The flow 500 of the method for providing video comprises the steps of:
step 501, in response to detecting that a user selects a target template video from a preset template video set, obtaining region indication information of the user.
Step 501 is the same as step 201 in the foregoing embodiment, and the above description for step 201 also applies to step 501, which is not described herein again.
Step 502, judging whether a finished product video corresponding to a target language type indicated by regional indication information of a user exists in at least one finished product video generated based on a target template video; and in response to detecting that the finished product video corresponding to the target language type indicated by the region indication information of the user exists, providing the finished product video corresponding to the target language type indicated by the region indication information of the user to the user.
In this embodiment, the execution subject matches, among at least one finishing video generated based on the target template video, whether a finishing video exists in a version corresponding to the target language type. And when detecting that the finished product video of the target language type version expected by the user exists, providing the finished product video of the target language type version for the user. Here, the at least one finished video generated based on the target template video may be a video that has been previously generated based on the same regional indication information of the other users and the same target template video. For example, based on that the regional indication information of the other users indicates that the target template video is the shanghai dialect and the selected target template video is the a video, the a video of the shanghai dialect version is generated. And when the situation that the selected target template video is the video A indicated by the regional indication information of the current user is detected, directly providing the video A of the prepared version of the Shanghai dialect for the user. Alternatively, the at least one finishing video generated based on the target template video may be a language type version of a video previously made and stored by a developer corresponding to the target template video. For example, for the video a, a developer may create and store an english version of the video a, a shanghai dialect version of the video a, a Shandongdong dialect version of the video a, and the like in advance, and when the user selects the video a, and finds the corresponding video a in the target language version according to the region indication information of the user, provide the video a to the user.
As can be seen from fig. 5, the process 500 of the method for providing a video in this embodiment determines a video in a target language type version expected by a user from at least one finished video generated based on a target template video, which improves the efficiency of determining the video in the target language type version expected by the user.
In some optional implementations of the embodiments described above in conjunction with fig. 2 and 5, providing, to the user, the video after converting the language type in the target template video into the target language type based on the region indication information and the target template video includes: acquiring audio information of a user; training a first language type conversion model corresponding to the audio information of the user based on the audio information of the user; and converting the language type in the target template video into the target language type by using the first language type conversion model.
In the implementation manner, the execution main body obtains audio information of a user, trains a first language type conversion model for converting a language type into a target language type corresponding to the audio information of the user by using the audio information of the user as a training sample, inputs a target template video selected by the user into the first language type conversion model, converts audio of the target template video into audio of the target language type, and then fuses the audio of the target language type with a picture of the template video according to a time axis to obtain a video of a target language type version. For example, if the audio information of the user corresponds to a Shandong dialect, a conversion model for converting the language type into the Shandong dialect is trained, and then the A video selected by the user is input into the Shandong language conversion model to generate the A video of the Shandong dialect version. Here, the audio information of the user may be derived from video or audio recorded by the user.
The video of the target language type generated by the implementation mode is more consistent with the tone quality of the user, and the video which is more interesting to the user can be provided.
In some optional implementations of the embodiments described above in conjunction with fig. 2 and 5, providing a video for converting a language type in a target template video into a target language type to a user based on the region indication information and the target template video includes: and converting the language type in the target template video into the target language type by adopting a pre-trained second language type conversion model.
In this implementation manner, the execution subject inputs the target template video to a pre-trained language type conversion model corresponding to the second target language type based on the region indication information of the user, and converts the language type in the target template video into the target language type to obtain the video of the target language type version.
The implementation mode can convert the language type based on a pre-trained model, so that the audio in the video obtained after the language type is converted is smoother and closer to the audio of a real user.
In this implementation, further, the second language type conversion model may be obtained by training:
the first step is as follows: acquiring a training sample set, wherein training samples in the training sample set comprise sample videos and videos of target language types corresponding to the sample videos.
In this alternative implementation, the executing entity may obtain an existing training sample set stored therein from a database server. As another example, the user may generate the training sample set by collecting training samples via the terminal device and storing the samples locally.
Here, the sample set may include at least one sample. Wherein the sample may include a sample video and a video of the target language type corresponding to the sample video.
The second step is that: and taking the sample video as input, taking the video of the target language type corresponding to the sample video as expected output, and training to obtain a language type conversion model.
In this optional implementation, an electronic device (e.g., a server) communicatively connected to the execution subject may input a sample video in a training sample into an initial neural network, obtain a sample prediction video, and train the initial neural network using a machine learning method with a video of a target language type corresponding to the sample video as an expected output. Specifically, the difference between the obtained sample predicted video and the video of the target language type corresponding to the sample video may be calculated by a preset loss function. Then, network parameters of the initial neural network can be iteratively adjusted based on the calculated difference, and when a preset training completion condition is met, the trained initial neural network is determined as a second language type conversion model.
Here, various implementations may be employed to adjust network parameters of the initial neural network based on differences between the generated sample predicted video and the video of the target language type to which the sample video corresponds. For example, a BP (back propagation) algorithm or an SGD (Stochastic Gradient Descent) algorithm may be used to adjust the network parameters of the initial neural network.
In some optional implementation manners, when the template video selected by the user is acquired, the execution main body may extract audio from the template video, convert the audio of the video of the target template by using the language type conversion model, convert the audio of the template video into the audio of the target language type, and then fuse the audio of the target language type with the picture of the template video according to a time axis to obtain the video of the target language type version.
Here, the second language type conversion model may be obtained by training the sample audio in the training sample set and the audio of the target language type corresponding to the sample audio.
According to the implementation mode, the language type conversion model is trained based on the sample data in advance, so that the accuracy and the reliability of the language type conversion model can be further improved, and the language type in the video can be converted more accurately.
Referring now to fig. 6, and referring now to fig. 6, a block diagram of an electronic device (e.g., a terminal device or server of fig. 1) 600 suitable for use in implementing embodiments of the present disclosure is shown. The terminal device shown in fig. 6 is only an example, and should not bring any limitation to the functions and the use range of the embodiments of the present disclosure.
As shown in fig. 6, electronic device 600 may include a processing means (e.g., central processing unit, graphics processor, etc.) 601 that may perform various appropriate actions and processes in accordance with a program stored in a Read Only Memory (ROM)602 or a program loaded from a storage means 608 into a Random Access Memory (RAM) 603. In the RAM603, various programs and data necessary for the operation of the electronic apparatus 600 are also stored. The processing device 601, the ROM 602, and the RAM603 are connected to each other via a bus 604. An input/output (I/O) interface 605 is also connected to bus 604.
Generally, the following devices may be connected to the I/O interface 605: input devices 606 including, for example, a touch screen, touch pad, keyboard, mouse, camera, microphone, accelerometer, gyroscope, etc.; an output device 607 including, for example, a Liquid Crystal Display (LCD), a speaker, a vibrator, and the like; storage 608 including, for example, tape, hard disk, etc.; and a communication device 609. The communication means 609 may allow the electronic device 600 to communicate with other devices wirelessly or by wire to exchange data. While fig. 6 illustrates an electronic device 600 having various means, it is to be understood that not all illustrated means are required to be implemented or provided. More or fewer devices may alternatively be implemented or provided. Each block shown in fig. 6 may represent one device or may represent multiple devices as desired.
In particular, according to an embodiment of the present disclosure, the processes described above with reference to the flowcharts may be implemented as computer software programs. For example, embodiments of the present disclosure include a computer program product comprising a computer program embodied on a computer readable medium, the computer program comprising program code for performing the method illustrated in the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network via the communication means 609, or may be installed from the storage means 608, or may be installed from the ROM 602. The computer program, when executed by the processing device 601, performs the above-described functions defined in the methods of embodiments of the present disclosure.
It should be noted that the computer readable medium of the embodiments of the present disclosure may be a computer readable signal medium or a computer readable storage medium or any combination of the two. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination of the foregoing. More specific examples of the computer readable storage medium may include, but are not limited to: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a Random Access Memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In embodiments of the disclosure, a computer readable storage medium may be any tangible medium that can contain, or store a program for use by or in connection with an instruction execution system, apparatus, or device. In embodiments of the present disclosure, however, a computer readable signal medium may comprise a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated data signal may take many forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof. A computer readable signal medium may also be any computer readable medium that is not a computer readable storage medium and that can communicate, propagate, or transport a program for use by or in connection with an instruction execution system, apparatus, or device. Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to: electrical wires, optical cables, RF (Radio Frequency), etc., or any suitable combination of the foregoing.
The computer readable medium may be embodied in the electronic device; or may exist separately without being assembled into the electronic device. The computer readable medium carries one or more programs which, when executed by the electronic device, cause the electronic device to: in response to detecting that a user selects a target template video from a preset template video set, acquiring region indication information of the user, wherein the region indication information is used for indicating a region associated with a target language type expected to be used by the user; and providing the video obtained after the language type in the target template video is converted into the target language type for the user based on the region indication information and the target template video.
Computer program code for carrying out operations for embodiments of the present disclosure may be written in any combination of one or more programming languages, including an object oriented programming language such as Java, Smalltalk, C + +, and conventional procedural programming languages, such as the "C" programming language or similar programming languages. The program code may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server. In the case of a remote computer, the remote computer may be connected to the user's computer through any type of network, including a Local Area Network (LAN) or a Wide Area Network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet service provider).
The flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present disclosure. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
The foregoing description is only exemplary of the preferred embodiments of the disclosure and is illustrative of the principles of the technology employed. It will be appreciated by those skilled in the art that the scope of the invention in the embodiments of the present disclosure is not limited to the specific combination of the above-mentioned features, but also encompasses other embodiments in which any combination of the above-mentioned features or their equivalents is made without departing from the inventive concept as defined above. For example, the above features and (but not limited to) technical features with similar functions disclosed in the embodiments of the present disclosure are mutually replaced to form the technical solution.

Claims (11)

1. A method for providing video, comprising:
in response to detecting that a user selects a target template video from a preset template video set, acquiring region indication information of the user, wherein the region indication information is used for indicating a region associated with a target language type expected to be used by the user;
and providing the video obtained by converting the language type in the target template video into the target language type for the user based on the region indication information and the target template video.
2. The method of claim 1, wherein the providing, to the user, the video that converts the language type in the target template video to the target language type based on the region indication information and the target template video comprises:
judging whether a finished product video corresponding to the target language type indicated by the region indication information of the user exists in at least one finished product video generated based on the target template video;
and in response to detecting that the finished product video corresponding to the target language type indicated by the region indication information of the user exists, providing the finished product video corresponding to the target language type indicated by the region indication information of the user to the user.
3. The method according to claim 1 or 2, wherein the providing, to the user, the video for converting the language type in the target template video into the target language type based on the region indication information and the target template video comprises:
acquiring audio information of a user;
training a first language type conversion model corresponding to the audio information of the user based on the audio information of the user;
and converting the language type in the target template video into the target language type by using the first language type conversion model.
4. The method according to claim 1 or 2, wherein the providing, to the user, the video for converting the language type in the target template video into the target language type based on the region indication information and the target template video comprises:
and converting the language type in the target template video into the target language type by adopting a pre-trained second language type conversion model.
5. The method of claim 4, wherein the second language type conversion model is trained by:
acquiring a training sample set, wherein training samples in the training sample set comprise sample videos and videos of target language types corresponding to the sample videos;
and taking the sample video as input, taking the video of the target language type corresponding to the sample video as expected output, and training to obtain the second language type conversion model.
6. The method of claim 1, wherein the obtaining of the regional indication information of the user comprises:
and responding to the fact that the user opens the positioning authority, and obtaining the position information of the user as the region indication information of the user.
7. The method of claim 7, wherein the obtaining of the regional indication information of the user comprises:
responding to the fact that the user does not start the positioning authority, and sending the positioning authority to the user to obtain prompt information of the video of the target language type;
and responding to the detection that the user starts the positioning authority according to the prompt message, and acquiring the position information of the user as the region indication information of the user.
8. The method of claim 1, wherein the obtaining of the regional indication information of the user comprises:
and acquiring the target region type selected by the user from the region type list as the region indication information of the user.
9. The method of claim 1, wherein the obtaining of the regional indication information of the user comprises:
acquiring audio information of a user;
and identifying a region associated with the language type used by the user based on the audio information as region indication information of the user.
10. An electronic device, comprising:
one or more processors;
a storage device having one or more programs stored thereon;
when executed by the one or more processors, cause the one or more processors to implement the method of any one of claims 1-9.
11. A computer-readable medium, on which a computer program is stored which, when being executed by a processor, carries out the method according to any one of claims 1-9.
CN201910891806.7A 2019-09-20 2019-09-20 Method and apparatus for providing video Pending CN110659387A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910891806.7A CN110659387A (en) 2019-09-20 2019-09-20 Method and apparatus for providing video

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910891806.7A CN110659387A (en) 2019-09-20 2019-09-20 Method and apparatus for providing video

Publications (1)

Publication Number Publication Date
CN110659387A true CN110659387A (en) 2020-01-07

Family

ID=69038282

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910891806.7A Pending CN110659387A (en) 2019-09-20 2019-09-20 Method and apparatus for providing video

Country Status (1)

Country Link
CN (1) CN110659387A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112788529A (en) * 2021-01-15 2021-05-11 拉扎斯网络科技(上海)有限公司 Information display method and device and electronic equipment
CN112925512A (en) * 2021-01-29 2021-06-08 北京字节跳动网络技术有限公司 Message sending method and device, electronic equipment and storage medium
CN114297457A (en) * 2021-12-31 2022-04-08 上海掌门科技有限公司 Method and equipment for acquiring regional version content

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1407795A (en) * 2001-08-30 2003-04-02 通用仪器公司 Device and method for providing TV speech-sounds with selected language
CN104252861A (en) * 2014-09-11 2014-12-31 百度在线网络技术(北京)有限公司 Video voice conversion method, video voice conversion device and server
CN105025319A (en) * 2015-07-09 2015-11-04 无锡天脉聚源传媒科技有限公司 Video pushing method and device
CN105260240A (en) * 2015-10-15 2016-01-20 珠海格力电器股份有限公司 Language switching method and device of application program
CN107515862A (en) * 2017-09-01 2017-12-26 北京百度网讯科技有限公司 Voice translation method, device and server
CN108737845A (en) * 2018-05-22 2018-11-02 北京百度网讯科技有限公司 Processing method, device, equipment and storage medium is broadcast live
CN109949793A (en) * 2019-03-06 2019-06-28 百度在线网络技术(北京)有限公司 Method and apparatus for output information

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1407795A (en) * 2001-08-30 2003-04-02 通用仪器公司 Device and method for providing TV speech-sounds with selected language
CN104252861A (en) * 2014-09-11 2014-12-31 百度在线网络技术(北京)有限公司 Video voice conversion method, video voice conversion device and server
CN105025319A (en) * 2015-07-09 2015-11-04 无锡天脉聚源传媒科技有限公司 Video pushing method and device
CN105260240A (en) * 2015-10-15 2016-01-20 珠海格力电器股份有限公司 Language switching method and device of application program
CN107515862A (en) * 2017-09-01 2017-12-26 北京百度网讯科技有限公司 Voice translation method, device and server
CN108737845A (en) * 2018-05-22 2018-11-02 北京百度网讯科技有限公司 Processing method, device, equipment and storage medium is broadcast live
CN109949793A (en) * 2019-03-06 2019-06-28 百度在线网络技术(北京)有限公司 Method and apparatus for output information

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112788529A (en) * 2021-01-15 2021-05-11 拉扎斯网络科技(上海)有限公司 Information display method and device and electronic equipment
CN112788529B (en) * 2021-01-15 2023-04-25 拉扎斯网络科技(上海)有限公司 Information display method and device and electronic equipment
CN112925512A (en) * 2021-01-29 2021-06-08 北京字节跳动网络技术有限公司 Message sending method and device, electronic equipment and storage medium
CN114297457A (en) * 2021-12-31 2022-04-08 上海掌门科技有限公司 Method and equipment for acquiring regional version content

Similar Documents

Publication Publication Date Title
CN109858445B (en) Method and apparatus for generating a model
EP3271917B1 (en) Communicating metadata that identifies a current speaker
US9635129B2 (en) Automatic application discovery, download, integration and launch
CN107731229B (en) Method and apparatus for recognizing speech
CN109993150B (en) Method and device for identifying age
CN107463700B (en) Method, device and equipment for acquiring information
CN109919244B (en) Method and apparatus for generating a scene recognition model
CN112509562B (en) Method, apparatus, electronic device and medium for text post-processing
CN109981787B (en) Method and device for displaying information
CN110659387A (en) Method and apparatus for providing video
CN109862100B (en) Method and device for pushing information
CN110084317B (en) Method and device for recognizing images
US11750898B2 (en) Method for generating target video, apparatus, server, and medium
CN110046571B (en) Method and device for identifying age
CN112165647B (en) Audio data processing method, device, equipment and storage medium
CN111897950A (en) Method and apparatus for generating information
CN113395538B (en) Sound effect rendering method and device, computer readable medium and electronic equipment
CN110990598A (en) Resource retrieval method and device, electronic equipment and computer-readable storage medium
CN111883139A (en) Method, apparatus, device and medium for screening target voices
US10997963B1 (en) Voice based interaction based on context-based directives
CN112309387A (en) Method and apparatus for processing information
CN112309389A (en) Information interaction method and device
JP2024507734A (en) Speech similarity determination method and device, program product
CN110188833B (en) Method and apparatus for training a model
CN113360704A (en) Voice playing method and device and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination