WO2018090911A1 - 一种文件处理方法、终端及服务器、计算机存储介质 - Google Patents

一种文件处理方法、终端及服务器、计算机存储介质 Download PDF

Info

Publication number
WO2018090911A1
WO2018090911A1 PCT/CN2017/110949 CN2017110949W WO2018090911A1 WO 2018090911 A1 WO2018090911 A1 WO 2018090911A1 CN 2017110949 W CN2017110949 W CN 2017110949W WO 2018090911 A1 WO2018090911 A1 WO 2018090911A1
Authority
WO
WIPO (PCT)
Prior art keywords
file
processed
image
recording
audio
Prior art date
Application number
PCT/CN2017/110949
Other languages
English (en)
French (fr)
Inventor
王通
罗谷才
Original Assignee
腾讯科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 腾讯科技(深圳)有限公司 filed Critical 腾讯科技(深圳)有限公司
Publication of WO2018090911A1 publication Critical patent/WO2018090911A1/zh
Priority to US16/297,900 priority Critical patent/US10855969B2/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/189Recording image signals; Reproducing recorded image signals
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/10File systems; File servers
    • G06F16/17Details of further file system functions
    • G06F16/176Support for shared access to files; File sharing support
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/71Indexing; Data structures therefor; Storage structures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/122Improving the 3D impression of stereoscopic images by modifying image signal contents, e.g. by filtering or adding monoscopic depth cues
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/139Format conversion, e.g. of frame-rate or size
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/161Encoding, multiplexing or demultiplexing different image signal components
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/332Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
    • H04N13/344Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/233Processing of audio elementary streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234309Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by transcoding between formats or standards, e.g. from MPEG-2 to MPEG-4 or from Quicktime to Realvideo
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234363Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the spatial resolution, e.g. for clients with a lower screen resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/414Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/414Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance
    • H04N21/41407Specialised client platforms, e.g. receiver in car or embedded in a mobile appliance embedded in a portable device, e.g. video client on a mobile phone, PDA, laptop
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/433Content storage operation, e.g. storage operation in response to a pause request, caching operations
    • H04N21/4334Recording operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • H04N21/4782Web browsing, e.g. WebTV
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/85406Content authoring involving a specific file format, e.g. MP4 format

Definitions

  • the present invention relates to a conversion technology, and in particular, to a file processing method, a terminal and a server, and a computer storage medium.
  • VR glasses are relatively common, and VR glasses not only support panoramic pictures, but also support 3D pictures.
  • APP application
  • the processed image file is combined with the VR glasses, it is necessary to project the distorted image file onto the glasses to have a 3D stereo effect, that is, it cannot be directly in the VR glasses. Watching, the root cause is still limited by the application range brought by the file format.
  • an embodiment of the present invention provides a file processing method, a terminal, and a server.
  • the computer storage medium solves at least the problems of the prior art.
  • a file processing method includes:
  • Obtaining file conversion logic which is logic generated on the server side according to a play plugin, a software development kit SDK, and a recording plugin loaded and loaded in a specified 3D runtime environment;
  • the composition is a first object file and a second object file that are directly used for virtual reality VR playback.
  • An obtaining unit configured to acquire file conversion logic, which is logic generated on a server side according to a play plug-in, a software development kit SDK, and a recording plug-in loaded and loaded in a specified 3D running environment;
  • a playing unit configured to play a first to-be-processed file and a second to-be-processed file obtained based on the source file, where the first to-be-processed file is a left video file of the source file displayed on the terminal screen, and the second to-be-processed file The right video file for the source file displayed on the terminal screen;
  • a first processing unit configured to perform audio and video decoding on the first to-be-processed file and the second to-be-processed file respectively according to the file conversion logic, to obtain audio files and images corresponding to the first to-be-processed file and the second to-be-processed file, respectively a file, performing distortion processing on the image file corresponding to the first to-be-processed file and the second to-be-processed file, respectively, and storing the obtained distortion processing result in an image file format in a memory for file sharing;
  • a second processing unit configured to receive a recording operation, and when the distortion processing result is provided to the recording thread for recording, the distortion processing result is respectively corresponding to the first to-be-processed file and the second to-be-processed according to the file conversion logic
  • the audio file of the file is encoded into a first object file and a second object file that are directly used for virtual reality VR playback.
  • a file processing method includes:
  • the server includes:
  • a generating unit configured to generate file conversion logic according to a playing plug-in, a software development kit SDK, and a recording plug-in loaded and loaded in a specified 3D running environment;
  • a feedback unit configured to respond to the download request initiated by the terminal, and feed back the file conversion logic to the terminal.
  • a computer storage medium according to an embodiment of the present invention, wherein the computer storage medium stores computer executable instructions, and the computer executable instructions are used to execute the file processing method of any of the above.
  • a file processing method is applied to a terminal, where the terminal includes one or more processors and a memory, and one or more programs, wherein the one or more programs are stored in a memory,
  • the program can include one or more modules each corresponding to a set of instructions, the one or more processors being configured to execute instructions;
  • the method includes:
  • Obtaining file conversion logic which is logic generated on the server side according to a play plugin, a software development kit SDK, and a recording plugin loaded and loaded in a specified 3D runtime environment;
  • the composition is a first object file and a second object file that are directly used for virtual reality VR playback.
  • a file processing method is applied to a server, where the server includes one or more processors and a memory, and one or more programs, wherein the one or more programs are stored in a memory.
  • the program can include one or more modules each corresponding to a set of instructions, the one or more processors being configured to execute instructions; the method comprising:
  • file conversion logic is used as a general processing mechanism.
  • a series of processes such as audio and video decoding and distortion processing can be performed to adapt any source file and terminal without installing a dedicated application on the terminal for processing.
  • the object targeted by the processing is a first to-be-processed file and a second to-be-processed file obtained based on the source file, where the first to-be-processed file is a left video file of the source file displayed on the terminal screen, and the second The file to be processed is the right video file of the source file displayed on the screen of the terminal, that is, it is processed for the left/right format of the source file, which achieves better distortion processing effect and does not need to be as prior art.
  • the source file processing results are screened to the VR for playback.
  • the obtained distortion processing result is stored in the image file format in the image memory for file sharing, so that it is not necessary to copy the operation multiple times, and only needs to acquire the required file during sharing, which not only improves efficiency, but also can improve efficiency. Avoid excessive CPU and memory usage.
  • the distortion processing result is provided to the recording thread for recording, and finally the encoding is synthesized into the first target file and the second target file directly used for VR playback.
  • FIG. 1 is a schematic diagram of an optional hardware structure of a mobile terminal implementing various embodiments of the present invention
  • FIG. 2 is a schematic diagram of a communication system of the mobile terminal shown in FIG. 1;
  • FIG. 3 is a schematic diagram of hardware entities of each party performing information interaction in an embodiment of the present invention.
  • FIG. 4 is a schematic diagram of a process for implementing a method according to an embodiment of the present invention.
  • FIG. 5 is a schematic diagram of a process for implementing another method according to an embodiment of the present invention.
  • FIG. 6 is a schematic diagram of a system architecture according to an embodiment of the present invention.
  • FIG. 7 is a schematic diagram of file conversion using a dedicated APP
  • Figure 8-9 is a schematic diagram of an application scenario before and after distortion of an image
  • FIG. 10 is a flowchart of a file conversion according to an embodiment of the present invention.
  • FIG. 11 is a schematic diagram of synthesizing a new video file after the distortion processing according to the embodiment of the present invention.
  • module A mobile terminal embodying various embodiments of the present invention will now be described with reference to the accompanying drawings.
  • suffixes such as “module,” “component,” or “unit” used to denote an element are merely illustrative of the embodiments of the present invention, and do not have a specific meaning per se. Therefore, “module” and “component” can be used in combination.
  • first, second, etc. are used herein to describe various elements (or various thresholds or various applications or various instructions or various operations), etc., these elements (or thresholds) Or application or instruction or operation) should not be limited by these terms. These terms are only used to distinguish one element (or threshold or application or instruction or operation) and another element (or threshold or application or instruction or operation).
  • first operation may be referred to as a second operation
  • second operation may also be referred to as a first operation
  • the first operation and the second operation are both operations, but the two are not the same The operation is only.
  • the steps in the embodiment of the present invention are not necessarily processed in the order of the steps described.
  • the steps may be selectively arranged to be reordered according to requirements, or the steps in the embodiment may be deleted, or the steps in the embodiment may be added.
  • the description of the steps in the embodiments of the present invention is only an optional combination of the steps, and does not represent a combination of the steps of the embodiments of the present invention.
  • the order of the steps in the embodiments is not to be construed as limiting the present invention.
  • the intelligent terminal (such as a mobile terminal) of the embodiment of the present invention can be implemented in various forms.
  • the mobile terminal described in the embodiments of the present invention may include, for example, a mobile phone, a smart phone, a notebook computer, a digital broadcast receiver, a personal digital assistant (PDA, Personal Digital Assistant), a tablet (PAD), a portable multimedia player ( Mobile terminals such as PMP (Portable Media Player), navigation devices, and the like, and fixed terminals such as digital TVs, desktop computers, and the like.
  • PDA Personal Digital Assistant
  • PAD tablet
  • PMP Portable Multimedia Player
  • navigation devices and the like
  • fixed terminals such as digital TVs, desktop computers, and the like.
  • the terminal is a mobile terminal.
  • those skilled in the art will appreciate that configurations in accordance with embodiments of the present invention can be applied to fixed type terminals in addition to components that are specifically for mobile purposes.
  • FIG. 1 is a schematic diagram of an optional hardware structure of a mobile terminal implementing various embodiments of the present invention.
  • the mobile terminal 100 may include a communication unit 110, an audio/video (A/V) input unit 120, a user input unit 130, an acquisition unit 140, a playback unit 141, a first processing unit 142, a second processing unit 143, an output unit 150,
  • Figure 1 illustrates a mobile terminal having various components, but it should be understood that not all illustrated components are required to be implemented. More or fewer components can be implemented instead. The elements of the mobile terminal will be described in detail below.
  • Communication unit 110 typically includes one or more components that permit radio communication between mobile terminal 100 and a wireless communication system or network (electrical communication can also be made by wire if the mobile terminal is replaced with a fixed terminal).
  • the communication unit when it is specifically a wireless communication unit, it may include at least one of a broadcast receiving unit 111, a mobile communication unit 112, a wireless internet unit 113, a short-range communication unit 114, and a location information unit 115, which are optional, according to different Demand can be added or deleted.
  • the broadcast receiving unit 111 receives a broadcast signal and/or broadcast associated information from an external broadcast management server via a broadcast channel.
  • the broadcast channel can include a satellite channel and/or a terrestrial channel.
  • Broadcast management The server may be a server that generates and transmits a broadcast signal and/or broadcast associated information or a server that receives the previously generated broadcast signal and/or broadcast associated information and transmits it to the terminal.
  • the broadcast signal may include a TV broadcast signal, a radio broadcast signal, a data broadcast signal, and the like.
  • the broadcast signal may further include a broadcast signal combined with a TV or radio broadcast signal.
  • the broadcast associated information may also be provided via the mobile communication network, and in this case, the broadcast associated information may be received by the mobile communication unit 112.
  • the broadcast signal may exist in various forms, for example, it may be an Electronic Program Guide (EPG) of Digital Multimedia Broadcasting (DMB), a digital video broadcast handheld (DVB-H, Digital Video Broadcasting-Handheld). ) exists in the form of an ESG (Electronic Service Guide) and the like.
  • EPG Electronic Program Guide
  • DMB Digital Multimedia Broadcasting
  • DVD-H Digital Video Broadcasting-Handheld
  • ESG Electronic Service Guide
  • the broadcast receiving unit 111 can receive a signal broadcast by using various types of broadcast systems.
  • the broadcast receiving unit 111 can use a forward direction such as DMB-T (Digital Multimedia Broadcasting-Terrestrial), Digital Multimedia Broadcasting-Satellite (DMB-S), DVB-H, Forward A digital broadcast system such as a data broadcast system of MediaFLO (Media Forward Link Only), Integrated Broadcast Digital Broadcasting (ISDB-T), or the like receives digital broadcast.
  • the broadcast receiving unit 111 can be constructed as various broadcast systems suitable for providing broadcast signals as well as the above-described digital broadcast system.
  • the broadcast signal and/or broadcast associated information received via the broadcast receiving unit 111 may be stored in the memory 160 (or other type of storage medium).
  • the mobile communication unit 112 transmits the radio signal to and/or receives a radio signal from at least one of a base station (e.g., an access point, a Node B, etc.), an external terminal, and a server.
  • a base station e.g., an access point, a Node B, etc.
  • Such radio signals may include voice call signals, video call signals, or various types of data transmitted and/or received in accordance with text and/or multimedia messages.
  • the wireless internet unit 113 supports wireless internet access of the mobile terminal.
  • the unit can be internally or externally coupled to the terminal.
  • the wireless internet access technology involved in the unit may include none Line-local area network (Wi-Fi, WLAN, Wireless Local Area Networks), wireless broadband (Wibro), Worldwide Interoperability for Microwave Access (Wimax), High Speed Downlink Packet Access (HSDPA), and the like.
  • the short-range communication unit 114 is a unit for supporting short-range communication.
  • Some examples of short-range communication technologies include Bluetooth, Radio Frequency Identification (RFID), Infrared Data Association (IrDA), Ultra Wideband (UWB), Zigbee, and the like.
  • the location information unit 115 is a unit for checking or acquiring location information of the mobile terminal.
  • a typical example of a location information unit is a Global Positioning System (GPS).
  • GPS Global Positioning System
  • the position information unit 115 calculates distance information and accurate time information from three or more satellites and applies triangulation to the calculated information, thereby accurately calculating three-dimensional current position information according to longitude, latitude, and altitude.
  • the method for calculating position and time information uses three satellites and corrects the calculated position and time information errors by using another satellite. Further, the position information unit 115 can calculate the speed information by continuously calculating the current position information in real time.
  • the A/V input unit 120 is for receiving an audio or video signal.
  • the A/V input unit 120 may include a camera 121 and a microphone 122 that processes image data of still pictures or video obtained by the image capturing device in a video capturing mode or an image capturing mode.
  • the processed image frame can be displayed on the display unit 151.
  • the image frames processed by the camera 121 may be stored in the storage unit 160 (or other storage medium) or transmitted via the communication unit 110, and two or more cameras 121 may be provided according to the configuration of the mobile terminal.
  • the microphone 122 can receive sound (audio data) via a microphone in an operation mode of a telephone call mode, a recording mode, a voice recognition mode, and the like, and can process such sound as audio data.
  • the processed audio (voice) data can be converted to a format output that can be transmitted to the mobile communication base station via the mobile communication unit 112 in the case of a telephone call mode.
  • Microphone 122 can implement various types of noise cancellation (or The algorithm is suppressed to cancel (or suppress) noise or interference generated during the process of receiving and transmitting an audio signal.
  • the user input unit 130 may generate key input data according to a command input by the user to control various operations of the mobile terminal.
  • the user input unit 130 allows the user to input various types of information, and may include a keyboard, a mouse, a touch pad (eg, a touch sensitive component that detects changes in resistance, pressure, capacitance, etc. due to contact), a scroll wheel, a shaker. Rod and so on.
  • a touch panel when the touch panel is superimposed on the display unit 151 in the form of a layer, a touch screen can be formed.
  • the obtaining unit 140 is configured to acquire file conversion logic, and the playing unit 141 is configured to play the first to-be-processed file and the second to-be-processed file obtained based on the source file, where the first to-be-processed file is a source file displayed on the terminal screen.
  • the second to-be-processed file is a right video file of the source file displayed on the screen of the terminal;
  • the first processing unit 142 is configured to process the first to-be-processed file and the second to-be-processed according to the file conversion logic
  • the file is separately decoded by audio and video, and an audio file and an image file respectively corresponding to the first to-be-processed file and the second to-be-processed file are obtained, and the image files corresponding to the first to-be-processed file and the second to-be-processed file are respectively subjected to distortion processing.
  • the second processing unit 143 is configured to receive the recording operation, and provide the distortion processing result to the recording thread for recording, according to the file conversion Logging the distortion processing result and the audio text corresponding to the first to-be-processed file and the second to-be-processed file, respectively Synthesis coding is used directly for the first virtual reality VR object file playback and the second target file.
  • the interface unit 170 serves as an interface through which at least one external device can connect with the mobile terminal 100.
  • the external device may include a wired or wireless headset port, an external power (or battery charger) port, a wired or wireless data port, a memory card port, a port for connecting a device having an identification unit, and an audio input/output. (I/O) port, video I/O port, headphone port, and more.
  • the identification unit may be stored to verify various information used by the user using the mobile terminal 100 and may include a User Identification Module (UIM), a customer identification unit (SIM, Subscriber Identity Module), Universal Subscriber Identity Module (USIM), and so on.
  • UIM User Identification Module
  • SIM Subscriber Identity Module
  • USB Universal Subscriber Identity Module
  • the device having the identification unit may take the form of a smart card, and thus the identification device may be connected to the mobile terminal 100 via a port or other connection device.
  • the interface unit 170 can be configured to receive input from an external device (eg, data information, power, etc.) and transmit the received input to one or more components within the mobile terminal 100 or can be used at the mobile terminal and external device Transfer data between.
  • the interface unit 170 may function as a path through which power is supplied from the base to the mobile terminal 100 or may be used as a transmission of various command signals allowing input from the base to the mobile terminal 100 The path to the terminal.
  • Various command signals or power input from the base can be used as signals for identifying whether the mobile terminal is accurately mounted on the base.
  • Output unit 150 is configured to provide an output signal (eg, an audio signal, a video signal, a vibration signal, etc.) in a visual, audio, and/or tactile manner.
  • the output unit 150 may include a display unit 151, an audio output unit 152, and the like.
  • the display unit 151 can display information processed in the mobile terminal 100.
  • the mobile terminal 100 can display a related user interface (UI) or a graphical user interface (GUI).
  • UI related user interface
  • GUI graphical user interface
  • the display unit 151 may display a captured image and/or a received image, a UI or GUI showing a video or image and related functions, and the like.
  • the display unit 151 can function as an input device and an output device.
  • the display unit 151 may include a Liquid Crystal Display (LCD), a Thin Film Transistor (LCD), an Organic Light-Emitting Diode (OLED) display, a flexible display, and a three-dimensional (3D) At least one of a display or the like.
  • LCD Liquid Crystal Display
  • LCD Thin Film Transistor
  • OLED Organic Light-Emitting Diode
  • a flexible display a three-dimensional (3D) At least one of a display or the like.
  • 3D three-dimensional
  • Some of these displays may be constructed to be transparent to allow the user to view from the outside, which may be referred to as A transparent display, which may be, for example, a transparent organic light emitting diode (TOLED) display or the like.
  • TOLED transparent organic light emitting diode
  • the mobile terminal 100 may include two or more display units (or other display devices), for example, the mobile terminal may include an external display unit (not shown) and an internal display unit (not shown) .
  • the touch screen can be used to detect touch input pressure as well as touch input position and touch input area.
  • the audio output unit 152 may convert audio data received by the communication unit 110 or stored in the memory 160 into audio when the mobile terminal is in a call signal receiving mode, a call mode, a recording mode, a voice recognition mode, a broadcast receiving mode, and the like.
  • the signal is output as a sound.
  • the audio output unit 152 can provide an audio output (eg, a call signal reception sound, a message reception sound, etc.) related to a specific function performed by the mobile terminal 100.
  • the audio output unit 152 may include a speaker, a buzzer, and the like.
  • the storage unit 160 may store a software program or the like that performs processing and control operations performed by the processing unit 180, or may temporarily store data (for example, a phone book, a message, a still image, a video, and the like) that has been output or is to be output. Moreover, the storage unit 160 may store data regarding various manners of vibration and audio signals that are output when a touch is applied to the touch screen.
  • the storage unit 160 may include at least one type of storage medium including a flash memory, a hard disk, a multimedia card, a card type memory (for example, SD or DX memory, etc.), a random access memory (RAM), Static Random Access Memory (SRAM), Read Only Memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), Programmable Read Only Memory (EEPROM) PROM, Programmable Read Only Memory), magnetic memory, magnetic disk, optical disk, and the like.
  • the mobile terminal 100 can cooperate with a network storage device that performs a storage function of the storage unit 160 through a network connection.
  • Processing unit 180 typically controls the overall operation of the mobile terminal. For example, the processing unit 180 Control and processing related to voice calls, data communications, video calls, and the like. As another example, the processing unit 180 can perform a pattern recognition process to recognize a handwriting input or a picture drawing input performed on the touch screen as a character or an image.
  • the power supply unit 190 receives external power or internal power under the control of the processing unit 180 and provides appropriate power required to operate the various components and components.
  • the various embodiments described herein can be implemented in a computer readable medium using, for example, computer software, hardware, or any combination thereof.
  • the embodiments described herein may use an Application Specific Integrated Circuit (ASIC), a Digital Signal Processing (DSP), a Digital Signal Processing Device (DSPD), Programmable Logic Device (PLD), Field Programmable Gate Array (FPGA), processor, controller, microcontroller, microprocessor, electronics designed to perform the functions described herein At least one of the units is implemented, and in some cases, such an implementation may be implemented in processing unit 180.
  • ASIC Application Specific Integrated Circuit
  • DSP Digital Signal Processing
  • DSPD Digital Signal Processing Device
  • PLD Programmable Logic Device
  • FPGA Field Programmable Gate Array
  • the software code can be implemented by a software application (or program) written in any suitable programming language, which can be stored in storage unit 160 and executed by processing unit 180.
  • the specific hardware entity of the storage unit 160 may be a memory
  • a specific hardware entity of the processing unit 180 may be a controller.
  • the mobile terminal has been described in terms of its function.
  • a slide type mobile terminal among various types of mobile terminals such as a folding type, a bar type, a swing type, a slide type mobile terminal, and the like will be described as an example. Therefore, the present invention can be applied to any type of mobile terminal, and is not limited to a slide type mobile terminal.
  • the mobile terminal 100 as shown in FIG. 1 may be configured to operate using a communication system such as a wired and wireless communication system and a satellite-based communication system that transmits data via frames or packets.
  • a communication system such as a wired and wireless communication system and a satellite-based communication system that transmits data via frames or packets.
  • a communication system in which a mobile terminal is operable according to an embodiment of the present invention will now be described with reference to FIG.
  • Such communication systems may use different air interfaces and/or physical layers.
  • the air interface used by the communication system includes, for example, Frequency Division Multiple Access (FDMA), Time Division Multiple Access (TDMA), Code Division Multiple Access (CDMA), and General Purpose Code Division Multiple Access (CDMA).
  • FDMA Frequency Division Multiple Access
  • TDMA Time Division Multiple Access
  • CDMA Code Division Multiple Access
  • CDMA General Purpose Code Division Multiple Access
  • UMTS Universal Mobile Telecommunications System
  • LTE Long Term Evolution
  • GSM Global System for Mobile Communications
  • the following description relates to a CDMA communication system, but such teachings are equally applicable to other types of systems.
  • the CDMA wireless communication system may include a plurality of mobile terminals 100, a plurality of base stations (BS) 270, a base station controller (BSC) 275, and a mobile switching center (MSC) 280.
  • the MSC 280 is configured to interface with a Public Switched Telephone Network (PSTN) 290.
  • PSTN Public Switched Telephone Network
  • the MSC 280 is also configured to interface with a BSC 275 that can be coupled to the BS 270 via a backhaul line.
  • the backhaul line can be constructed in accordance with any of a number of known interfaces including, for example, E1/T1, ATM, IP, PPP, Frame Relay, HDSL, ADSL, or xDSL. It will be appreciated that the system as shown in FIG. 2 can include multiple BSCs 275.
  • Each BS 270 can serve one or more partitions (or regions), with each partition covered by a multi-directional antenna or an antenna pointing in a particular direction radially away from the BS 270. Alternatively, each partition may be covered by two or more antennas for diversity reception. Each BS 270 can be configured to support multiple frequency allocations, and each frequency allocation has a particular frequency spectrum (eg, 1.25 MHz, 5 MHz, etc.).
  • BS 270 may also be referred to as a Base Transceiver Subsystem (BTS) or other equivalent terminology.
  • BTS Base Transceiver Subsystem
  • the term "base station” can be used to generally mean a single BSC 275 and at least one BS 270.
  • a base station can also be referred to as a "cell station.”
  • each partition of a particular BS 270 may be referred to as multiple cellular stations.
  • a broadcast transmitter (BT, Broadcast Transmitter) 295 transmits a broadcast signal to the mobile terminal 100 operating within the system.
  • a broadcast receiving unit 111 as shown in FIG. 1 is provided at the mobile terminal 100 to receive a broadcast signal transmitted by the BT 295.
  • several satellites 300 are shown, for example GPS satellites 300 may be employed. The satellite 300 helps locate at least one of the plurality of mobile terminals 100.
  • a plurality of satellites 300 are depicted, but it is understood that useful positioning information can be obtained using any number of satellites.
  • the location information unit 115 as shown in FIG. 1 is typically configured to cooperate with the satellite 300 to obtain desired positioning information. Instead of GPS tracking technology or in addition to GPS tracking technology, other techniques that can track the location of the mobile terminal can be used. Additionally, at least one GPS satellite 300 can selectively or additionally process satellite DMB transmissions.
  • BS 270 receives reverse link signals from various mobile terminals 100.
  • Mobile terminal 100 typically participates in calls, messaging, and other types of communications.
  • Each reverse link signal received by a particular base station is processed within a particular BS 270.
  • the obtained data is forwarded to the relevant BSC 275.
  • the BSC 275 provides call resource allocation and coordinated mobility management functions including a soft handoff procedure between the BSs 270.
  • the BSC 275 also routes the received data to the MSC 280, which provides additional routing services for interfacing with the PSTN 290.
  • PSTN 290 interfaces with MSC 280
  • MSC 280 interfaces with BSC 275
  • BSC 275 controls BS 270 accordingly to transmit forward link signals to mobile terminal 100.
  • the mobile communication unit 112 of the communication unit 110 in the mobile terminal accesses the mobile communication based on necessary data (including user identification information and authentication information) of the mobile communication network (such as 2G/3G/4G mobile communication network) built in the mobile terminal.
  • the network transmits mobile communication data (including uplink mobile communication data and downlink) for mobile terminal users such as web browsing and network multimedia broadcasting. Mobile communication data).
  • the wireless internet unit 113 of the communication unit 110 implements a function of a wireless hotspot by operating a related protocol function of the wireless hotspot, and the wireless hotspot supports access of a plurality of mobile terminals (any mobile terminal other than the mobile terminal) by multiplexing the mobile communication unit 112.
  • the mobile communication connection with the mobile communication network transmits mobile communication data (including uplink mobile communication data and downlink mobile communication data) for the mobile terminal user's web browsing, network multimedia playback, etc., since the mobile terminal is substantially multiplexed
  • the mobile communication connection between the mobile terminal and the communication network transmits the mobile communication data, so the traffic of the mobile communication data consumed by the mobile terminal is included in the communication tariff of the mobile terminal by the charging entity on the communication network side, thereby consuming the subscription used by the mobile terminal.
  • the data traffic of mobile communication data included in the communication tariff included in the communication tariff.
  • FIG. 3 is a schematic diagram of hardware entities of each party performing information interaction in the embodiment of the present invention, and FIG. 3 includes: a terminal device 1 and a server 2.
  • the terminal device 1 is composed of terminal devices 11-14, and the terminal device performs information interaction with the server through a wired network or a wireless network.
  • Terminal equipment includes mobile phones, desktops, PCs, all-in-ones, and the like.
  • the terminal may initiate a download request to the server to obtain file conversion logic, where the file conversion logic may be in the form of a general installation package, and the universal installation package is a general installation obtained after being processed by the windows platform. The package does not require a mobile phone to install a special APP.
  • the general installation package can be directly used for playback and distortion processing.
  • the operation object for the common installation package Based on the left and right format files of the same source file (such as a video file), the operation object for the common installation package not only achieves better distortion processing effect, but also does not need to be processed for the source file as in the prior art.
  • the processing results are screened to VR glasses for playback, which is more convenient and faster.
  • the obtained distortion processing result is stored in the image file format in the image file for file sharing, so that the copying operation is not required, and only the required file is obtained by sharing during recording, which not only improves efficiency, but also can improve efficiency. Avoid excessive CPU and memory usage.
  • the distortion processing result is provided to the recording thread for recording, and finally
  • the code is synthesized into a first object file and a second object file that are directly used for VR playback.
  • the file conversion logic such as a universal installation package
  • the terminal obtains the file conversion logic (such as a general installation package), and directly uses the file conversion logic (such as a universal installation package). Just do it.
  • processing logic 10 executed by the terminal device is as shown in FIG. 3, and the processing logic 10 includes:
  • S1 Obtain file conversion logic, where the file conversion logic is logic generated on the server side according to the play plug-in, SDK, and recording plug-in loaded in the specified 3D running environment; S2, and the file-based conversion logic is played based on the source file.
  • the left video file and the right video file the left video file is a file that is played based on the source file and displayed on the left side of the terminal screen
  • the right video file is a file that is played based on the source file and displayed on the right side of the terminal screen;
  • S3 according to the The file conversion logic performs audio and video decoding corresponding to the left video file and the right video file to obtain an audio file and an image file respectively;
  • S4 the image file is distorted, and the obtained image distortion processing result is stored in the image file format in the image memory.
  • FIG. 3 is only a system architecture example for implementing the embodiment of the present invention.
  • the embodiment of the present invention is not limited to the system structure described in FIG. 3 above. According to the system architecture described in FIG. 3, various embodiments of the method of the present invention are proposed. .
  • a file processing method includes: acquiring file conversion logic, where the file conversion logic is a playback plug-in and software running on a server side according to a specified 3D running environment.
  • the file conversion logic can be used for playback, audio and video decomposition, distortion, target file recording and other operations.
  • the file conversion logic is in the form of an SDK installation package.
  • the installation package is a general installation package obtained after being processed by the windows platform.
  • the terminal does not need to install a special APP, and any terminal only needs to download the universal installation package to the terminal local
  • the installation package can be directly used for playback, distortion processing, etc., that is, the source file is processed by the universal installation package to obtain a target file, and the target file can be further shared to the network.
  • the first to-be-processed file and the second to-be-processed file obtained based on the source file are first played, the first to-be-processed file.
  • the second to-be-processed file is the right video file 102 of the source file displayed on the terminal screen.
  • the image files respectively corresponding to the first to-be-processed file and the second to-be-processed file are subjected to distortion processing, and the obtained distortion processing result is stored in the image memory in the image file format for file sharing 104.
  • the composition is a first object file and a second object file 105 that are directly used for virtual reality VR playback.
  • VR technology is a computer simulation system that can create and experience virtual worlds. It uses a computer to generate a simulation environment, providing interactive 3D dynamic visual and physical behavior simulation of multi-source information fusion, so as to immerse users in In the simulation environment.
  • VR mainly includes simulation environment, perception, natural skills and sensing equipment.
  • the simulation environment is a computer-generated, real-time, dynamic, three-dimensional, realistic image. Perception means that the ideal VR should have the perception that everyone has.
  • there are also perceptions such as hearing, touch, force, and motion, and even smell and taste, also known as multi-perception.
  • Natural skills are people The head rotation, eyes, gestures, or other human behavioral actions, the computer processes the data that is appropriate for the participant's actions, and responds to the user's input in real time, and feeds back to the user's facial features.
  • a sensing device is a three-dimensional interactive device.
  • VR glasses With the development of VR, major VR vendors have launched their own VR glasses. For this new human-computer interaction method through VR glasses, due to its more natural and more stereoscopic environment simulation, combined with the increasing popularity of mobile terminals. Mobile VR has entered the life of mass consumers faster. VR glasses not only support panoramic images, but also support 3D images. However, the current 3D video playback requires a special APP to be converted, which is not only troublesome, but also does not support 3D playback of the webpage. The reason is that the APP installed on the terminal is usually a video player, and there is no web browsing function, so 3D video playback of web pages is not supported. At present, most 3D videos are left and right formats. This format cannot be viewed directly in VR glasses. Because VR glasses need to project the distorted image to the glasses, there will be 3D stereo. effect.
  • processing logic based on file transcoding can be played through any built-in player of the mobile phone terminal, and the user can conveniently convert the file.
  • file-based transcoding processing logic such as file conversion logic
  • the generated file will be shared to the web page with one click and played on the web page.
  • the prior art processes the original source file, and finally, it also needs to be projected into the VR glasses for screen display, and the format is limited, and the present invention directly
  • the source file is processed based on the left and right format videos required by the VR glasses display, and is not displayed in the format. Subsequently, the VR can directly watch, and no screen is required.
  • the distorted image is shared by the memory, avoiding multiple copies during the recording process, avoiding excessive CPU and memory usage, thereby affecting the overall processing efficiency, realizing efficient and delay-free image transmission, and saving CPU and memory resources. .
  • VLC refers to video decoding and audio decoding
  • Cardboard refers to image distortion
  • OBS refers to encoding and compressing the distorted image and audio.
  • VLC refers to video decoding and audio decoding
  • Cardboard refers to image distortion
  • OBS refers to encoding and compressing the distorted image and audio.
  • VLC refers to video decoding and audio decoding
  • Cardboard refers to image distortion
  • OBS refers to encoding and compressing the distorted image and audio.
  • VLC is a full-format full-format player. It is a playback plug-in for VLC in Unity. VLC can decode the audio and image of 3D video separately, and then, for processing. The key image is distorted.
  • OBS is a recording tool for the Windows platform, and its own plug-in can record for the window. Since it is necessary to perform multiple copy operations for window recording, the efficiency is very low, and the CPU and memory are high. This program optimizes here, two plug-ins are written separately, one for Unity, and the distorted images are shared by video memory. Another plugin is used to get the image from the video memory in the OBS and hand it directly to the recording thread. In the image copying process, since the operations are all performed in the video memory, the copy operation basically consumes no memory and consumes no CPU. OBS will compress the encoded image and sound, and finally save it as the target video file.
  • a file processing method includes: acquiring file conversion logic, where the file conversion logic is a play plugin and SDK that are loaded and run according to the specified 3D running environment on the server side. And recording the logic 201 generated by the plug-in, in the subsequent specific processing, the file conversion logic can be used for playing, audio and video decomposition, distortion, target file recording and the like.
  • the file conversion logic is in the form of an SDK installation package,
  • the installation package is a general installation package obtained after processing by the WINDOW platform. It does not require a terminal to install a special APP. Any terminal only needs to download the universal installation package to the terminal local, and then directly use the installation package to realize playback, distortion processing, etc.
  • the source file is processed by the general installation package to obtain the target file, and the target file can be further shared on the network.
  • the first to-be-processed file and the second to-be-processed file obtained based on the source file are first played, the first to-be-processed file.
  • the second to-be-processed file is the right video file 202 of the source file displayed on the terminal screen.
  • Audio file and image file 203 And performing audio and video decoding on the first to-be-processed file and the second to-be-processed file respectively according to the file conversion logic, to obtain a first audio file and a first image file corresponding to the first to-be-processed file, and corresponding to the second to-be-processed file.
  • the second audio file and the second image file are respectively subjected to audio and video decoding according to the file conversion logic, and respectively obtained to correspond to the first to-be-processed file and the second to-be-processed file respectively. Audio file and image file 203.
  • the first information 204 related to the display of the terminal screen where the first information includes: a screen aspect ratio At least one of the information in the resolution.
  • the second information 205 related to the performance of the terminal includes at least one of a model of the terminal and a performance indicator of the terminal processor.
  • the related fourth image file 206 stores the third image file and the fourth image file in the video memory for file sharing 207, so that the obtained distortion processing result is stored in the image file format in the image file for file sharing.
  • Receiving a recording operation when the third image file and the fourth image file are provided to a recording thread for recording, encoding the first audio file and the third image file into a code according to the file conversion logic
  • An object file, the second audio file and the fourth image file are encoded into a second object file, and the first object file and the first object file
  • the two object files are used directly for VR playback 208.
  • VR technology is a computer simulation system that can create and experience virtual worlds. It uses a computer to generate a simulation environment, providing interactive 3D dynamic visual and physical behavior simulation of multi-source information fusion, so as to immerse users in In the simulation environment.
  • VR mainly includes simulation environment, perception, natural skills and sensing equipment.
  • the simulation environment is a computer-generated, real-time, dynamic, three-dimensional, realistic image. Perception means that the ideal VR should have the perception that everyone has.
  • there are also perceptions such as hearing, touch, force, and motion, and even smell and taste, also known as multi-perception.
  • Natural skills refer to the rotation of the person's head, eyes, gestures, or other human behaviors.
  • the computer processes the data that is appropriate to the actions of the participants, responds to the user's input in real time, and feeds back to the user's facial features. .
  • a sensing device is a three-dimensional interactive device.
  • the first object file and the second object file are combined with the VR glasses, and the left and right format videos required for directly displaying the source file through the VR glasses are
  • the basic processing therefore, is not limited by the format display and application scope, VR glasses can directly use the above target file to view on any terminal or webpage, without screen processing, improve processing efficiency, and make the file format for each
  • the range of terminals and applications is compatible.
  • the distortion of the terminal is adjusted by using the corresponding distortion parameter in advance in the embodiment of the present invention. To ensure that the final playback is clear and undistorted.
  • the distortion parameter is related to the aspect ratio of the screen, the resolution, and the type of the terminal such as the mobile terminal.
  • the prior art is to process the original source file, and finally, it needs to be projected into the VR glasses to display the screen, which has a limitation on the format.
  • the left and right formats required for directly displaying the source file through the VR glasses are displayed.
  • the video is processed on the basis of the format, and is not displayed in the format.
  • the VR can directly watch, and no screen is required.
  • the distorted image is shared by the memory, avoiding Multiple copies during the recording process, avoiding excessive CPU and memory usage, thus affecting the overall processing efficiency, achieving efficient, delay-free image transmission, saving CPU and memory resources.
  • VLC refers to video decoding and audio decoding
  • Cardboard refers to image distortion
  • OBS refers to encoding and compressing the distorted image and audio.
  • VLC refers to video decoding and audio decoding
  • Cardboard refers to image distortion
  • OBS refers to encoding and compressing the distorted image and audio.
  • VLC refers to video decoding and audio decoding
  • Cardboard refers to image distortion
  • OBS refers to encoding and compressing the distorted image and audio.
  • VLC is a full-format full-format player. It is a playback plug-in for VLC in Unity. VLC can decode the audio and image of 3D video separately, and then, for processing. The key image is distorted.
  • OBS is a recording tool for the Windows platform, and its own plug-in can record for the window. Since it is necessary to perform multiple copy operations for window recording, the efficiency is very low, and the CPU and memory are high. This program optimizes here, two plug-ins are written separately, one for Unity, and the distorted images are shared by video memory. Another plugin is used to get the image from the video memory in the OBS and hand it directly to the recording thread. In the image copying process, since the operations are all performed in the video memory, the copy operation basically consumes no memory and consumes no CPU. OBS will compress the encoded image and sound, and finally save it as the target video file.
  • the method further includes: receiving a file sharing operation, and sharing the first object file and the second object file to the target user according to a target user address. Or, the first object file and the second object file are shared according to the webpage connection address to the corresponding webpage.
  • the method further includes: storing the first object file and the second object file in a video file of the same format. Storage.
  • the terminal may initiate a download request to the server to obtain file conversion logic, where the file conversion logic may be in the form of a general installation package, and the universal installation package is processed by the windows platform.
  • the general installation package you do not need to install a special APP on the mobile phone.
  • the operation object for the common installation package not only achieves better distortion processing effect, but also does not need to be processed for the source file as in the prior art.
  • the processing results are screened to VR glasses for playback, which is more convenient and faster.
  • the obtained distortion processing result is stored in the image file format in the image file for file sharing, so that the copying operation is not required, and only the required file is obtained by sharing during recording, which not only improves efficiency, but also can improve efficiency. Avoid excessive CPU and memory usage.
  • the distortion processing result is provided to the recording thread for recording, and finally the encoding is synthesized into the first target file and the second target file directly used for VR playback.
  • the file conversion logic such as a universal installation package
  • the terminal obtains the file conversion logic (such as a general installation package), and directly uses the file conversion logic (such as a universal installation package). Just do it.
  • the terminal 41 includes: an obtaining unit 411, configured to acquire file conversion logic; a playing unit 412, configured to play a first to-be-processed file and a second to-be-processed file obtained based on the source file, where the first to-be-processed file is displayed a left video file of the source file on the terminal screen, the second to-be-processed file is a right video file of the source file displayed on the screen of the terminal; the first processing unit 413 is configured to process the first to be processed according to the file conversion logic
  • the file and the second to-be-processed file are respectively decoded by audio and video, and the audio file and the image file respectively corresponding to the first to-be-processed file and the second to-be-processed file are respectively obtained, which respectively correspond to the first to-be-processed file and the second to-be-processed file.
  • the image file is subjected to distortion processing, and the obtained distortion processing result is stored in the image file format in the image file for file sharing; the second processing The unit 414 is configured to receive a recording operation, and when the distortion processing result is provided to the recording thread for recording, the distortion processing result is corresponding to the first to-be-processed file and the second to-be-processed file respectively according to the file conversion logic.
  • the audio file is encoded and synthesized into a first object file and a second object file that are directly used for virtual reality VR playback.
  • the server 42 includes: a generating unit 421, configured to generate file conversion logic according to a playback plug-in (such as VLC), an SDK (such as a Cardboard SDK), and a recording plug-in (such as OBS) loaded and loaded in a specified 3D running environment (such as Unity);
  • the unit 422 is configured to feed back the file conversion logic to the terminal in response to the download request initiated by the terminal.
  • processing logic based on file transcoding can be played through any built-in player of the mobile phone terminal, and the user can conveniently convert the file.
  • file-based transcoding processing logic such as file conversion logic
  • the generated file will be shared to the web page with one click and played on the web page.
  • the prior art processes the original source file, and finally, it also needs to be projected into the VR glasses for screen display, and the format is limited, and the present invention directly
  • the source file is processed based on the left and right format videos required by the VR glasses display, and is not displayed in the format. Subsequently, the VR can directly watch, and no screen is required.
  • the distorted image is shared by the memory, avoiding multiple copies during the recording process, avoiding excessive CPU and memory usage, thereby affecting the overall processing efficiency, realizing efficient and delay-free image transmission, and saving CPU and memory resources. .
  • VLC refers to video decoding, audio decoding
  • Cardboard Refers to image distortion
  • OBS refers to encoding and compressing the distorted image and audio to save the file.
  • VLC refers to video decoding, audio decoding
  • Cardboard Refers to image distortion
  • OBS refers to encoding and compressing the distorted image and audio to save the file.
  • VLC refers to video decoding, audio decoding
  • Cardboard Refers to image distortion
  • OBS refers to encoding and compressing the distorted image and audio to save the file.
  • the role of the Cardboard SDK is to make distortions.
  • the Cardboard SDK can adapt to the distortion parameters according to the aspect ratio and resolution of the screen, as well as preset a variety of mobile phone models.
  • the Cardboard SDK itself is a very comprehensive development package. This program only uses the distortion function, so it needs to be cropped. There is no player in any video format in Unity.
  • VLC is a full-format full-format player. It is a playback plug-in for VLC in Unity. VLC can decode the audio and image of 3D video separately, and then, for processing. The key image is distorted.
  • OBS is a recording tool for the Windows platform, and its own plug-in can record for the window. Since it is necessary to perform multiple copy operations for window recording, the efficiency is very low, and the CPU and memory are high. This program optimizes here, two plug-ins are written separately, one for Unity, and the distorted images are shared by video memory. Another plugin is used to get the image from the video memory in the OBS and hand it directly to the recording thread. In the image copying process, since the operations are all performed in the video memory, the copy operation basically consumes no memory and consumes no CPU. OBS will compress the encoded image and sound, and finally save it as the target video file.
  • the first processing unit in the terminal is further configured to perform audio and video decoding on the first to-be-processed file and the second to-be-processed file according to the file conversion logic, to obtain a corresponding a first audio file and a first image file of the first to-be-processed file, and a second audio file and a second image file corresponding to the second to-be-processed file.
  • the first processing unit in the terminal is further configured to: acquire first information related to display of the terminal screen; and acquire second information related to performance of the terminal itself; Distorting the first image file and the second image file with a message and/or the second information to obtain a third image file associated with the first image file and a fourth image file associated with the second image file
  • the image file stores the third image file and the fourth image file in a video memory for file sharing.
  • the first information includes: a screen width At least one of high ratio and resolution; the second information includes at least one of a model number of the terminal and a performance indicator of the terminal processor.
  • the second processing unit in the terminal is further configured to: receive a recording operation, and provide the third image file and the fourth image file to a recording thread for recording, And encoding the first audio file and the third image file into a first target file according to the file conversion logic, and synthesizing the second audio file and the fourth image file into a second target file.
  • the terminal further includes: a sharing unit, configured to receive a file sharing operation, and share the first object file and the second object file with the target user address according to the target user address. Or, the first object file and the second object file are shared according to the webpage connection address to the corresponding webpage.
  • the terminal further includes: a storage unit, configured to store the first object file and the second object file in a video file of the same format.
  • the generating unit in the server is further configured to: import the SDK in the specified 3D running environment; load and run a playing plugin in the specified 3D running environment, The playing plug-in performs audio and video decoding on the first to-be-processed file and the second to-be-processed file obtained based on the source file, to obtain an audio file and an image file respectively corresponding to the first to-be-processed file and the second to-be-processed file;
  • the image file is distorted, the image file corresponding to the first to-be-processed file and the second to-be-processed file is distorted by the SDK; and the distortion processing is obtained before the recording thread is opened by the recording plug-in for recording.
  • the result is stored in the video file format in the image file format for file sharing.
  • the recording plug-in in the server includes: a first plug-in related to the specified 3D running environment memory and a second plug-in responsive to the recording operation;
  • the server includes: a communication unit, And communicating with the display memory by the first plug-in, extracting the distortion processing result from the video memory; and recording unit for responding to the recording operation,
  • the second plug-in communicates with the first plug-in, and provides the distortion processing result to the recording thread for recording, and the distortion processing result and the audio corresponding to the first to-be-processed file and the second to-be-processed file respectively
  • the file encoding is synthesized into a first object file and a second object file directly used for virtual reality VR playback.
  • the processor for data processing may be implemented by using a microprocessor, a central processing unit (CPU), a DSP, or an FPGA when performing processing; for an storage medium, including an operation instruction,
  • the operation instruction may be computer executable code, and the steps in the flow of the information processing method of the embodiment of the present invention are implemented by the operation instruction.
  • VR glasses not only support panoramic pictures, but also support 3D pictures.
  • the current 3D video playback requires an APP to be installed, which is not only troublesome, but also does not support 3D playback of the webpage.
  • FIG. 7 the process of using a video containing video content, game content, and video player for VR playback is performed. In this process, after the audio and video are decoded by the 3D video source, the image decoded by the video is distorted. The audio is finally played through the speaker, and the distorted image is projected onto the VR screen for playback.
  • the processing method based on this process is the processing of the 3D video source, and this format cannot be directly viewed in the VR glasses, because the VR glasses need to project the distorted image to Glasses will have a 3D effect. Need to install a dedicated app, or support web page playback, these issues are not in line with mass consumers Demand.
  • the VLC+Unity3D+CardboardSDK+OBS combination method can be used to perform distortion transformation on the left and right format videos and save the video files in the same format, for example, any video that can be easily used in the mobile terminal. Player, web page to play. Bring VR glasses and you can experience 3D video effects.
  • the VLC+Unity3D+CardboardSDK+OBS combination is used to generate file conversion logic and run on the Windows platform. Then, according to the download request of the terminal, the file conversion logic is fed back to the terminal for direct use, so as to implement audio, video decoding, distortion processing, encoding synthesis, recording and saving, and webpage sharing functions for the left and right format files of the source file.
  • the Cardboard SDK plugin package is imported.
  • the role of the Cardboard SDK is to make distortions.
  • the Cardboard SDK can adapt to the distortion parameters according to the aspect ratio and resolution of the screen, as well as preset a variety of mobile phone models.
  • the Cardboard SDK itself is a very comprehensive development package. This program only uses the distortion function, so it needs to be cropped.
  • VLC is a full-format full-format player.
  • the images before and after the distortion are shown in Fig.
  • Fig. 8 is the image before the distortion
  • Fig. 9 is the image after the distortion.
  • OBS is a recording tool under the Windows platform, and its own plug-in can record for the window. Since it is necessary to perform multiple copy operations for window recording, the efficiency is very low, and the CPU and memory are high.
  • This program optimizes here two plug-ins are written separately, one for Unity, and the distorted images are shared by video memory.
  • Another plugin is used to get the image from the video memory in the OBS and hand it directly to the recording thread.
  • the copy operation basically consumes no memory and consumes no CPU.
  • OBS will compress the encoded image and sound, and finally save it as a new video file.
  • VLC refers to: video decoding
  • audio decoding Cardboard refers to: image distortion
  • OBS refers to: encoding and compressing the distorted image and audio, and saving the file.
  • FIG. 10 is a file conversion process according to an embodiment of the present invention.
  • a left-right format 3D source video derived from a 3D video source (such as a 3D video file) is sounded based on the file conversion logic.
  • the playback plug-in VLC is used to play the left and right format 3D source video, and the left and right format 3D source video are respectively separated into audio and video, and the image decoded by the video decoding is distorted and then the image is distorted.
  • the processing result is put into the video memory for sharing.
  • the image distortion processing result shared by the audio file and the video memory is obtained, and the two are encoded and synthesized, and a new video file is obtained, which is directly on the VR.
  • the new video file is played, and the interaction process between the distortion process and the OBS recording process is as shown in FIG. 11.
  • the deformed video needs to be distorted to eliminate the influence of the deformation.
  • the original video file is obtained, the playback plug-in VLC is decoded, the video frame and the audio frame are separated (the audio and video are separately decoded for the left and right format 3D source video), and the audio frame is placed in the virtual sound card, only Distort the video frame.
  • the distortion processing result obtained by distorting the video frame is put into the video memory, and the benefit of sharing in the video memory is that the distortion processing result is stored in the image file format in the image memory for file sharing, thereby eliminating the need for multiple copy operations and only recording When you go to the share to get the files you need, you can not only improve efficiency, but also avoid excessive CPU and memory usage.
  • the distortion processing result of the video shared in the audio and the video memory is captured, and after receiving the recording operation, the distortion processing result is provided to the recording thread for recording, and the two are encoded. Synthesize and maintain, get a new video file, and play the new video file directly on the VR.
  • VR Capture is a recording sharing tool for VR PC glasses. It converts the left and right format videos recorded by VR Capture, and performs distortion conversion to generate new videos. file. VR Capture will generate After the file, share it with one click.
  • the program also provides an interface DLL to directly call the start conversion and end conversion. It also provides a callback for the conversion progress.
  • a computer storage medium may be a memory including a computer program, which may be executed by a processor of the data processing device to complete the steps of the method in the foregoing embodiment.
  • the computer storage medium may be a memory such as FRAM, ROM, PROM, EPROM, EEPROM, Flash Memory, magnetic surface memory, optical disk, or CD-ROM; or may be various devices including one or any combination of the above memories, such as a mobile phone. , computers, tablet devices, personal digital assistants, etc.
  • the computer readable storage medium having stored thereon a computer program that, when executed by the processor, executes the following file processing method.
  • the computer readable storage medium when the computer readable storage medium is located on the terminal side, the following content is included:
  • Obtaining file conversion logic which is logic generated on the server side according to a play plugin, a software development kit SDK, and a recording plugin loaded and loaded in a specified 3D runtime environment;
  • the audio file of the two to-be-processed files is encoded into a first object file and a second object file that are directly used for virtual reality VR playback.
  • the computer program when executed by the processor, it also executes:
  • the computer program when executed by the processor, it also executes:
  • the related fourth image file stores the third image file and the fourth image file in a video memory for file sharing.
  • the first information includes at least one of a screen aspect ratio and a resolution.
  • the second information includes at least one of a model of the terminal and a performance indicator of the terminal processor.
  • the computer program when executed by the processor, it also executes:
  • the computer program when executed by the processor, it also executes:
  • the computer program when executed by the processor, it also executes:
  • the first object file and the second object file are stored in a video file of the same format.
  • the computer program when executed by the processor, it also executes:
  • the image file corresponding to the first to-be-processed file and the second to-be-processed file is distorted by the SDK;
  • the obtained distortion processing result is stored in the image memory in the image file format for file sharing.
  • the recording plug-in includes: a first plug-in associated with the specified 3D runtime environment and a second plug-in responsive to the recording operation, the computer program being executed by the processor, further executing:
  • the second plug-in communicates with the first plug-in, and provides the distortion processing result to the recording thread for recording, and the distortion processing result is respectively corresponding to the first to-be-processed file and the second to-be-processed
  • the audio file of the file is encoded and synthesized for direct use in virtual reality.
  • the disclosed apparatus and method may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner such as: multiple units or components may be combined, or Can be integrated into another system, or some features can be ignored or not executed.
  • the coupling, or direct coupling, or communication connection of the components shown or discussed may be indirect coupling or communication connection through some interfaces, devices or units, and may be electrical, mechanical or other forms. of.
  • the units described above as separate components may or may not be physically separated, and the components displayed as the unit may or may not be physical units, that is, may be located in one place or distributed to multiple network units; Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may be separately used as one unit, or two or more units may be integrated into one unit;
  • the unit can be implemented in the form of hardware or in the form of hardware plus software functional units.
  • the foregoing program may be stored in a computer readable storage medium, and the program is executed when executed.
  • the foregoing steps include the steps of the foregoing method embodiments; and the foregoing storage medium includes: a removable storage device, a ROM, a RAM, a magnetic disk, or an optical disk, and the like, which can store program codes.
  • the above-described integrated unit of the present invention may be stored in a computer readable storage medium if it is implemented in the form of a software functional unit and sold or used as a standalone product.
  • the technical solution of the embodiments of the present invention may be embodied in the form of a software product in essence or in the form of a software product stored in a software product.
  • the storage medium includes instructions for causing a computer device (which may be a personal computer, server, or network device, etc.) to perform all or part of the methods described in various embodiments of the present invention.
  • the foregoing storage medium includes various media that can store program codes, such as a mobile storage device, a ROM, a RAM, a magnetic disk, or an optical disk.
  • the file conversion logic is used as a general processing mechanism, and a series of processes such as audio and video decoding and distortion processing can be performed, thereby adapting any source file and terminal, and no special application needs to be installed on the terminal for processing. Achieve better distortion processing.
  • the obtained distortion processing result is stored in the image file format in the image file for file sharing, so that it is not necessary to copy the operation multiple times, and only needs to obtain the required file during sharing, which not only improves efficiency, but also avoids excessive occupation. CPU and memory.
  • the distortion processing result is provided to the recording thread for recording, and finally the encoding is synthesized into an object file directly used for VR playback.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • Telephone Function (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Telephonic Communication Services (AREA)

Abstract

本发明公开了一种文件处理方法、终端及服务器、计算机存储介质,其中,所述方法包括:获取文件转换逻辑,所述文件转换逻辑为在服务器侧根据在指定3D运行环境中加载运行的播放插件、软件开发工具包(SDK)和录制插件所生成的逻辑;播放基于源文件得到的第一待处理文件和第二待处理文件,根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享;接收录制操作,将所述畸变处理结果提供给录制线程进行录制、编码合成为直接用于虚拟现实(VR)播放的第一目标文件和第二目标文件。

Description

一种文件处理方法、终端及服务器、计算机存储介质
相关申请的交叉引用
本申请基于申请号为201611009068.1、申请日为2016年11月16日的中国专利申请提出,并要求该中国专利申请的优先权,该中国专利申请的全部内容在此引入本申请作为参考。
技术领域
本发明涉及转换技术,尤其涉及一种文件处理方法、终端及服务器、计算机存储介质。
背景技术
随着人机交互技术的发展,智能终端的大量普及,信息的传输和交互方式越来越便捷和新颖。基于虚拟现实(VR,Virtual Reality)信息传输和交互的场景中,VR眼镜是比较常见的,VR眼镜不仅支持全景的画面,而且还支持3D的画面。然而,要支持3D视频的播放,需要安装专门的应用(APP)进行处理才可以,这会增加额外的处理成本,处理效率低下,不但麻烦,而且处理得到的文件格式还不支持网页上的3D视频播放,应用范围非常局限,处理得到的图像文件与VR眼镜相结合时,是需要把畸变后图像文件投射到眼镜上,才会有3D立体效果,也就是说,不能直接在VR眼镜中正常观看,根源还是因为文件格式带来的应用范围限制。然而,相关技术中,对于这些问题,尚无有效解决方案。
发明内容
有鉴于此,本发明实施例提供了一种文件处理方法、终端及服务器、 计算机存储介质,至少解决了现有技术存在的问题。
本发明实施例的技术方案是这样实现的:
本发明实施例的一种文件处理方法,所述方法包括:
获取文件转换逻辑,所述文件转换逻辑为在服务器侧根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件所生成的逻辑;
播放基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件;
根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件,将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享;
接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
本发明实施例的一种终端,所述终端包括:
获取单元,用于获取文件转换逻辑,所述文件转换逻辑为在服务器侧根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件所生成的逻辑;
播放单元,用于播放基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件;
第一处理单元,用于根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件,将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享;
第二处理单元,用于接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
本发明实施例的一种文件处理方法,所述方法包括:
根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件生成文件转换逻辑;
响应终端发起的下载请求,反馈所述文件转换逻辑给终端。
本发明实施例的一种服务器,所述服务器包括:
生成单元,用于根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件生成文件转换逻辑;
反馈单元,用于响应终端发起的下载请求,反馈所述文件转换逻辑给终端。
本发明实施例的一种计算机存储介质,所述计算机存储介质中存储有计算机可执行指令,所述计算机可执行指令用于执行上述任一项的文件处理方法。
本发明实施例的一种文件处理方法,应用于终端,终端包括有一个或多个处理器以及存储器,以及一个或一个以上的程序,其中,所述一个或一个以上的程序存储于存储器中,所述程序可以包括一个或一个以上的每一个对应于一组指令的模块,所述一个或多个处理器被配置为执行指令; 所述方法包括:
获取文件转换逻辑,所述文件转换逻辑为在服务器侧根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件所生成的逻辑;
播放基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件;
根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件,将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享;
接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
本发明实施例的一种文件处理方法,应用于服务器,服务器包括有一个或多个处理器以及存储器,以及一个或一个以上的程序,其中,所述一个或一个以上的程序存储于存储器中,所述程序可以包括一个或一个以上的每一个对应于一组指令的模块,所述一个或多个处理器被配置为执行指令;所述方法包括:
根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件生成文件转换逻辑;
响应终端发起的下载请求,反馈所述文件转换逻辑给终端。
采用本发明实施例,一方面,以文件转换逻辑作为通用的处理机制, 可以进行音视频解码、畸变处理等一系列处理,从而自适应任何源文件和终端,不需要在终端上安装专门的应用进行处理。而且,该处理所针对的对象是基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件,也就是说,是针对源文件的左/右格式进行处理,这样会达到更好的畸变处理效果,也不需要如现有技术一样后续将源文件处理结果分别投屏到VR上播放。另一方面,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享,从而无需多次拷贝操作,只需要录制时去共享中获取所需文件即可,不仅可以提高效率,而且可以避免过多占用CPU和内存。在接收录制操作后,将所述畸变处理结果提供给录制线程进行录制,最后编码合成为直接用于VR播放的第一目标文件和第二目标文件。
附图说明
图1为实现本发明各个实施例的移动终端一个可选的硬件结构示意图;
图2为如图1所示的移动终端的通信***示意图;
图3为本发明实施例中进行信息交互的各方硬件实体的示意图;
图4为本发明实施例一方法实现流程的示意图;
图5为本发明实施例又一方法实现流程的示意图;
图6为本发明实施例一***架构的示意图;
图7为采用专门APP进行文件转换的示意图;
图8-9为一图像畸变前、畸变后的应用场景示意图;
图10为采用本发明实施例一文件转换流程图;
图11为采用本发明实施例一畸变处理后合成新视频文件的示意图。
具体实施方式
下面结合附图对技术方案的实施作进一步的详细描述。
现在将参考附图描述实现本发明各个实施例的移动终端。在后续的描述中,使用用于表示元件的诸如“模块”、“部件”或“单元”的后缀仅为了有利于本发明实施例的说明,其本身并没有特定的意义。因此,"模块"与"部件"可以混合地使用。
在下面的详细说明中,陈述了众多的具体细节,以便彻底理解本发明。不过,对于本领域的普通技术人员来说,显然可在没有这些具体细节的情况下实践本发明。在其他情况下,没有详细说明公开的公知方法、过程、组件、电路和网络,以避免不必要地使实施例的各个方面模糊不清。
另外,本文中尽管多次采用术语“第一”、“第二”等来描述各种元件(或各种阈值或各种应用或各种指令或各种操作)等,不过这些元件(或阈值或应用或指令或操作)不应受这些术语的限制。这些术语只是用于区分一个元件(或阈值或应用或指令或操作)和另一个元件(或阈值或应用或指令或操作)。例如,第一操作可以被称为第二操作,第二操作也可以被称为第一操作,而不脱离本发明的范围,第一操作和第二操作都是操作,只是二者并不是相同的操作而已。
本发明实施例中的步骤并不一定是按照所描述的步骤顺序进行处理,可以按照需求有选择的将步骤打乱重排,或者删除实施例中的步骤,或者增加实施例中的步骤,本发明实施例中的步骤描述只是可选的顺序组合,并不代表本发明实施例的所有步骤顺序组合,实施例中的步骤顺序不能认为是对本发明的限制。
本发明实施例中的术语“和/或”指的是包括相关联的列举项目中的一个或多个的任何和全部的可能组合。还要说明的是:当用在本说明书中时,“包括/包含”指定所陈述的特征、整数、步骤、操作、元件和/或组件的存 在,但是不排除一个或多个其他特征、整数、步骤、操作、元件和/或组件和/或它们的组群的存在或添加。
本发明实施例的智能终端(如移动终端)可以以各种形式来实施。例如,本发明实施例中描述的移动终端可以包括诸如移动电话、智能电话、笔记本电脑、数字广播接收器、个人数字助理(PDA,Personal Digital Assistant)、平板电脑(PAD)、便携式多媒体播放器(PMP,Portable Media Player)、导航装置等等的移动终端以及诸如数字TV、台式计算机等等的固定终端。下面,假设终端是移动终端。然而,本领域技术人员将理解的是,除了特别用于移动目的的元件之外,根据本发明的实施方式的构造也能够应用于固定类型的终端。
图1为实现本发明各个实施例的移动终端一个可选的硬件结构示意图。
移动终端100可以包括通信单元110、音频/视频(A/V)输入单元120、用户输入单元130、获取单元140、播放单元141、第一处理单元142、第二处理单元143、输出单元150、存储单元160、接口单元170、处理单元180和电源单元190等等。图1示出了具有各种组件的移动终端,但是应理解的是,并不要求实施所有示出的组件。可以替代地实施更多或更少的组件。将在下面详细描述移动终端的元件。
通信单元110通常包括一个或多个组件,其允许移动终端100与无线通信***或网络之间的无线电通信(如果将移动终端用固定终端代替,也可以通过有线方式进行电通信)。例如,通信单元具体为无线通信单元时可以包括广播接收单元111、移动通信单元112、无线互联网单元113、短程通信单元114和位置信息单元115中的至少一个,这些单元是可选的,根据不同需求可以增删。
广播接收单元111经由广播信道从外部广播管理服务器接收广播信号和/或广播相关信息。广播信道可以包括卫星信道和/或地面信道。广播管理 服务器可以是生成并发送广播信号和/或广播相关信息的服务器或者接收之前生成的广播信号和/或广播相关信息并且将其发送给终端的服务器。广播信号可以包括TV广播信号、无线电广播信号、数据广播信号等等。而且,广播信号可以进一步包括与TV或无线电广播信号组合的广播信号。广播相关信息也可以经由移动通信网络提供,并且在该情况下,广播相关信息可以由移动通信单元112来接收。广播信号可以以各种形式存在,例如,其可以以数字多媒体广播(DMB,Digital Multimedia Broadcasting)的电子节目指南(EPG,Electronic Program Guide)、数字视频广播手持(DVB-H,Digital Video Broadcasting-Handheld)的电子服务指南(ESG,Electronic Service Guide)等等的形式而存在。广播接收单元111可以通过使用各种类型的广播***接收信号广播。特别地,广播接收单元111可以通过使用诸如多媒体广播-地面(DMB-T,Digital Multimedia Broadcasting-Terrestrial)、数字多媒体广播-卫星(DMB-S,Digital Multimedia Broadcasting-Satellite)、DVB-H,前向链路媒体(MediaFLO,Media Forward Link Only)的数据广播***、地面数字广播综合服务(ISDB-T,Integrated Services Digital Broadcasting-Terrestrial)等等的数字广播***接收数字广播。广播接收单元111可以被构造为适合提供广播信号的各种广播***以及上述数字广播***。经由广播接收单元111接收的广播信号和/或广播相关信息可以存储在存储器160(或者其它类型的存储介质)中。
移动通信单元112将无线电信号发送到基站(例如,接入点、节点B等等)、外部终端以及服务器中的至少一个和/或从其接收无线电信号。这样的无线电信号可以包括语音通话信号、视频通话信号、或者根据文本和/或多媒体消息发送和/或接收的各种类型的数据。
无线互联网单元113支持移动终端的无线互联网接入。该单元可以内部或外部地耦接到终端。该单元所涉及的无线互联网接入技术可以包括无 线局域网络(Wi-Fi,WLAN,Wireless Local Area Networks)、无线宽带(Wibro)、全球微波互联接入(Wimax)、高速下行链路分组接入(HSDPA,High Speed Downlink Packet Access)等等。
短程通信单元114是用于支持短程通信的单元。短程通信技术的一些示例包括蓝牙、射频识别(RFID,Radio Frequency Identification)、红外数据协会(IrDA,Infrared Data Association)、超宽带(UWB,Ultra Wideband)、紫蜂等等。
位置信息单元115是用于检查或获取移动终端的位置信息的单元。位置信息单元的典型示例是全球定位***(GPS,Global Positioning System)。根据当前的技术,位置信息单元115计算来自三个或更多卫星的距离信息和准确的时间信息并且对于计算的信息应用三角测量法,从而根据经度、纬度和高度准确地计算三维当前位置信息。当前,用于计算位置和时间信息的方法使用三颗卫星并且通过使用另外的一颗卫星校正计算出的位置和时间信息的误差。此外,位置信息单元115能够通过实时地连续计算当前位置信息来计算速度信息。
A/V输入单元120用于接收音频或视频信号。A/V输入单元120可以包括相机121和麦克风122,相机121对在视频捕获模式或图像捕获模式中由图像捕获装置获得的静态图片或视频的图像数据进行处理。处理后的图像帧可以显示在显示单元151上。经相机121处理后的图像帧可以存储在存储单元160(或其它存储介质)中或者经由通信单元110进行发送,可以根据移动终端的构造提供两个或更多相机121。麦克风122可以在电话通话模式、记录模式、语音识别模式等等运行模式中经由麦克风接收声音(音频数据),并且能够将这样的声音处理为音频数据。处理后的音频(语音)数据可以在电话通话模式的情况下转换为可经由移动通信单元112发送到移动通信基站的格式输出。麦克风122可以实施各种类型的噪声消除(或 抑制)算法以消除(或抑制)在接收和发送音频信号的过程中产生的噪声或者干扰。
用户输入单元130可以根据用户输入的命令生成键输入数据以控制移动终端的各种操作。用户输入单元130允许用户输入各种类型的信息,并且可以包括键盘、鼠标、触摸板(例如,检测由于被接触而导致的电阻、压力、电容等等的变化的触敏组件)、滚轮、摇杆等等。特别地,当触摸板以层的形式叠加在显示单元151上时,可以形成触摸屏。
获取单元140,用于获取文件转换逻辑;播放单元141,用于播放基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件;第一处理单元142,用于根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件,将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享;第二处理单元143,用于接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
接口单元170用作至少一个外部装置与移动终端100连接可以通过的接口。例如,外部装置可以包括有线或无线头戴式耳机端口、外部电源(或电池充电器)端口、有线或无线数据端口、存储卡端口、用于连接具有识别单元的装置的端口、音频输入/输出(I/O)端口、视频I/O端口、耳机端口等等。识别单元可以是存储用于验证用户使用移动终端100的各种信息并且可以包括用户识别单元(UIM,User Identify Module)、客户识别单元 (SIM,Subscriber Identity Module)、通用客户识别单元(USIM,Universal Subscriber Identity Module)等等。另外,具有识别单元的装置(下面称为"识别装置")可以采取智能卡的形式,因此,识别装置可以经由端口或其它连接装置与移动终端100连接。接口单元170可以用于接收来自外部装置的输入(例如,数据信息、电力等等)并且将接收到的输入传输到移动终端100内的一个或多个元件或者可以用于在移动终端和外部装置之间传输数据。
另外,当移动终端100与外部底座连接时,接口单元170可以用作允许通过其将电力从底座提供到移动终端100的路径或者可以用作允许从底座输入的各种命令信号通过其传输到移动终端的路径。从底座输入的各种命令信号或电力可以用作用于识别移动终端是否准确地安装在底座上的信号。输出单元150被构造为以视觉、音频和/或触觉方式提供输出信号(例如,音频信号、视频信号、振动信号等等)。输出单元150可以包括显示单元151、音频输出单元152等等。
显示单元151可以显示在移动终端100中处理的信息。例如,移动终端100可以显示相关用户界面(UI,User Interface)或图形用户界面(GUI,Graphical User Interface)。当移动终端100处于视频通话模式或者图像捕获模式时,显示单元151可以显示捕获的图像和/或接收的图像、示出视频或图像以及相关功能的UI或GUI等等。
同时,当显示单元151和触摸板以层的形式彼此叠加以形成触摸屏时,显示单元151可以用作输入装置和输出装置。显示单元151可以包括液晶显示器(LCD,Liquid Crystal Display)、薄膜晶体管LCD(TFT-LCD,Thin Film Transistor-LCD)、有机发光二极管(OLED,Organic Light-Emitting Diode)显示器、柔性显示器、三维(3D)显示器等等中的至少一种。这些显示器中的一些可以被构造为透明状以允许用户从外部观看,这可以称为 透明显示器,典型的透明显示器可以例如为透明有机发光二极管(TOLED)显示器等等。根据特定想要的实施方式,移动终端100可以包括两个或更多显示单元(或其它显示装置),例如,移动终端可以包括外部显示单元(未示出)和内部显示单元(未示出)。触摸屏可用于检测触摸输入压力以及触摸输入位置和触摸输入面积。
音频输出单元152可以在移动终端处于呼叫信号接收模式、通话模式、记录模式、语音识别模式、广播接收模式等等模式下时,将通信单元110接收的或者在存储器160中存储的音频数据转换音频信号并且输出为声音。而且,音频输出单元152可以提供与移动终端100执行的特定功能相关的音频输出(例如,呼叫信号接收声音、消息接收声音等等)。音频输出单元152可以包括扬声器、蜂鸣器等等。
存储单元160可以存储由处理单元180执行的处理和控制操作的软件程序等等,或者可以暂时地存储已经输出或将要输出的数据(例如,电话簿、消息、静态图像、视频等等)。而且,存储单元160可以存储关于当触摸施加到触摸屏时输出的各种方式的振动和音频信号的数据。
存储单元160可以包括至少一种类型的存储介质,所述存储介质包括闪存、硬盘、多媒体卡、卡型存储器(例如,SD或DX存储器等等)、随机访问存储器(RAM,Random Access Memory)、静态随机访问存储器(SRAM,Static Random Access Memory)、只读存储器(ROM,Read Only Memory)、电可擦除可编程只读存储器(EEPROM,Electrically Erasable Programmable Read Only Memory)、可编程只读存储器(PROM,Programmable Read Only Memory)、磁性存储器、磁盘、光盘等等。而且,移动终端100可以与通过网络连接执行存储单元160的存储功能的网络存储装置协作。
处理单元180通常控制移动终端的总体操作。例如,处理单元180执 行与语音通话、数据通信、视频通话等等相关的控制和处理。又如,处理单元180可以执行模式识别处理,以将在触摸屏上执行的手写输入或者图片绘制输入识别为字符或图像。
电源单元190在处理单元180的控制下接收外部电力或内部电力并且提供操作各元件和组件所需的适当的电力。
这里描述的各种实施方式可以以使用例如计算机软件、硬件或其任何组合的计算机可读介质来实施。对于硬件实施,这里描述的实施方式可以通过使用特定用途集成电路(ASIC,Application Specific Integrated Circuit)、数字信号处理器(DSP,Digital Signal Processing)、数字信号处理装置(DSPD,Digital Signal Processing Device)、可编程逻辑装置(PLD,Programmable Logic Device)、现场可编程门阵列(FPGA,Field Programmable Gate Array)、处理器、控制器、微控制器、微处理器、被设计为执行这里描述的功能的电子单元中的至少一种来实施,在一些情况下,这样的实施方式可以在处理单元180中实施。对于软件实施,诸如过程或功能的实施方式可以与允许执行至少一种功能或操作的单独的软件单元来实施。软件代码可以由以任何适当的编程语言编写的软件应用程序(或程序)来实施,软件代码可以存储在存储单元160中并且由处理单元180执行。其中,存储单元160的一个具体硬件实体可以为存储器,处理单元180的一个具体硬件实体可以为控制器。
至此,已经按照其功能描述了移动终端。下面,为了简要起见,将描述诸如折叠型、直板型、摆动型、滑动型移动终端等等的各种类型的移动终端中的滑动型移动终端作为示例。因此,本发明能够应用于任何类型的移动终端,并且不限于滑动型移动终端。
如图1中所示的移动终端100可以被构造为利用经由帧或分组发送数据的诸如有线和无线通信***以及基于卫星的通信***来操作。
现在将参考图2描述其中根据本发明实施例的移动终端能够操作的通信***。
这样的通信***可以使用不同的空中接口和/或物理层。例如,由通信***使用的空中接口包括例如频分多址(FDMA,Frequency Division Multiple Access)、时分多址(TDMA,Time Division Multiple Access)、码分多址(CDMA,Code Division Multiple Access)和通用移动通信***(UMTS,Universal Mobile Telecommunications System)(特别地,长期演进(LTE,Long Term Evolution))、全球移动通信***(GSM)等等。作为非限制性示例,下面的描述涉及CDMA通信***,但是这样的教导同样适用于其它类型的***。
参考图2,CDMA无线通信***可以包括多个移动终端100、多个基站(BS,Base Station)270、基站控制器(BSC,Base Station Controller)275和移动交换中心(MSC,Mobile Switching Center)280。MSC280被构造为与公共电话交换网络(PSTN,Public Switched Telephone Network)290形成接口。MSC280还被构造为与可以经由回程线路耦接到BS270的BSC275形成接口。回程线路可以根据若干已知的接口中的任一种来构造,所述接口包括例如E1/T1、ATM、IP、PPP、帧中继、HDSL、ADSL或xDSL。将理解的是,如图2中所示的***可以包括多个BSC275。
每个BS 270可以服务一个或多个分区(或区域),由多向天线或指向特定方向的天线覆盖的每个分区放射状地远离BS 270。或者,每个分区可以由用于分集接收的两个或更多天线覆盖。每个BS 270可以被构造为支持多个频率分配,并且每个频率分配具有特定频谱(例如,1.25MHz,5MHz等等)。
分区与频率分配的交叉可以被称为CDMA信道。BS 270也可以被称为基站收发器子***(BTS,Base Transceiver Station)或者其它等效术语。在 这样的情况下,术语“基站”可以用于笼统地表示单个BSC275和至少一个BS 270。基站也可以被称为“蜂窝站”。或者,特定BS 270的各分区可以被称为多个蜂窝站。
如图2中所示,广播发射器(BT,Broadcast Transmitter)295将广播信号发送给在***内操作的移动终端100。如图1中所示的广播接收单元111被设置在移动终端100处以接收由BT295发送的广播信号。在图2中,示出了几个卫星300,例如可以采用GPS卫星300。卫星300帮助定位多个移动终端100中的至少一个。
在图2中,描绘了多个卫星300,但是理解的是,可以利用任何数目的卫星获得有用的定位信息。如图1中所示的位置信息单元115通常被构造为与卫星300配合以获得想要的定位信息。替代GPS跟踪技术或者在GPS跟踪技术之外,可以使用可以跟踪移动终端的位置的其它技术。另外,至少一个GPS卫星300可以选择性地或者额外地处理卫星DMB传输。
作为无线通信***的一个典型操作,BS 270接收来自各种移动终端100的反向链路信号。移动终端100通常参与通话、消息收发和其它类型的通信。特定基站接收的每个反向链路信号被在特定BS 270内进行处理。获得的数据被转发给相关的BSC275。BSC275提供通话资源分配和包括BS 270之间的软切换过程的协调的移动管理功能。BSC275还将接收到的数据路由到MSC280,其提供用于与PSTN290形成接口的额外的路由服务。类似地,PSTN290与MSC280形成接口,MSC280与BSC275形成接口,并且BSC275相应地控制BS 270以将正向链路信号发送到移动终端100。
移动终端中通信单元110的移动通信单元112基于移动终端内置的接入移动通信网络(如2G/3G/4G等移动通信网络)的必要数据(包括用户识别信息和鉴权信息)接入移动通信网络为移动终端用户的网页浏览、网络多媒体播放等业务传输移动通信数据(包括上行的移动通信数据和下行的 移动通信数据)。
通信单元110的无线互联网单元113通过运行无线热点的相关协议功能而实现无线热点的功能,无线热点支持多个移动终端(移动终端之外的任意移动终端)接入,通过复用移动通信单元112与移动通信网络之间的移动通信连接为移动终端用户的网页浏览、网络多媒体播放等业务传输移动通信数据(包括上行的移动通信数据和下行的移动通信数据),由于移动终端实质上是复用移动终端与通信网络之间的移动通信连接传输移动通信数据的,因此移动终端消耗的移动通信数据的流量由通信网络侧的计费实体计入移动终端的通信资费,从而消耗移动终端签约使用的通信资费中包括的移动通信数据的数据流量。
图3为本发明实施例中进行信息交互的各方硬件实体的示意图,图3中包括:终端设备1和服务器2。其中,终端设备1由终端设备11-14构成,终端设备通过有线网络或者无线网络与服务器进行信息交互。终端设备包括手机、台式机、PC机、一体机等类型。采用本发明实施例,终端可以向服务器发起下载请求,以获取文件转换逻辑,该文件转换逻辑在实际应用中可以为通用安装包的形式,该通用安装包是经windows平台处理后得到的通用安装包,不需要手机安装专门的APP,只需要下载该通用安装包到终端本地后,就可以直接使用该通用安装包实现播放、畸变处理等操作。基于该通用安装包所针对的操作对象为同一源文件(如视频文件)的左、右格式文件,不仅会达到更好的畸变处理效果,而且不需要如现有技术一样是针对源文件的处理,最终将处理结果分别投屏到VR眼镜上播放,更加方便快捷。本发明实施例将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享,从而无需多次拷贝操作,只需要录制时去共享中获取所需文件即可,不仅可以提高效率,而且可以避免过多占用CPU和内存。在接收录制操作后,将所述畸变处理结果提供给录制线程进行录制,最后 编码合成为直接用于VR播放的第一目标文件和第二目标文件。可见,采用本发明实施例,在后台的windows平台处理生成文件转换逻辑(如通用安装包),终端得到该文件转换逻辑(如通用安装包),直接使用该文件转换逻辑(如通用安装包)进行处理即可。
具体的,通过终端设备执行的处理逻辑10如图3所示,处理逻辑10包括:
S1、获取文件转换逻辑,所述文件转换逻辑为在服务器侧根据在指定3D运行环境中加载运行的播放插件、SDK和录制插件所生成的逻辑;S2、根据文件转换逻辑播放基于源文件得到的左视频文件和右视频文件,左视频文件为基于源文件进行播放并显示于终端屏幕左侧的文件,右视频文件为基于源文件进行播放并显示于终端屏幕右侧的文件;S3、根据所述文件转换逻辑进行左视频文件和右视频文件对应的音视频解码,分别得到音频文件和图像文件;S4、将图像文件进行畸变处理,将得到的图像畸变处理结果以图像文件格式存储于显存中进行文件共享;S5、接收录制操作,将图像畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述图像畸变处理结果和所述音频文件编码合成为直接用于VR播放的目标文件,该目标文件为分别对应初始文件(左视频文件和右视频文件)的第一目标文件和第二目标文件,从而,最终将第一目标文件和第二目标文件直接用于VR眼镜上的播放。
上述图3的例子只是实现本发明实施例的一个***架构实例,本发明实施例并不限于上述图3所述的***结构,基于上述图3所述的***架构,提出本发明方法各个实施例。
本发明实施例的一种文件处理方法,如图4所示,所述方法包括:获取文件转换逻辑,所述文件转换逻辑为在服务器侧根据在指定3D运行环境中加载运行的播放插件、软件开发工具包(SDK)和录制插件所生成的逻 辑101,以便后续的具体处理中,可以使用该文件转换逻辑进行播放、音视频分解、畸变、目标文件录制等一系列操作。该文件转换逻辑是SDK安装包形式存在时,该安装包为经windows平台处理后得到的通用安装包,不需要终端安装专门的APP,任意终端只需要下载该通用安装包到终端本地后,就可以直接使用该安装包实现播放,畸变处理等操作,即使用该通用安装包对源文件进行处理以得到目标文件,进一步还可以将该目标文件分享到网络上。在使用该文件转换逻辑进行播放、音视频分解、畸变、目标文件录制等一系列操作中,首先播放基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件102。根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件103。将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享104。接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件105。
VR技术是一种可以创建和体验虚拟世界的计算机仿真***,它利用计算机生成一种模拟环境,提供多源信息融合的交互式的三维动态视景和实体行为的***仿真,以便使用户沉浸到该模拟环境中。VR主要包括模拟环境、感知、自然技能和传感设备等方面。该模拟环境是由计算机生成的、实时动态的三维立体逼真图像。感知是指理想的VR应该具有一切人所具有的感知。除计算机图形技术所生成的视觉感知外,还有听觉、触觉、力觉、运动等感知,甚至还包括嗅觉和味觉等,也称为多感知。自然技能是指人 的头部转动,眼睛、手势、或其他人体行为动作,由计算机来处理与参与者的动作相适应的数据,并对用户的输入作出实时响应,并分别反馈到用户的五官。传感设备是指三维交互设备。
随着VR的发展,各大VR厂商推出了各自的VR眼镜,对于这种通过VR眼镜得到的新的人机交互方式,由于其更自然和更立体的环境模拟,结合移动终端的日益普及性,移动VR更快的走进了大众消费者的生活,VR眼镜不仅支持全景的画面,而且还支持3D的画面。可是目前3D视频的播放都需要安装专门APP才可以进行转换,不但麻烦,而且还不支持网页的3D播放,其原因是:通常终端上安装的APP只是视频播放器,没有网页浏览功能,所以对于网页的3D视频播放是不支持的,目前3D视频大部分都是左右格式,这种格式不能直接在VR眼镜中正常观看,因为VR眼镜是需要把畸变后图像投射到眼镜,才会有3D立体效果。
在实际应用中,以移动VR场景为例,采用本发明实施例,基于文件转码的处理逻辑(如文件转换逻辑),通过手机终端任意的内置播放器即可播放,用户还可以方便将转换后的目标视频文件上传至视频网站,方便分享,从而在网页里就可以体验3D视频。对于该基于文件转码的处理逻辑(如文件转换逻辑)而言,播放,解码,畸变、录制,保存,分享,采用VLC+Unity3D+CardboardSDK+OBS组合的方式,生成新的视频文件,具体的,是针对左、右格式视频做畸变转换并保存为相同格式视频文件,可以轻松在手机终端的任意视频播放器,不需要专门的APP进行处理。将生成后的文件,一键分享至网页上,在网页页面上进行播放。
采用本发明实施例,除了区别于现有技术,现有技术是对原始的源文件进行的处理,最终,还需要投射到VR眼镜中投屏显示,对格式有限制,而本发明是直接将源文件通过VR眼镜显示所需要的左右格式视频为基础进行处理,不受格式显示,后续,VR直接可以进行观看,不需要投屏。而 且,畸变后的图像通过显存共享,避免录制过程中的多次拷贝,避免过多占用CPU和内存,从而影响到总的处理效率,实现了高效、无延迟图像传输,节省了CPU和内存资源。
对于如何在后台windows生成上述基于文件转码的处理逻辑(如文件转换逻辑)而言,其中,VLC指视频解码,音频解码;Cardboard指图像畸变;OBS指将畸变后的图像和音频进行编码压缩,保存文件。运行在Windows平台的Unity环境中,导入Cardboard SDK插件包。Cardboard SDK的作用是做畸变。Cardboard SDK可以根据屏幕宽高比、分辨率来适配畸变参数,还有预置多种种手机型号。Cardboard SDK本身是很全的开发包,本方案这里只用了畸变功能,所以需要对其进行裁剪。Unity中没有任意视频格式的播放器,VLC是全平台的全格式播放器,做一个VLC在Unity中的播放插件,利用VLC可以将3D视频的音频和图像分别解码出来,然后,对于需要处理的重点图像进行畸变处理。OBS作为windows平台的录制工具,自带插件可以针对窗口进行录制。由于针对窗口录像需要做多次拷贝操作,效率很低,占用CPU和内存都很高。本方案对这里进行优化,分别写了两个插件,一个用于Unity中,将畸变后的图像通过显存共享。另一个插件用于在OBS中,从显存中拿到图像,直接交给录制线程。这样在图像拷贝过程中,因为都是在显存中操作,拷贝操作基本无消耗内存,无消耗CPU。OBS将畸变后的图像和声音,通过压缩编码,最后保存为目标视频文件。
本发明实施例的一种文件处理方法,如图5所示,所述方法包括:获取文件转换逻辑,所述文件转换逻辑为在服务器侧根据在指定3D运行环境中加载运行的播放插件、SDK和录制插件所生成的逻辑201,以便后续的具体处理中,可以使用该文件转换逻辑进行播放、音视频分解、畸变、目标文件录制等一系列操作。该文件转换逻辑是SDK安装包形式存在时,该 安装包为经WINDOW平台处理后得到的通用安装包,不需要终端安装专门的APP,任意终端只需要下载该通用安装包到终端本地后,就可以直接使用该安装包实现播放,畸变处理等操作,即使用该通用安装包对源文件进行处理以得到目标文件,进一步还可以将该目标文件分享到网络上。在使用该文件转换逻辑进行播放、音视频分解、畸变、目标文件录制等一系列操作中,首先播放基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件202。根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到对应第一待处理文件的第一音频文件和第一图像文件,及对应第二待处理文件的第二音频文件和第二图像文件,即根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件203。在将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理的过程中,获取与终端屏幕显示有关的第一信息204,所述第一信息包括:屏幕宽高比、分辨率中的至少一种信息。获取与终端自身性能有关的第二信息205,所述第二信息包括:终端的型号、终端处理器性能指标中的至少一种信息。根据所述第一信息和/或所述第二信息对所述第一图像文件和所述第二图像文件进行畸变处理,得到与第一图像文件相关的第三图像文件和与第二图像文件相关的第四图像文件206,将所述第三图像文件和所述第四图像文件存储于显存中进行文件共享207,以便将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享。接收录制操作,将所述第三图像文件和所述第四图像文件提供给录制线程进行录制时,根据所述文件转换逻辑将所述第一音频文件和所述第三图像文件编码合成为第一目标文件,将所述第二音频文件和所述第四图像文件编码合成为第二目标文件,将第一目标文件和第 二目标文件直接用于VR播放208。
VR技术是一种可以创建和体验虚拟世界的计算机仿真***,它利用计算机生成一种模拟环境,提供多源信息融合的交互式的三维动态视景和实体行为的***仿真,以便使用户沉浸到该模拟环境中。VR主要包括模拟环境、感知、自然技能和传感设备等方面。该模拟环境是由计算机生成的、实时动态的三维立体逼真图像。感知是指理想的VR应该具有一切人所具有的感知。除计算机图形技术所生成的视觉感知外,还有听觉、触觉、力觉、运动等感知,甚至还包括嗅觉和味觉等,也称为多感知。自然技能是指人的头部转动,眼睛、手势、或其他人体行为动作,由计算机来处理与参与者的动作相适应的数据,并对用户的输入作出实时响应,并分别反馈到用户的五官。传感设备是指三维交互设备。
采用本发明实施例,将所述第一目标文件和所述第二目标文件,与VR眼镜结合的场景中,由于本发明实施例是直接将源文件通过VR眼镜显示所需要的左右格式视频为基础进行处理,因此,不受格式显示及应用范围的限制,VR眼镜可以直接使用上述目标文件,在任意终端或网页上进行观看,不需要投屏处理,提高了处理效率,使得文件格式对于各种终端和应用范围具备兼容性。其中,各种终端播放上述目标文件势必会带来不同程度的畸变,因此,在本发明实施例中,预先通过文件转换逻辑,针对不同终端存在的畸变可能性,采用对应的畸变参数进行调整,以确保最终播放的效果是清晰和无畸变的。具体的,畸变参数,与屏幕宽高比、分辨率、终端如手机终端的型号等信息有关。
现有技术是对原始的源文件进行的处理,最终,还需要投射到VR眼镜中投屏显示,对格式有限制,而本发明实施例是直接将源文件通过VR眼镜显示所需要的左右格式视频为基础进行处理,不受格式显示,后续,VR直接可以进行观看,不需要投屏。而且,畸变后的图像通过显存共享,避免 录制过程中的多次拷贝,避免过多占用CPU和内存,从而影响到总的处理效率,实现了高效、无延迟图像传输,节省了CPU和内存资源。
对于如何在后台windows生成上述基于文件转码的处理逻辑(如文件转换逻辑)而言,其中,VLC指视频解码,音频解码;Cardboard指图像畸变;OBS指将畸变后的图像和音频进行编码压缩,保存文件。运行在Windows平台的Unity环境中,导入Cardboard SDK插件包。Cardboard SDK的作用是做畸变。Cardboard SDK可以根据屏幕宽高比、分辨率来适配畸变参数,还有预置多种种手机型号。Cardboard SDK本身是很全的开发包,本方案这里只用了畸变功能,所以需要对其进行裁剪。Unity中没有任意视频格式的播放器,VLC是全平台的全格式播放器,做一个VLC在Unity中的播放插件,利用VLC可以将3D视频的音频和图像分别解码出来,然后,对于需要处理的重点图像进行畸变处理。OBS作为windows平台的录制工具,自带插件可以针对窗口进行录制。由于针对窗口录像需要做多次拷贝操作,效率很低,占用CPU和内存都很高。本方案对这里进行优化,分别写了两个插件,一个用于Unity中,将畸变后的图像通过显存共享。另一个插件用于在OBS中,从显存中拿到图像,直接交给录制线程。这样在图像拷贝过程中,因为都是在显存中操作,拷贝操作基本无消耗内存,无消耗CPU。OBS将畸变后的图像和声音,通过压缩编码,最后保存为目标视频文件。
基于上述各个实施例,本发明实施例一实施方式中,所述方法还包括:接收文件分享操作,将所述第一目标文件和所述第二目标文件按照目标用户地址分享给所述目标用户;或者,将所述第一目标文件和所述第二目标文件按照网页连接地址分享到对应的网页上。
基于上述各个实施例,本发明实施例一实施方式中,所述方法还包括:将所述第一目标文件和所述第二目标文件以相同格式的视频文件进行存 储。
本发明实施例的一种文件处理***,终端可以向服务器发起下载请求,以获取文件转换逻辑,该文件转换逻辑在实际应用中可以为通用安装包的形式,该通用安装包是经windows平台处理后得到的通用安装包,不需要手机安装专门的APP,只需要下载该通用安装包到终端本地后,就可以直接使用该通用安装包实现播放、畸变处理等操作。基于该通用安装包所针对的操作对象为同一源文件(如视频文件)的左、右格式文件,不仅会达到更好的畸变处理效果,而且不需要如现有技术一样是针对源文件的处理,最终将处理结果分别投屏到VR眼镜上播放,更加方便快捷。本发明实施例将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享,从而无需多次拷贝操作,只需要录制时去共享中获取所需文件即可,不仅可以提高效率,而且可以避免过多占用CPU和内存。在接收录制操作后,将所述畸变处理结果提供给录制线程进行录制,最后编码合成为直接用于VR播放的第一目标文件和第二目标文件。可见,采用本发明实施例,在后台的windows平台处理生成文件转换逻辑(如通用安装包),终端得到该文件转换逻辑(如通用安装包),直接使用该文件转换逻辑(如通用安装包)进行处理即可。如图6所示,包括:终端41和服务器42。其中,终端41包括:获取单元411,用于获取文件转换逻辑;播放单元412,用于播放基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件;第一处理单元413,用于根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件,将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享;第二处理 单元414,用于接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。服务器42包括:生成单元421,用于根据在指定3D运行环境(如Unity)中加载运行的播放插件(如VLC)、SDK(如Cardboard SDK)和录制插件(如OBS)生成文件转换逻辑;反馈单元422,用于响应终端发起的下载请求,反馈所述文件转换逻辑给终端。
在实际应用中,以移动VR场景为例,采用本发明实施例,基于文件转码的处理逻辑(如文件转换逻辑),通过手机终端任意的内置播放器即可播放,用户还可以方便将转换后的目标视频文件上传至视频网站,方便分享,从而在网页里就可以体验3D视频。对于该基于文件转码的处理逻辑(如文件转换逻辑)而言,播放,解码,畸变、录制,保存,分享,采用VLC+Unity3D+CardboardSDK+OBS组合的方式,生成新的视频文件,具体的,是针对左、右格式视频做畸变转换并保存为相同格式视频文件,可以轻松在手机终端的任意视频播放器,不需要专门的APP进行处理。将生成后的文件,一键分享至网页上,在网页页面上进行播放。
采用本发明实施例,除了区别于现有技术,现有技术是对原始的源文件进行的处理,最终,还需要投射到VR眼镜中投屏显示,对格式有限制,而本发明是直接将源文件通过VR眼镜显示所需要的左右格式视频为基础进行处理,不受格式显示,后续,VR直接可以进行观看,不需要投屏。而且,畸变后的图像通过显存共享,避免录制过程中的多次拷贝,避免过多占用CPU和内存,从而影响到总的处理效率,实现了高效、无延迟图像传输,节省了CPU和内存资源。
对于在上述服务器侧如何在后台windows生成上述基于文件转码的处理逻辑(如文件转换逻辑)而言,其中,VLC指视频解码,音频解码;Cardboard 指图像畸变;OBS指将畸变后的图像和音频进行编码压缩,保存文件。运行在Windows平台的Unity环境中,导入Cardboard SDK插件包。Cardboard SDK的作用是做畸变。Cardboard SDK可以根据屏幕宽高比、分辨率来适配畸变参数,还有预置多种种手机型号。Cardboard SDK本身是很全的开发包,本方案这里只用了畸变功能,所以需要对其进行裁剪。Unity中没有任意视频格式的播放器,VLC是全平台的全格式播放器,做一个VLC在Unity中的播放插件,利用VLC可以将3D视频的音频和图像分别解码出来,然后,对于需要处理的重点图像进行畸变处理。OBS作为windows平台的录制工具,自带插件可以针对窗口进行录制。由于针对窗口录像需要做多次拷贝操作,效率很低,占用CPU和内存都很高。本方案对这里进行优化,分别写了两个插件,一个用于Unity中,将畸变后的图像通过显存共享。另一个插件用于在OBS中,从显存中拿到图像,直接交给录制线程。这样在图像拷贝过程中,因为都是在显存中操作,拷贝操作基本无消耗内存,无消耗CPU。OBS将畸变后的图像和声音,通过压缩编码,最后保存为目标视频文件。
在本发明实施例一实施方式中,终端中的所述第一处理单元,进一步用于:根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到对应第一待处理文件的第一音频文件和第一图像文件,及对应第二待处理文件的第二音频文件和第二图像文件。
在本发明实施例一实施方式中,终端中的所述第一处理单元,进一步用于:获取与终端屏幕显示有关的第一信息;获取与终端自身性能有关的第二信息;根据所述第一信息和/或所述第二信息对所述第一图像文件和所述第二图像文件进行畸变处理,得到与第一图像文件相关的第三图像文件和与第二图像文件相关的第四图像文件,将所述第三图像文件和所述第四图像文件存储于显存中进行文件共享。其中,所述第一信息包括:屏幕宽 高比、分辨率中的至少一种信息;所述第二信息包括:终端的型号、终端处理器性能指标中的至少一种信息。
在本发明实施例一实施方式中,终端中的所述第二处理单元,进一步用于:接收录制操作,将所述第三图像文件和所述第四图像文件提供给录制线程进行录制时,根据所述文件转换逻辑将所述第一音频文件和所述第三图像文件编码合成为第一目标文件,将所述第二音频文件和所述第四图像文件编码合成为第二目标文件。
在本发明实施例一实施方式中,所述终端还包括:分享单元,用于接收文件分享操作,将所述第一目标文件和所述第二目标文件按照目标用户地址分享给所述目标用户;或者,将所述第一目标文件和所述第二目标文件按照网页连接地址分享到对应的网页上。
在本发明实施例一实施方式中,所述终端还包括:存储单元,用于将所述第一目标文件和所述第二目标文件以相同格式的视频文件进行存储。
在本发明实施例一实施方式中,服务器中的所述生成单元,进一步用于:在所述指定3D运行环境中导入所述SDK;在所述指定3D运行环境中加载并运行播放插件,通过所述播放插件对基于源文件得到的第一待处理文件和第二待处理文件进行音视频解码,以得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件;监控到图像文件存在畸变时,通过所述SDK将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理;通过所述录制插件开启录制线程进行录制之前,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享。
在本发明实施例一实施方式中,服务器中的所述录制插件包括:与所述指定3D运行环境显存相关的第一插件和响应录制操作的第二插件;所述服务器包括:通信单元,用于通过所述第一插件与所述显存进行通信,从所述显存中提取出所述畸变处理结果;录制单元,用于响应录制操作,所 述第二插件与所述第一插件进行通信,将所述畸变处理结果提供给录制线程进行录制,将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
其中,对于用于数据处理的处理器而言,在执行处理时,可以采用微处理器、中央处理器(CPU,Central Processing Unit)、DSP或FPGA实现;对于存储介质来说,包含操作指令,该操作指令可以为计算机可执行代码,通过所述操作指令来实现上述本发明实施例信息处理方法流程中的各个步骤。
这里需要指出的是:以上涉及终端和服务器项的描述,与上述方法描述是类似的,同方法的有益效果描述,不做赘述。对于本发明终端和服务器实施例中未披露的技术细节,请参照本发明方法流程描述的实施例所描述内容。
以一个现实应用场景为例对本发明实施例阐述如下:
VR技术的发展带来VR眼镜的普及,这种新的人机交互方式,受限于PC VR的硬件成本,而移动VR更能符合大众消费者的使用需求。VR眼镜不仅支持全景的画面,而且还支持3D的画面。可是目前3D视频的播放都需要安装APP才可以,不但麻烦,还不支持网页的3D播放,如图7所示为使用一个包含视频内容,游戏内容,视频播放器的APP进行VR播放的流程,该流程中,对3D视频源进行音、视频解码后,对视频解码得到的图像进行畸变处理。最终将音频通过扬声器播放,将畸变处理的图像投射到VR屏幕上播放。由于目前3D视频大部分都是左右格式,基于该流程的处理方式,是对3D视频源的处理,而这种格式不能直接在VR眼镜中正常观看,因为VR眼镜是需要把畸变后图像投射到眼镜,才会有3D立体效果。需要安装专门的APP,也不支持网页播放,这些问题都不符合大众消费者 的需求。
对于上述问题,采用本发明实施例,可以采用VLC+Unity3D+CardboardSDK+OBS组合的方式,实现了针对左右格式视频做畸变转换并保存为相同格式视频文件,比如,可以轻松在手机终端的任意视频播放器,网页页面进行播放。带上VR眼镜,就可以体验3D视频效果。采用VLC+Unity3D+CardboardSDK+OBS组合的方式来生成文件转换逻辑,运行在Windows平台。之后根据终端的下载请求将该文件转换逻辑反馈给终端直接使用,以实现针对源文件的左右格式文件的音、视频解码、畸变处理、编码合成、录制和保存,及网页分享功能。
具体的,就文件转换逻辑而言,在Unity环境中,导入Cardboard SDK插件包。Cardboard SDK的作用是做畸变,Cardboard SDK可以根据屏幕宽高比、分辨率来适配畸变参数,还有预置多种手机型号。Cardboard SDK本身是很全的开发包,本方案这里只用了畸变功能,所以需要对其进行裁剪。Unity中没有任意视频格式的播放器,VLC是全平台的全格式播放器,这里需要做一个VLC在Unity中的播放插件,利用VLC可以将3D视频的音频和图像分别解码出来,然后重点图像进行畸变处理。畸变的前后的图像如图8-9所示,图8为畸变前的图像,图9为畸变后的图像。OBS是windows平台下一个录制工具,自带插件可以针对窗口进行录制。由于针对窗口录像需要做多次拷贝操作,效率很低,占用CPU和内存都很高。本方案对这里进行优化,分别写了两个插件,一个用于Unity中,将畸变后的图像通过显存共享。另一个插件用于在OBS中,从显存中拿到图像,直接交给录制线程。这样在图像拷贝过程中,因为都是在显存中操作,拷贝操作基本无消耗内存,无消耗CPU。OBS将畸变后的图像和声音,通过压缩编码,最后保存为新的视频文件,这个文件可以直接在手机自带播放器中播放,也可以上传至视频网站,用网页播放。大大减少了用户分享、体验3D视频的 成本。其中,VLC指:视频解码,音频解码;Cardboard指:图像畸变;OBS指:将畸变后的图像和音频进行编码压缩,保存文件。
如图10所示为采用本发明实施例的文件转换流程,该流程中,对源自于3D视频源(如3D视频文件)的左、右格式的3D源视频基于上述文件转换逻辑进行音、视频解码、畸变处理、编码合成、录制和保存。其中,利用播放插件VLC对左、右格式的3D源视频进行播放,对左、右格式的3D源视频分别进行音、视频解码分离,对视频解码分离得到的图像进行畸变处理后再将图像畸变处理结果放入显存中进行共享,通过录制插件OBS进行录制时,得到音频文件和显存中共享的该图像畸变处理结果,将二者进行编码合成并保持,得到新的视频文件,在VR上直接播放该新的视频文件,畸变进程与OBS录制进程间的交互处理如图11所示,在畸变过程中,需要对变形的视频进行畸变处理,以消除变形的影响。具体的,获取原始视频文件,播放插件VLC解码,将视频帧和音频帧分离(是对左、右格式的3D源视频分别进行音、视频解码分离),将音频帧放入虚拟声卡中,仅仅对视频帧进行畸变处理。将对视频帧进行畸变处理得到的畸变处理结果放入显存,在显存中进行共享的好处是:畸变处理结果以图像文件格式存储于显存中进行文件共享,从而无需多次拷贝操作,只需要录制时去共享中获取所需文件即可,不仅可以提高效率,而且可以避免过多占用CPU和内存。在OBS进程中,通过录制插件OBS进行录制时,捕获音频和显存中共享的对视频的畸变处理结果,接收录制操作后,将所述畸变处理结果提供给录制线程进行录制,将二者进行编码合成并保持,得到新的视频文件,在VR上直接播放该新的视频文件。
基于该文件转换流程的一个实际应用中,本方案是VR Capture中的一个功能,VR Capture是一个VR PC眼镜的录制分享工具,将VR Capture录制的左右格式视频,做畸变转换,生成新的视频文件。VR Capture将生成 后的文件,一键分享至网页上。本方案还对外提供一个接口DLL,直接调用开始转换和结束转换即可。同时还提供转换进度的回调。
本发明实施例的一种计算机存储介质,可以是包括计算机程序的存储器,上述计算机程序可由数据处理装置的处理器执行,以完成前述实施例中方法所述步骤。计算机存储介质可以是FRAM、ROM、PROM、EPROM、EEPROM、Flash Memory、磁表面存储器、光盘、或CD-ROM等存储器;也可以是包括上述存储器之一或任意组合的各种设备,如移动电话、计算机、平板设备、个人数字助理等。
该计算机可读存储介质,其上存储有计算机程序,该计算机程序被处理器运行时,执行如下文件处理方法。
一方面,该计算机可读存储介质位于终端侧时,包括如下内容:
一实施例中,该计算机程序被处理器运行时,执行:
获取文件转换逻辑,所述文件转换逻辑为在服务器侧根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件所生成的逻辑;
播放基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件;
根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件,将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享;
接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第 二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
一实施例中,该计算机程序被处理器运行时,还执行:
根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到对应第一待处理文件的第一音频文件和第一图像文件,及对应第二待处理文件的第二音频文件和第二图像文件。
一实施例中,该计算机程序被处理器运行时,还执行:
获取与终端屏幕显示有关的第一信息;
获取与终端自身性能有关的第二信息;
根据所述第一信息和/或所述第二信息对所述第一图像文件和所述第二图像文件进行畸变处理,得到与第一图像文件相关的第三图像文件和与第二图像文件相关的第四图像文件,将所述第三图像文件和所述第四图像文件存储于显存中进行文件共享。
所述第一信息包括:屏幕宽高比、分辨率中的至少一种信息。
所述第二信息包括:终端的型号、终端处理器性能指标中的至少一种信息。
一实施例中,该计算机程序被处理器运行时,还执行:
接收录制操作,将所述第三图像文件和所述第四图像文件提供给录制线程进行录制时,根据所述文件转换逻辑将所述第一音频文件和所述第三图像文件编码合成为第一目标文件,将所述第二音频文件和所述第四图像文件编码合成为第二目标文件。
一实施例中,该计算机程序被处理器运行时,还执行:
接收文件分享操作,将所述第一目标文件和所述第二目标文件按照目标用户地址分享给所述目标用户;或者,将所述第一目标文件和所述第二目标文件按照网页连接地址分享到对应的网页上。
一实施例中,该计算机程序被处理器运行时,还执行:
将所述第一目标文件和所述第二目标文件以相同格式的视频文件进行存储。
另一方面,该计算机可读存储介质位于服务器侧时,包括如下内容:
一实施例中,该计算机程序被处理器运行时,执行:
根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件生成文件转换逻辑;
响应终端发起的下载请求,反馈所述文件转换逻辑给终端。
一实施例中,该计算机程序被处理器运行时,还执行:
在所述指定3D运行环境中导入所述SDK;
在所述指定3D运行环境中加载并运行播放插件,通过所述播放插件对基于源文件得到的第一待处理文件和第二待处理文件进行音视频解码,以得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件;
监控到图像文件存在畸变时,通过所述SDK将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理;
通过所述录制插件开启录制线程进行录制之前,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享。
一实施例中,所述录制插件包括:与所述指定3D运行环境显存相关的第一插件和响应录制操作的第二插件,该计算机程序被处理器运行时,还执行:
通过所述第一插件与所述显存进行通信,从所述显存中提取出所述畸变处理结果;
响应录制操作,所述第二插件与所述第一插件进行通信,将所述畸变处理结果提供给录制线程进行录制,将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现 实VR播放的第一目标文件和第二目标文件。
在本申请所提供的几个实施例中,应该理解到,所揭露的设备和方法,可以通过其它的方式实现。以上所描述的设备实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,如:多个单元或组件可以结合,或可以集成到另一个***,或一些特征可以忽略,或不执行。另外,所显示或讨论的各组成部分相互之间的耦合、或直接耦合、或通信连接可以是通过一些接口,设备或单元的间接耦合或通信连接,可以是电性的、机械的或其它形式的。
上述作为分离部件说明的单元可以是、或也可以不是物理上分开的,作为单元显示的部件可以是、或也可以不是物理单元,即可以位于一个地方,也可以分布到多个网络单元上;可以根据实际的需要选择其中的部分或全部单元来实现本实施例方案的目的。
另外,在本发明各实施例中的各功能单元可以全部集成在一个处理单元中,也可以是各单元分别单独作为一个单元,也可以两个或两个以上单元集成在一个单元中;上述集成的单元既可以采用硬件的形式实现,也可以采用硬件加软件功能单元的形式实现。
本领域普通技术人员可以理解:实现上述方法实施例的全部或部分步骤可以通过程序指令相关的硬件来完成,前述的程序可以存储于一计算机可读取存储介质中,该程序在执行时,执行包括上述方法实施例的步骤;而前述的存储介质包括:移动存储设备、ROM、RAM、磁碟或者光盘等各种可以存储程序代码的介质。
或者,本发明上述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,也可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明实施例的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一 个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机、服务器、或者网络设备等)执行本发明各个实施例所述方法的全部或部分。而前述的存储介质包括:移动存储设备、ROM、RAM、磁碟或者光盘等各种可以存储程序代码的介质。
以上所述,仅为本发明的具体实施方式,但本发明的保护范围并不局限于此,任何熟悉本技术领域的技术人员在本发明揭露的技术范围内,可轻易想到变化或替换,都应涵盖在本发明的保护范围之内。因此,本发明的保护范围应以所述权利要求的保护范围为准。
工业实用性
采用发明实施例,以文件转换逻辑作为通用的处理机制,可以进行音视频解码、畸变处理等一系列处理,从而自适应任何源文件和终端,不需要在终端上安装专门的应用进行处理,会达到更好的畸变处理效果。将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享,从而无需多次拷贝操作,只需要录制时去共享中获取所需文件即可,不仅可以提高效率,而且可以避免过多占用CPU和内存。在接收录制操作后,将所述畸变处理结果提供给录制线程进行录制,最后编码合成为直接用于VR播放的目标文件。

Claims (31)

  1. 一种文件处理方法,所述方法包括:
    获取文件转换逻辑,所述文件转换逻辑为在服务器侧根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件所生成的逻辑;
    播放基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件;
    根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件,将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享;
    接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
  2. 根据权利要求1所述的方法,其中,所述根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件,包括:
    根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到对应第一待处理文件的第一音频文件和第一图像文件,及对应第二待处理文件的第二音频文件和第二图像文件。
  3. 根据权利要求2所述的方法,其中,将分别对应第一待处理文件和 第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享,包括:
    获取与终端屏幕显示有关的第一信息;
    获取与终端自身性能有关的第二信息;
    根据所述第一信息和/或所述第二信息对所述第一图像文件和所述第二图像文件进行畸变处理,得到与第一图像文件相关的第三图像文件和与第二图像文件相关的第四图像文件,将所述第三图像文件和所述第四图像文件存储于显存中进行文件共享。
  4. 根据权利要求3所述的方法,其中,所述第一信息包括:屏幕宽高比、分辨率中的至少一种信息;
    所述第二信息包括:终端的型号、终端处理器性能指标中的至少一种信息。
  5. 根据权利要求3所述的方法,其中,所述接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于VR播放的第一目标文件和第二目标文件,包括:
    接收录制操作,将所述第三图像文件和所述第四图像文件提供给录制线程进行录制时,根据所述文件转换逻辑将所述第一音频文件和所述第三图像文件编码合成为第一目标文件,将所述第二音频文件和所述第四图像文件编码合成为第二目标文件。
  6. 根据权利要求1至5任一项所述的方法,其中,所述方法还包括:
    接收文件分享操作,将所述第一目标文件和所述第二目标文件按照目标用户地址分享给所述目标用户;或者,将所述第一目标文件和所述第二目标文件按照网页连接地址分享到对应的网页上。
  7. 根据权利要求1至5任一项所述的方法,其中,所述方法还包括:
    将所述第一目标文件和所述第二目标文件以相同格式的视频文件进行存储。
  8. 一种终端,所述终端包括:
    获取单元,用于获取文件转换逻辑,所述文件转换逻辑为在服务器侧根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件所生成的逻辑;
    播放单元,用于播放基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件;
    第一处理单元,用于根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件,将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享;
    第二处理单元,用于接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
  9. 根据权利要求8所述的终端,其中,所述第一处理单元,进一步用于:
    根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到对应第一待处理文件的第一音频文件和第一图像文件,及对应第二待处理文件的第二音频文件和第二图像文件。
  10. 根据权利要求9所述的终端,其中,所述第一处理单元,进一步用于:
    获取与终端屏幕显示有关的第一信息;
    获取与终端自身性能有关的第二信息;
    根据所述第一信息和/或所述第二信息对所述第一图像文件和所述第二图像文件进行畸变处理,得到与第一图像文件相关的第三图像文件和与第二图像文件相关的第四图像文件,将所述第三图像文件和所述第四图像文件存储于显存中进行文件共享。
  11. 根据权利要求10所述的终端,其中,所述第一信息包括:屏幕宽高比、分辨率中的至少一种信息;
    所述第二信息包括:终端的型号、终端处理器性能指标中的至少一种信息。
  12. 根据权利要求10所述的终端,其中,所述第二处理单元,进一步用于:
    接收录制操作,将所述第三图像文件和所述第四图像文件提供给录制线程进行录制时,根据所述文件转换逻辑将所述第一音频文件和所述第三图像文件编码合成为第一目标文件,将所述第二音频文件和所述第四图像文件编码合成为第二目标文件。
  13. 根据权利要求8至12任一项所述的终端,其中,所述终端还包括:
    分享单元,用于接收文件分享操作,将所述第一目标文件和所述第二目标文件按照目标用户地址分享给所述目标用户;或者,将所述第一目标文件和所述第二目标文件按照网页连接地址分享到对应的网页上。
  14. 根据权利要求8至12任一项所述的终端,其中,所述终端还包括:
    存储单元,用于将所述第一目标文件和所述第二目标文件以相同格式的视频文件进行存储。
  15. 一种文件处理方法,所述方法包括:
    根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK 和录制插件生成文件转换逻辑;
    响应终端发起的下载请求,反馈所述文件转换逻辑给终端。
  16. 根据权利要求15所述的方法,其中,所述根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件生成文件转换逻辑,包括:
    在所述指定3D运行环境中导入所述SDK;
    在所述指定3D运行环境中加载并运行播放插件,通过所述播放插件对基于源文件得到的第一待处理文件和第二待处理文件进行音视频解码,以得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件;
    监控到图像文件存在畸变时,通过所述SDK将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理;
    通过所述录制插件开启录制线程进行录制之前,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享。
  17. 根据权利要求16所述的方法,其中,所述录制插件包括:与所述指定3D运行环境显存相关的第一插件和响应录制操作的第二插件;
    通过所述第一插件与所述显存进行通信,从所述显存中提取出所述畸变处理结果;
    响应录制操作,所述第二插件与所述第一插件进行通信,将所述畸变处理结果提供给录制线程进行录制,将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
  18. 一种服务器,所述服务器包括:
    生成单元,用于根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件生成文件转换逻辑;
    反馈单元,用于响应终端发起的下载请求,反馈所述文件转换逻辑给 终端。
  19. 根据权利要求18所述的服务器,其中,所述生成单元,进一步用于:
    在所述指定3D运行环境中导入所述SDK;
    在所述指定3D运行环境中加载并运行播放插件,通过所述播放插件对基于源文件得到的第一待处理文件和第二待处理文件进行音视频解码,以得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件;
    监控到图像文件存在畸变时,通过所述SDK将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理;
    通过所述录制插件开启录制线程进行录制之前,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享。
  20. 根据权利要求19所述的服务器,其中,所述录制插件包括:与所述指定3D运行环境显存相关的第一插件和响应录制操作的第二插件;
    所述服务器包括:
    通信单元,用于通过所述第一插件与所述显存进行通信,从所述显存中提取出所述畸变处理结果;
    录制单元,用于响应录制操作,所述第二插件与所述第一插件进行通信,将所述畸变处理结果提供给录制线程进行录制,将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
  21. 一种计算机存储介质,所述计算机存储介质中存储有计算机可执行指令,所述计算机可执行指令用于执行权利要求1至7、15-17任一项所述的文件处理方法。
  22. 一种文件处理方法,应用于终端,终端包括有一个或多个处理器以及存储器,以及一个或一个以上的程序,其中,所述一个或一个以上的 程序存储于存储器中,所述程序可以包括一个或一个以上的每一个对应于一组指令的模块,所述一个或多个处理器被配置为执行指令;所述方法包括:
    获取文件转换逻辑,所述文件转换逻辑为在服务器侧根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件所生成的逻辑;
    播放基于源文件得到的第一待处理文件和第二待处理文件,所述第一待处理文件为显示于终端屏幕上源文件的左视频文件,所述第二待处理文件为显示于终端屏幕上源文件的右视频文件;
    根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件,将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享;
    接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
  23. 根据权利要求22所述的方法,其中,所述根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件,包括:
    根据所述文件转换逻辑对第一待处理文件和第二待处理文件分别进行音视频解码,得到对应第一待处理文件的第一音频文件和第一图像文件,及对应第二待处理文件的第二音频文件和第二图像文件。
  24. 根据权利要求23所述的方法,其中,将分别对应第一待处理文件 和第二待处理文件的所述图像文件进行畸变处理,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享,包括:
    获取与终端屏幕显示有关的第一信息;
    获取与终端自身性能有关的第二信息;
    根据所述第一信息和/或所述第二信息对所述第一图像文件和所述第二图像文件进行畸变处理,得到与第一图像文件相关的第三图像文件和与第二图像文件相关的第四图像文件,将所述第三图像文件和所述第四图像文件存储于显存中进行文件共享。
  25. 根据权利要求24所述的方法,其中,所述第一信息包括:屏幕宽高比、分辨率中的至少一种信息;
    所述第二信息包括:终端的型号、终端处理器性能指标中的至少一种信息。
  26. 根据权利要求24所述的方法,其中,所述接收录制操作,将所述畸变处理结果提供给录制线程进行录制时,根据所述文件转换逻辑将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于VR播放的第一目标文件和第二目标文件,包括:
    接收录制操作,将所述第三图像文件和所述第四图像文件提供给录制线程进行录制时,根据所述文件转换逻辑将所述第一音频文件和所述第三图像文件编码合成为第一目标文件,将所述第二音频文件和所述第四图像文件编码合成为第二目标文件。
  27. 根据权利要求22至26任一项所述的方法,其中,所述方法还包括:
    接收文件分享操作,将所述第一目标文件和所述第二目标文件按照目标用户地址分享给所述目标用户;或者,将所述第一目标文件和所述第二目标文件按照网页连接地址分享到对应的网页上。
  28. 根据权利要求22至26任一项所述的方法,其中,所述方法还包括:
    将所述第一目标文件和所述第二目标文件以相同格式的视频文件进行存储。
  29. 一种文件处理方法,应用于服务器,服务器包括有一个或多个处理器以及存储器,以及一个或一个以上的程序,其中,所述一个或一个以上的程序存储于存储器中,所述程序可以包括一个或一个以上的每一个对应于一组指令的模块,所述一个或多个处理器被配置为执行指令;所述方法包括:
    根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件生成文件转换逻辑;
    响应终端发起的下载请求,反馈所述文件转换逻辑给终端。
  30. 根据权利要求29所述的方法,其中,所述根据在指定3D运行环境中加载运行的播放插件、软件开发工具包SDK和录制插件生成文件转换逻辑,包括:
    在所述指定3D运行环境中导入所述SDK;
    在所述指定3D运行环境中加载并运行播放插件,通过所述播放插件对基于源文件得到的第一待处理文件和第二待处理文件进行音视频解码,以得到分别对应第一待处理文件和第二待处理文件的音频文件和图像文件;
    监控到图像文件存在畸变时,通过所述SDK将分别对应第一待处理文件和第二待处理文件的所述图像文件进行畸变处理;
    通过所述录制插件开启录制线程进行录制之前,将得到的畸变处理结果以图像文件格式存储于显存中进行文件共享。
  31. 根据权利要求30所述的方法,其中,所述录制插件包括:与所述指定3D运行环境显存相关的第一插件和响应录制操作的第二插件;
    通过所述第一插件与所述显存进行通信,从所述显存中提取出所述畸变处理结果;
    响应录制操作,所述第二插件与所述第一插件进行通信,将所述畸变处理结果提供给录制线程进行录制,将所述畸变处理结果和分别对应第一待处理文件和第二待处理文件的所述音频文件编码合成为直接用于虚拟现实VR播放的第一目标文件和第二目标文件。
PCT/CN2017/110949 2016-11-16 2017-11-14 一种文件处理方法、终端及服务器、计算机存储介质 WO2018090911A1 (zh)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/297,900 US10855969B2 (en) 2016-11-16 2019-03-11 File processing method, terminal, server, and computer storage medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201611009068.1 2016-11-16
CN201611009068.1A CN108111874B (zh) 2016-11-16 2016-11-16 一种文件处理方法、终端及服务器

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/297,900 Continuation US10855969B2 (en) 2016-11-16 2019-03-11 File processing method, terminal, server, and computer storage medium

Publications (1)

Publication Number Publication Date
WO2018090911A1 true WO2018090911A1 (zh) 2018-05-24

Family

ID=62146082

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/110949 WO2018090911A1 (zh) 2016-11-16 2017-11-14 一种文件处理方法、终端及服务器、计算机存储介质

Country Status (3)

Country Link
US (1) US10855969B2 (zh)
CN (1) CN108111874B (zh)
WO (1) WO2018090911A1 (zh)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113595869A (zh) * 2021-06-28 2021-11-02 青岛海尔科技有限公司 语音播放方法和装置、存储介质及电子装置

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112823519B (zh) * 2018-11-26 2023-10-13 深圳市欢太科技有限公司 视频解码方法、装置、电子设备、计算机可读存储介质
CN110597773B (zh) * 2019-08-28 2022-07-29 北京梦想绽放科技有限公司 在计算机设备和虚拟现实设备之间共享文件的方法和装置
CN112218140A (zh) * 2020-09-02 2021-01-12 中国第一汽车股份有限公司 视频同步播放方法、装置、***和存储介质
CN112862978A (zh) * 2020-12-31 2021-05-28 江西格灵如科科技有限公司 一种vr课件终端及***
CN114501079A (zh) * 2022-01-29 2022-05-13 京东方科技集团股份有限公司 用于对多媒体数据进行处理的方法及相关设备
CN116055481B (zh) * 2023-04-03 2023-06-02 银河航天(北京)网络技术有限公司 一种共享文件的方法、***、装置及存储介质

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN205071176U (zh) * 2015-07-22 2016-03-02 深圳多新哆技术有限责任公司 一种信号转换装置
US20160101356A1 (en) * 2014-01-02 2016-04-14 Ubitus Inc. System And Method For Delivering Media Over Network
CN105872521A (zh) * 2016-04-26 2016-08-17 乐视控股(北京)有限公司 2d视频播放方法及装置
CN105892643A (zh) * 2015-12-31 2016-08-24 乐视致新电子科技(天津)有限公司 基于虚拟现实的多界面统一显示***及方法

Family Cites Families (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101808250B (zh) * 2009-02-13 2012-10-24 北京邮电大学 基于双路视觉的立体影像合成方法及***
CA2857691A1 (en) * 2011-12-02 2013-06-06 Jerry G. Aguren Wide field-of-view 3d stereo vision platform with dynamic control of immersive or heads-up display operation
JP6051605B2 (ja) * 2012-06-13 2016-12-27 ソニー株式会社 表示装置、および表示制御方法、並びにプログラム
US9538160B1 (en) * 2013-04-11 2017-01-03 Nextvr Inc. Immersive stereoscopic video acquisition, encoding and virtual reality playback methods and apparatus
CN103533305B (zh) * 2013-10-10 2018-01-12 国电南瑞科技股份有限公司 一种b/s架构的免插件通用视频监控***
US10182257B2 (en) * 2014-07-31 2019-01-15 Clipchamp Ip Pty Ltd Client-side video transcoding and processing
US10049495B2 (en) * 2015-01-14 2018-08-14 Hashplay Inc. System and method for providing virtual reality content
US10419737B2 (en) * 2015-04-15 2019-09-17 Google Llc Data structures and delivery methods for expediting virtual reality playback
US10491711B2 (en) * 2015-09-10 2019-11-26 EEVO, Inc. Adaptive streaming of virtual reality data
CN105455285B (zh) * 2015-12-31 2019-02-12 北京小鸟看看科技有限公司 一种虚拟现实头盔适配方法
US20170206708A1 (en) * 2016-01-19 2017-07-20 Immersv, Inc. Generating a virtual reality environment for displaying content
JP6708444B2 (ja) * 2016-03-04 2020-06-10 キヤノン株式会社 画像処理装置および画像処理方法
KR101870865B1 (ko) * 2016-04-15 2018-06-26 재단법인 실감교류인체감응솔루션연구단 삼차원 정보증강 비디오 씨쓰루 디스플레이 장치 및 방법, 렉티피케이션 장치
KR20180011539A (ko) * 2016-07-25 2018-02-02 삼성전자주식회사 영상의 처리를 위한 전자 장치
CA3034384A1 (en) * 2016-08-22 2018-03-01 Magic Leap, Inc. Virtual, augmented, and mixed reality systems and methods
JP6732617B2 (ja) * 2016-09-21 2020-07-29 株式会社ソニー・インタラクティブエンタテインメント 情報処理装置および画像生成方法
US10445925B2 (en) * 2016-09-30 2019-10-15 Sony Interactive Entertainment Inc. Using a portable device and a head-mounted display to view a shared virtual reality space

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160101356A1 (en) * 2014-01-02 2016-04-14 Ubitus Inc. System And Method For Delivering Media Over Network
CN205071176U (zh) * 2015-07-22 2016-03-02 深圳多新哆技术有限责任公司 一种信号转换装置
CN105892643A (zh) * 2015-12-31 2016-08-24 乐视致新电子科技(天津)有限公司 基于虚拟现实的多界面统一显示***及方法
CN105872521A (zh) * 2016-04-26 2016-08-17 乐视控股(北京)有限公司 2d视频播放方法及装置

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113595869A (zh) * 2021-06-28 2021-11-02 青岛海尔科技有限公司 语音播放方法和装置、存储介质及电子装置
CN113595869B (zh) * 2021-06-28 2023-10-24 青岛海尔科技有限公司 语音播放方法和装置、存储介质及电子装置

Also Published As

Publication number Publication date
CN108111874A (zh) 2018-06-01
CN108111874B (zh) 2020-01-31
US10855969B2 (en) 2020-12-01
US20190208178A1 (en) 2019-07-04

Similar Documents

Publication Publication Date Title
WO2018090911A1 (zh) 一种文件处理方法、终端及服务器、计算机存储介质
US10439678B2 (en) Method and system for transfering data between plurality of devices
CN106454121B (zh) 双摄像头拍照方法及装置
CN104902075B (zh) 多屏互动方法及***
CN106484416B (zh) 一种信息处理方法及终端
KR101788060B1 (ko) 영상표시장치 및 이를 이용한 콘텐츠 관리방법
WO2017202271A1 (zh) 一种信息处理方法及终端、计算机存储介质
CN106502693A (zh) 一种图像显示方法和装置
WO2020119464A1 (zh) 一种视频拆分方法及电子设备
CN106485689A (zh) 一种图片处理方法和装置
CN106453939B (zh) 一种流媒体接入方法及装置
CN106686301A (zh) 照片拍摄方法及装置
CN105827866A (zh) 一种移动终端及控制方法
KR20120027876A (ko) 이동 단말기 및 그의 3d 이미지 합성방법
CN106484470A (zh) 应用程序安装包的处理方法、装置及服务器
WO2020034150A1 (zh) 一种广告图片的显示方法、上传方法及装置
WO2017054616A1 (zh) 一种视频图片显示方法、装置及一种图片显示方法
CN106373110A (zh) 一种图像融合的方法及装置
US20160295256A1 (en) Digital content streaming from digital tv broadcast
CN109168029B (zh) 一种调整分辨率的方法、设备和计算机可存储介质
CN107454464B (zh) 一种信息处理方法及终端
WO2018040751A1 (zh) 图像生成装置及方法、图像处理设备及存储介质
WO2017185800A1 (zh) 一种控制方法及电子设备、存储介质
CN106709882A (zh) 一种图像融合的方法及装置
CN107992505B (zh) 一种网页信息处理方法及终端

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17871582

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17871582

Country of ref document: EP

Kind code of ref document: A1