WO2024017125A9 - 一种信号传输方法及装置 - Google Patents

一种信号传输方法及装置 Download PDF

Info

Publication number
WO2024017125A9
WO2024017125A9 PCT/CN2023/107128 CN2023107128W WO2024017125A9 WO 2024017125 A9 WO2024017125 A9 WO 2024017125A9 CN 2023107128 W CN2023107128 W CN 2023107128W WO 2024017125 A9 WO2024017125 A9 WO 2024017125A9
Authority
WO
WIPO (PCT)
Prior art keywords
message
video
signal
valid
audio
Prior art date
Application number
PCT/CN2023/107128
Other languages
English (en)
French (fr)
Other versions
WO2024017125A1 (zh
Inventor
周伟光
韦志贵
何剑鸿
Original Assignee
华为技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为技术有限公司 filed Critical 华为技术有限公司
Publication of WO2024017125A1 publication Critical patent/WO2024017125A1/zh
Publication of WO2024017125A9 publication Critical patent/WO2024017125A9/zh

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/16Implementation or adaptation of Internet protocol [IP], of transmission control protocol [TCP] or of user datagram protocol [UDP]
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/0803Configuration setting
    • H04L41/0823Configuration setting characterised by the purposes of a change of settings, e.g. optimising configuration for enhancing reliability
    • H04L41/0826Configuration setting characterised by the purposes of a change of settings, e.g. optimising configuration for enhancing reliability for reduction of network costs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/08Configuration management of networks or network elements
    • H04L41/0803Configuration setting
    • H04L41/0823Configuration setting characterised by the purposes of a change of settings, e.g. optimising configuration for enhancing reliability
    • H04L41/0833Configuration setting characterised by the purposes of a change of settings, e.g. optimising configuration for enhancing reliability for reduction of network energy consumption
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/63Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
    • H04N21/647Control signaling between network components and server or clients; Network processes for video distribution between server and clients, e.g. controlling the quality of the video stream, by dropping packets, protecting content from unauthorised alteration within the network, monitoring of network load, bridging between two different networks, e.g. between IP and wireless
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/16Analogue secrecy systems; Analogue subscription systems

Definitions

  • the embodiments of the present application relate to the field of audio and video processing technology, and in particular, to a signal transmission method and device.
  • the video signal when video signals are transmitted between video source devices such as computers, laptops, mobile phones, and tablet computers and video sink devices such as monitors, the video signal must first be converted into a DP stream through the display port (DP) controller in the video source device, and then converted into a USB4/TBT message through the DP input (IN) adapter in the universal serial bus (USB)4/Thunderbolt (TBT) router (router), and then transmitted to the video sink device through the USB4/TBT network.
  • the DP output (OUT) adapter in the USB4/TBT router in the video sink device restores it to a DP stream based on the message, and then restores it to a video signal through the DP controller.
  • the video signal needs to be processed by the DP controller and then sent to the USB/TBT adapter for processing, which increases the transmission delay of the video signal.
  • both the video source device and the video sink device need to integrate hardware devices such as the DP controller, which will also bring additional costs and power consumption.
  • the embodiments of the present application provide a signal transmission method and apparatus to reduce the transmission delay of a video signal, and reduce the cost and power consumption of a video signal sending device (such as a video source device) and a receiving device (a video sink device).
  • a video signal sending device such as a video source device
  • a receiving device such as a video sink device
  • the embodiment of the present application provides a signal transmission method, which can be applied to a first device, or a component of the first device (such as a processor, a chip, or a chip system, etc.), or a device used in combination with the first device, etc.
  • the first device can be a video source device such as a set-top box, a game console, etc., and can also be a terminal device, a computer, a portable device, etc. that sends video signals, etc.
  • the following is taken as an example for application to the first device.
  • the method includes: obtaining a video signal, the video signal includes a vertical synchronization signal, a horizontal synchronization signal, a pixel clock signal, a display enable signal, and a video data signal; obtaining valid video data in the video data signal according to the pixel clock signal and the display enable signal; encapsulating a vertical blanking message according to the vertical synchronization signal, encapsulating a horizontal blanking message according to the horizontal synchronization signal, and encapsulating a valid video message according to the valid video data; and compounding the vertical blanking message, the horizontal blanking message, and the valid video message into a video stream and sending it.
  • the vertical blanking message, the horizontal blanking message, and the valid video message can be compounded into a video stream through a stream multiplexer (such as a video stream multiplexer, an audio and video stream multiplexer).
  • a stream multiplexer such as a video stream multiplexer, an audio and video stream multiplexer.
  • compounding multiple types of messages into a video stream may refer to transmitting multiple types of messages through one video stream, and the multiple types of messages may carry the identifier of the video stream.
  • compounding the vertical blanking message, the horizontal blanking message, and the effective video message into a video stream may refer to transmitting the vertical blanking message, the horizontal blanking message, and the effective video message through one video stream, and one video stream includes the vertical blanking message, the horizontal blanking message, and the effective video message, and the vertical blanking message, the horizontal blanking message, and the effective video message may carry the identifier of the video stream.
  • the video signal can be directly packaged into a message for transmission.
  • the video signal does not need to be processed by the DP controller and then sent to the USB/TBT adapter for packaging and processing, which can reduce the transmission delay of the video signal.
  • sending a video stream includes sending the video stream through a transport layer and a physical layer of a unified multimedia interconnection (UMI).
  • UMI unified multimedia interconnection
  • the video signal can be packaged into a UMI message by an audio-visual sending adapter (or a video sending adapter) in a sending device, and the video stream obtained by compounding the UMI message is sent to the audio-visual receiving adapter (or a video receiving adapter) in the corresponding receiving device through the transport layer and physical layer (including the electrical layer and the logical layer) of the UMI network, and the audio-visual receiving adapter converts the UMI message into a video signal to reduce the cost and power consumption of the video signal sending device and the receiving device.
  • the method also includes: obtaining descriptive information located in a blanking area in a video data signal according to a pixel clock signal and a display enable signal; encapsulating a descriptive information message according to the descriptive information; and combining the vertical blanking message, the horizontal blanking message, and the valid video message into a video stream and sending it, including: combining the vertical blanking message, the horizontal blanking message, the valid video message, and the descriptive information message into a video stream and sending it.
  • encapsulating the descriptive information message according to the descriptive information includes: when the video signal also includes an associated audio signal, encapsulating the descriptive information message according to the descriptive information and the audio signal.
  • the accompanying data such as the configuration and control information of audio and video
  • the method also includes: when the video signal also includes an associated audio signal, encapsulating an audio sampling message according to the audio signal; compounding the vertical blanking message, the horizontal blanking message, the valid video message and the descriptive information message into a video stream and sending it, including: compounding the vertical blanking message, the horizontal blanking message, the valid video message, the descriptive information message and the audio sampling message into a video stream and sending it.
  • the audio signal and the descriptive information are transmitted separately, which is beneficial for the receiving end to quickly process the audio signal when there is a pure audio processing device, such as an amplifier (sound system), at the receiving end.
  • a pure audio processing device such as an amplifier (sound system)
  • the method also includes: encrypting the valid video message and the audio sampling message according to the encryption parameters and the encryption key to obtain the encrypted valid video message and the encrypted audio sampling message; encapsulating the encryption description message according to the encryption parameters, and encapsulating the key distribution message according to the encryption key; and compounding the vertical blanking message, the horizontal blanking message, the valid video message, the descriptive information message and the audio sampling message into a video stream and sending it, including: compounding the vertical blanking message, the horizontal blanking message, the encrypted valid video message, the descriptive information message, the encrypted audio sampling message, the encrypted description message and the key distribution message into a video stream and sending it.
  • the above design can achieve confidential transmission of audio and video data and meet the transmission needs of users in different scenarios.
  • vertical blanking messages, encryption description messages, and key distribution messages are repeatedly sent multiple times in a video stream.
  • the vertical blanking message in the video message is used to transmit the vertical synchronization signal, indicating the beginning of a frame of video image, and the encryption description message and key distribution message are used to decrypt the valid video data, if these messages are transmitted incorrectly, the problem of incorrectly outputting the video signal may occur.
  • sending multiple messages repeatedly can improve the reliability of video signal transmission.
  • the message headers of the vertical blanking message, encryption description message, and key distribution message include a retry flag.
  • the retry flag is 1, it is used to instruct the receiving end to check whether the message is correct and initiate a retransmission request if the message is incorrect.
  • the receiving end can be instructed to check whether the vertical blanking message, encryption description message and key distribution message are correct, thereby improving the reliability of signal transmission.
  • a vertical blanking message is encapsulated according to a vertical synchronization signal, including: encapsulating a vertical blanking message according to a vertical synchronization signal and video frame control information, wherein the video frame control information includes a pixel clock parameter of a pixel clock signal.
  • the video frame control information is carried in the vertical blanking message, which is beneficial for the receiving end to recover the video signal.
  • encapsulating a vertical blanking message according to a vertical synchronization signal includes encapsulating a vertical blanking message according to a vertical synchronization signal and a display enable (DE) signal.
  • DE display enable
  • a horizontal blanking message may be encapsulated according to a horizontal blanking region (HBlank) end signal.
  • the horizontal blanking message may also include a vertical blanking region flag (VBlank_tage).
  • the embodiment of the present application provides a signal transmission method, which can be applied to a second device, or a component of the second device (such as a processor, a chip, or a chip system, etc.), or a device used in combination with the second device, etc.
  • the second device can be a video sink device such as a display, a television, etc., and can also be a terminal device, a computer, a portable device, etc. for receiving video signals, etc. The following is taken as an example for application to the second device.
  • the method includes: receiving a video stream, the video stream includes a vertical blanking message, a horizontal blanking message, and a valid video message; generating a vertical synchronization signal according to the vertical blanking message, and generating a horizontal synchronization signal according to the horizontal blanking message; generating a video data signal and a display enable signal according to the vertical synchronization signal, the horizontal synchronization signal, the pixel clock signal, and the valid video data carried by the valid video message; outputting a video signal, the video signal includes a vertical synchronization signal, a horizontal synchronization signal, a pixel clock signal, a display enable signal, and a video data signal.
  • the video stream can be demultiplexed by a stream demultiplexer (such as a video stream demultiplexer, an audio and video stream demultiplexer), etc., to obtain a vertical blanking message, a horizontal blanking message, and a valid video message.
  • a stream demultiplexer such as a video stream demultiplexer, an audio and video stream demultiplexer, etc.
  • demultiplexing the video stream to obtain multiple types of messages may refer to, according to the identifier of the message, Identify (or divide) multiple types of messages in a video stream.
  • Identify or divide multiple types of messages in a video stream.
  • demultiplex the video stream to obtain vertical blanking messages, horizontal blanking messages and valid video messages which may refer to identifying (or dividing) vertical blanking messages, horizontal blanking messages and valid video messages in the video stream according to the identifier of the vertical blanking message (such as 4b'0001), the identifier of the horizontal blanking message (such as 4b'0010) and the identifier of the valid video message (such as 4b'0101).
  • receiving a video stream includes: receiving the video stream through a transport layer and a physical layer of a UMI.
  • the video stream also includes a descriptive information message, which generates a video data signal and a display enable signal according to the vertical synchronization signal, the horizontal synchronization signal, the pixel clock signal, and the valid video data carried by the valid video message, including: generating a video data signal and a display enable signal according to the vertical synchronization signal, the horizontal synchronization signal, the pixel clock signal, the valid video data carried by the valid video message, and the descriptive information carried by the descriptive information message.
  • a descriptive information message which generates a video data signal and a display enable signal according to the vertical synchronization signal, the horizontal synchronization signal, the pixel clock signal, and the valid video data carried by the valid video message, including: generating a video data signal and a display enable signal according to the vertical synchronization signal, the horizontal synchronization signal, the pixel clock signal, the valid video data carried by the valid video message, and the descriptive information carried by the descriptive information message.
  • the method further includes: generating an audio signal based on the descriptive information message, wherein the output video signal also includes the audio signal.
  • the video stream also includes an audio sampling message
  • the method further includes: generating an audio signal according to the audio sampling message; wherein the output video signal also includes an audio signal.
  • the video stream also includes an encrypted description message and a key distribution message
  • the valid video message is an encrypted valid video message.
  • the method also includes: obtaining encryption parameters carried by the encrypted description message, obtaining an encryption key carried by the key distribution message; decrypting the encrypted valid video message according to the encryption parameters and the encryption key to obtain a decrypted valid video message.
  • the video stream also includes an encrypted description message and a key distribution message
  • the valid video message and the audio sampling message are encrypted valid video messages and encrypted audio sampling messages.
  • the method also includes: obtaining encryption parameters carried by the encrypted description message, obtaining encryption keys carried by the key distribution message; decrypting the encrypted valid video message and the encrypted audio sampling message according to the encryption parameters and the encryption key to obtain decrypted valid video messages and decrypted audio sampling messages.
  • the video stream includes multiple vertical blanking messages, encrypted description messages, and key distribution messages that are sent repeatedly and continuously
  • the method also includes: selecting vertical blanking messages, encrypted description messages, and key distribution messages that pass the correctness check from the multiple vertical blanking messages, encrypted description messages, and key distribution messages that are sent repeatedly and continuously.
  • the message headers of the vertical blanking message, the encryption description message, and the key distribution message include a retry flag.
  • the Retry flag is 1, it is used to instruct the receiving end to check whether the message is correct and to initiate a retransmission request when the message is incorrect.
  • the method also includes: when any of the vertical blanking message, the encryption description message, and the key distribution message is incorrect, initiating retransmission of the message.
  • the method also includes: obtaining video frame control information carried by a vertical blanking message, the video frame control information including a pixel clock parameter; and generating a pixel clock signal according to the video frame control information.
  • an embodiment of the present application provides a signal transmission device, which has the function of implementing the method in the first aspect above, and the function can be implemented by hardware, or by hardware executing corresponding software.
  • the hardware or software includes one or more modules corresponding to the above functions, such as an interface unit and a processing unit.
  • the device may be a chip or an integrated circuit.
  • the device includes a memory and a processor, the memory is used to store instructions executed by the processor, and when the instructions are executed by the processor, the device can execute the method of the first aspect mentioned above.
  • the device may be a complete set of first equipment.
  • an embodiment of the present application provides a signal transmission device, which has the function of implementing the method in the second aspect above, and the function can be implemented by hardware, or by hardware executing corresponding software.
  • the hardware or software includes one or more modules corresponding to the above functions, such as an interface unit and a processing unit.
  • the device may be a chip or an integrated circuit.
  • the device includes a memory and a processor, the memory is used to store instructions executed by the processor, and when the instructions are executed by the processor, the device can execute the method of the second aspect mentioned above.
  • the apparatus may be a complete set of the second device.
  • an embodiment of the present application provides a signal transmission device, which includes an interface circuit and a processor, and the processor and the interface circuit are coupled to each other.
  • the processor is used to implement the method of the first aspect through a logic circuit or an execution instruction.
  • the interface circuit is used to receive a signal from other signal transmission devices other than the signal transmission device and transmit it to the processor or send a signal from the processor to other signal transmission devices other than the signal transmission device. It can be understood that the interface circuit can be a transceiver or a transceiver or a transceiver or an input-output interface.
  • the signal transmission device may further include a memory for storing instructions executed by the processor or storing input data required by the processor to execute instructions or storing data generated after the processor executes instructions.
  • the memory may be a physically independent unit or may be coupled to the processor, or the processor may include the memory.
  • an embodiment of the present application provides a signal transmission device, which includes an interface circuit and a processor, and the processor and the interface circuit are coupled to each other.
  • the processor is used to implement the method of the second aspect above through a logic circuit or an execution instruction.
  • the interface circuit is used to receive a signal from other signal transmission devices other than the signal transmission device and transmit it to the processor or send a signal from the processor to other signal transmission devices other than the signal transmission device. It can be understood that the interface circuit can be a transceiver or a transceiver or a transceiver or an input-output interface.
  • the signal transmission device may further include a memory for storing instructions executed by the processor or storing input data required by the processor to execute instructions or storing data generated after the processor executes instructions.
  • the memory may be a physically independent unit or may be coupled to the processor, or the processor may include the memory.
  • an embodiment of the present application provides a communication system, which includes a first device and a second device.
  • the first device can implement the method of the first aspect above
  • the second device can implement the method of the second aspect above.
  • an embodiment of the present application provides a computer-readable storage medium, in which a computer program or instruction is stored.
  • a computer program or instruction is stored.
  • the method of the first aspect or the second aspect mentioned above can be implemented.
  • an embodiment of the present application further provides a computer program product, including a computer program or instructions, which, when executed by a processor, can implement the method of the first or second aspect above.
  • an embodiment of the present application also provides a chip system, which includes: a processor and a memory, the processor is coupled to the memory, the memory is used to store programs or instructions, and when the program or instructions are executed by the processor, the method of the first or second aspect mentioned above can be implemented.
  • FIG1 is a schematic diagram of an image display provided in an embodiment of the present application.
  • FIG2 is a schematic diagram of video signal generation provided in an embodiment of the present application.
  • FIG3 is a schematic diagram of video transmission provided in an embodiment of the present application.
  • FIG4 is a schematic diagram of a signal transmission method according to an embodiment of the present application.
  • FIG5 is a schematic diagram of bidirectional multi-time audio stream transmission provided by an embodiment of the present application.
  • FIG6 is a second schematic diagram of a signal transmission method provided in an embodiment of the present application.
  • FIG7 is a schematic diagram of timing parameters of a video signal provided in an embodiment of the present application.
  • FIG8 is one of the logic diagrams of the video transmission adapter provided in an embodiment of the present application.
  • FIG9 is one of the schematic diagrams of the video message structure provided in an embodiment of the present application.
  • FIG10 is a schematic diagram of a vertical blanking message format provided in an embodiment of the present application.
  • FIG11 is a schematic diagram of a horizontal blanking message format provided in an embodiment of the present application.
  • FIG12 is a schematic diagram of a valid video message format provided in an embodiment of the present application.
  • FIG13 is a schematic diagram of the arrangement of valid video data provided in an embodiment of the present application.
  • FIG14 is a schematic diagram of the arrangement of pixel components of a video in RGB format provided in an embodiment of the present application.
  • FIG15 is a schematic diagram of the arrangement of pixel components of a YCbCr444 format video provided in an embodiment of the present application.
  • FIG16 is a schematic diagram of the arrangement of pixel components of a YCbCr422 format video provided in an embodiment of the present application;
  • FIG17 is one of the data arrangement diagrams of the video provided in the embodiment of the present application.
  • FIG18 is one of schematic diagrams of valid video messages provided in an embodiment of the present application.
  • FIG19 is a second schematic diagram of video data arrangement provided in an embodiment of the present application.
  • FIG20 is a second schematic diagram of a valid video message provided in an embodiment of the present application.
  • FIG21 is a schematic diagram of a descriptive information message format provided in an embodiment of the present application.
  • FIG22 is a schematic diagram of a descriptive information message encapsulation provided in an embodiment of the present application.
  • FIG23 is a schematic diagram of audio sampling message encapsulation provided in an embodiment of the present application.
  • FIG24 is a schematic diagram of an audio sampling message provided in an embodiment of the present application.
  • FIG25 is a second schematic diagram of a video message structure provided in an embodiment of the present application.
  • FIG26 is one of the logic diagrams of the video receiving adapter provided in an embodiment of the present application.
  • FIG27 is a schematic diagram of clock recovery provided in an embodiment of the present application.
  • FIG28 is a schematic diagram of video signal transmission and recovery provided by an embodiment of the present application.
  • FIG29 is a schematic diagram of video data filling provided in an embodiment of the present application.
  • FIG30 is one of the schematic diagrams of the structure of a communication device provided in an embodiment of the present application.
  • FIG31 is a second schematic diagram of the structure of the communication device provided in an embodiment of the present application.
  • Video signal timing originates from cathode ray tube (CRT) monitors/televisions. This type of monitor uses an electron gun to emit an electron beam. The electron beam hits the phosphor on the screen to make the phosphor glow. The intensity of the photon beam controls the brightness of the pixel.
  • the electron beam uses raster scanning technology to display the image. Referring to the image display schematic diagram shown in Figure 1, before starting to display a frame of the image, the electron beam needs to return from the lower right corner to the upper left corner. During this process, the electron beam needs to be blanked (no image is displayed). This period of time is called vertical blanking. After displaying a line of data, the electron beam needs to return from the rightmost side of the current line to the leftmost side of the next line. In this process, blanking is also required. This period of time is called horizontal blanking.
  • CTR cathode ray tube
  • Video signal referring to the schematic diagram of video signal generation shown in FIG2, the video signal usually includes several sub-signals such as vertical synchronization signal (VSYNC), horizontal synchronization signal (HSYNC), display enable signal (DE), pixel clock signal, video data signal (or pixel data signal), etc.
  • VSYNC vertical synchronization signal
  • HSELNC horizontal synchronization signal
  • DE display enable signal
  • pixel clock signal video data signal (or pixel data signal), etc.
  • Video data signal when the level is high, it indicates the beginning of a frame of video image; horizontal synchronization signal: when the level is high, it indicates the beginning of a line of video image; display enable signal: when the level is high, it indicates that the current pixel data is valid video pixel data; video data signal (or pixel data signal): pixel data transmission path, usually composed of 24/30/36/48 lines (video signal lines), any transmission path (i.e., video signal line) transmits data corresponding to one pixel in each pixel clock cycle.
  • the video signal when transmitting video signals between a video source device and a video sink device, the video signal needs to be first converted into a DP stream through the DP controller in the video source device, and then converted into a USB4/TBT message through the DP input (IN) adapter in the USB4/TBT router (router), and then transmitted to the video sink device through the USB4/TBT network, and the DP output (OUT) adapter in the USB4/TBT router in the video sink device restores the DP stream according to the message, and then restores it to a video signal through the DP controller.
  • the DP input (IN) adapter in the USB4/TBT router router
  • the DP output (OUT) adapter in the USB4/TBT router in the video sink device restores the DP stream according to the message, and then restores it to a video signal through the DP controller.
  • the video signal needs to be processed by the DP controller and then sent to the USB/TBT adapter for processing, which increases the transmission delay of the video signal, and both the video source device and the video sink device need to integrate an additional DP controller, and the adapter needs to integrate the electrical layer and logic layer of DP, which will bring additional costs and power consumption.
  • the present application provides a signal transmission method and apparatus, which aims to reduce the transmission delay of the video signal and reduce the cost and power consumption of the video source device and the video sink device by directly packaging the video signal into a message for transmission.
  • the audio-visual signal can be converted into a unified multimedia interconnection interface (UMI) message through an audio-visual transmission adapter (or a video transmission adapter).
  • UMI unified multimedia interconnection interface
  • the UMI message is transmitted to the corresponding audio-visual receiving adapter (or a video receiving adapter) through the UMI network, and the audio-visual receiving adapter converts the UMI message into an audio-visual signal.
  • the UMI standard supports bidirectional multi-view and audio service stream transmission, and the arrangement of audio-visual data in the adaptation layer is independent of the transport layer. Bidirectional multi-time audio stream transmission based on UMI can be shown in FIG5.
  • the adaptation layer may include one or more audio-visual transmission adapters and one or more audio-visual receiving adapters.
  • the audio-visual signal is packaged into a UMI message by the audio-visual transmission adapter of the adaptation layer
  • the audio-visual stream composed of the UMI message is transmitted through the transport layer and the physical layer (including the electrical layer and the logical layer) and is received by the audio-visual receiving adapter at the receiving end, and the audio-visual receiving adapter converts the UMI message into an audio-visual signal.
  • the video and audio signals are not limited to being transmitted via UMI, but can also be supported in general high-speed data transmission interfaces such as PCIe.
  • PCIe general high-speed data transmission interfaces
  • ordinal numbers such as “first” and “second” mentioned in the embodiments of the present application are used to distinguish multiple objects, and are not used to limit the size, content, order, timing, priority or importance of multiple objects.
  • the first device and the second device do not mean that the priority or importance of the two devices is different.
  • the number of nouns means “singular noun or plural noun", that is, “one or more”.
  • At least one means one or more
  • plural means two or more.
  • “And/or” describes the association relationship of associated objects, indicating that three relationships may exist.
  • a and/or B can mean: A exists alone, A and B exist at the same time, and B exists alone, where A and B can be singular or plural.
  • the character "/” generally indicates that the related objects before and after are in an "or” relationship.
  • A/B means: A or B.
  • “At least one of the following" or similar expressions refers to any combination of these items, including any combination of single or plural items.
  • At least one of a, b, or c means: a, b, c, a and b, a and c, b and c, or a and b and c, where a, b, and c can be single or plural.
  • Figure 6 is a schematic diagram of a signal transmission method provided by an embodiment of the present application.
  • Figure 6 takes the first device and the second device as the execution subjects as an example to illustrate the method, wherein the first device can be a set-top box, a game console or other video source device, the second device can be a display, a television or other video sink device, the first device can also be a terminal device, a computer, a portable device, etc. for sending video signals, etc., the second device can also be a terminal device, a computer, a portable device, etc. for receiving video signals, etc., and the present application does not limit the execution subject of the method.
  • the first device in Figure 6 can also be a component of the first device (such as a processor, a chip, or a chip system, etc.), or a device used in combination with the first device;
  • the second device in Figure 6 can also be a component of the second device (such as a processor, a chip, or a chip system, etc.), or a device used in combination with the second device.
  • the method includes:
  • the first device acquires a video signal, where the video signal includes a vertical synchronization signal, a horizontal synchronization signal, a pixel clock signal, a display enable signal, and a video data signal.
  • the first device obtains valid video data in the video data signal according to the pixel clock signal and the display enable signal.
  • the first device encapsulates a vertical blanking message according to a vertical synchronization signal and a horizontal blanking area (HBlank) end signal, encapsulates a horizontal blanking message according to an HBlank end signal, and encapsulates a valid video message according to valid video data.
  • HBlank horizontal blanking area
  • the horizontal blanking interval end signal may also be referred to as a horizontal blanking end signal.
  • the first device combines the vertical blanking message, the horizontal blanking message and the effective video message into a video stream and sends it.
  • the second device receives the video stream.
  • the video signal may be generated by the first device, for example, by a video signal generator in the first device, or may be obtained by the first device from other devices, which is not limited in the present application.
  • the video signal includes video sub-signals: a horizontal synchronization signal (HSYNC), a vertical synchronization signal (VSYNC), a display enable signal (DE), a pixel clock signal, and a video data signal video sub-signal (rectangular area in FIG. 7 ).
  • the video signal timing parameter refers to the relationship between the video sub-signals in time, and the specific timing parameters are shown in Table 1.
  • Each transmission path of the video data signal (or pixel data signal) transmits data corresponding to one pixel in each pixel clock cycle.
  • the video data signal (or pixel data signal) has no transmission of valid pixels in the vertical blanking area and the horizontal blanking area.
  • the display enable signal is at a high level, it indicates that the current pixel data is transmitting valid video pixel data. Therefore, in an embodiment of the present application, valid video data (i.e., valid pixel data) in the video data signal can be obtained based on the pixel clock signal and the display enable signal. This reduces the transmission bandwidth of the video signal.
  • the first device after acquiring the video signal, can acquire (or capture) the valid video data in the video data signal according to the pixel clock signal and the display enable signal, and encapsulate the active video packet (active video packet, AVP) according to the valid video data, and the active video packet AVP can carry the valid video data.
  • the first device can also encapsulate the vertical blanking packet (vertical blanking packet, VBP) according to the vertical synchronization signal, and encapsulate the horizontal blanking packet (horizontal blanking packet, HBP) according to the horizontal synchronization signal.
  • a vertical blanking packet VBP is generated, indicating the start of the vertical synchronization signal, that is, indicating the start of a frame of video image
  • a horizontal blanking packet HBP is generated, indicating the start of a line of video image.
  • the first device may also obtain (or capture) descriptive information in a blanking area (such as a vertical blanking area and a horizontal blanking area) in the video data signal according to the pixel clock signal and the display enable signal; encapsulate a descriptive information packet (DIP) according to the descriptive information, and the descriptive information packet DIP may carry the descriptive information.
  • the first device may also multiplex the descriptive information packet DIP (such as by multiplexing through a stream multiplexer, etc.) and send it in the video stream.
  • the first device can also start content protection (such as encryption), encrypt the valid video message AVP (such as advanced digital content protection system technology (ADCP) encryption, etc.), and can encapsulate the key distribution message (key distribution packet, KDP) according to the encryption key used in encryption, the encryption algorithm/encryption method used in encryption, and other encryption parameters to encapsulate the encryption description message (encryption description packet, EDP), and the encrypted valid video message AVP', descriptive information message, key distribution message KDP carrying encryption key information, encryption description message EDP carrying encryption parameters, and vertical blanking message VBP, horizontal blanking message HBP, etc.
  • content protection such as encryption
  • AVP advanced digital content protection system technology
  • the video stream can be sent through the UMI network.
  • the receiving second device or the video receiving adapter in the second device
  • the transport layer and physical layer including the electrical layer and the logical layer
  • the video stream may not include the encryption description message EDP and the key distribution message KDP.
  • multiple types of messages are compounded (such as by a stream multiplexer, etc.) into a video stream, which may refer to transmitting multiple types of messages through one video stream, and the multiple types of messages may carry the identification of the video stream.
  • the above-mentioned compounding of the vertical blanking message, the horizontal blanking message, and the effective video message into a video stream may refer to transmitting the vertical blanking message, the horizontal blanking message, and the effective video message through one video stream, including the vertical blanking message, the horizontal blanking message, and the effective video message in one video stream, and the vertical blanking message, the horizontal blanking message, and the effective video message may carry the identification of the video stream.
  • the first device may encapsulate the audio signal (or audio data) and the descriptive information together, such as in a descriptive information message DIP.
  • the first device may transmit the audio signal (or audio data) and the descriptive information separately.
  • the first device may encapsulate the descriptive information message DIP only based on the descriptive information, and may encapsulate the audio sampling message (audio sample packet, ASP) separately based on the audio signal, and carry the information of the audio signal (or the audio data corresponding to the audio signal) in the audio sampling message ASP.
  • the first device may also encrypt the audio sampling message ASP according to the transmission requirements, and transmit the encrypted audio sampling message ASP’.
  • the first device may also encrypt one or more messages of the valid video message AVP and the audio sampling message ASP as required.
  • the first device may set the content protection identifier (such as the CP flag) in the message header to 1 or 0 according to the content protection (such as encryption) requirements.
  • the first device (such as the content protection module in the first device) may process according to the CP flag in the message header. If the CP flag is 1, the message is encrypted. If the CP flag is 0, it is not encrypted.
  • the structure of a video message (such as a vertical blanking message, a horizontal blanking message, an effective video message, etc.) in a video stream may be as shown in FIG. 9 , and the video message may include a message header of 4 bytes (a total of 32 bits) and a payload portion.
  • the meaning of each field in the message header of the video message may be as shown in Table 2.
  • Video packets can be divided into vertical blanking packets VBP, horizontal blanking packets HBP, descriptive information packets DIP, effective video packets AVP, encryption description packets EDP, key distribution packets KDP, etc. according to the packet type (or function).
  • VBP Vertical blanking message
  • the vertical blanking message VBP marks the beginning of a frame.
  • the format of the vertical blanking message VBP can be shown in Figure 10, including a message header of 4 bytes and a message body of 28 bytes, a total of 32 bytes.
  • the first device such as the video transmission adapter in the first device
  • receives the VBS signal it immediately generates a vertical blanking message VBP.
  • the meaning of each field in the message header of the vertical blanking message VBP can be shown in Table 3.
  • the message body (ie, payload part) of the vertical blanking message VBP can carry frame-level control information (video frame control, VFC) for describing information related to the video frame.
  • VFC video frame control
  • the timing, color space, and color depth of the current video frame can be carried.
  • the message body (ie, VFC) information of the vertical blanking message can be as shown in Table 4, including the number of horizontal active pixels (HActivePixels), the number of horizontal blanking area pixels (HBlankPixels), the number of horizontal blanking area front shoulder pixels (HFrontPorch), the number of HSync pixels (HSyncPixels), the number of video frame active lines (VActiveLines), the number of vertical blanking area lines (VBlankLines), the number of vertical blanking area front shoulder lines (VFrontPorch), and the number of VSync lines (VSyncLines).
  • the horizontal blanking message HBP can be used to transmit a line start signal.
  • the horizontal blanking message HBP is immediately generated.
  • the format of the horizontal blanking message HBP can be shown in FIG. 11, wherein the horizontal blanking message can only have a message header, and the meaning of each field of the message header can be shown in Table 5.
  • Effective Video Message AVP is used to transmit effective video data (also called effective video pixel data) of the video signal.
  • effective video data also called effective video pixel data
  • the format of the effective video message AVP may be shown in FIG. 12.
  • the effective video message may be composed of a message header (4 bytes) and a payload (used to carry effective video data, i.e., effective video pixel data).
  • the meaning of each field of the message header of the vertical blanking message may be shown in Table 6.
  • the following arrangement rules can be used for each pixel component in the valid video data: for continuous video pixels, the first device (or the video adapter module of the first device) can prioritize the pixel data at the first position, and then arrange the pixel data after the first position in the order of reception.
  • the video signal is a red, green, blue (RGB) video
  • the arrangement of the valid video data carried by the valid video message can be shown in Figure 13, where Pixel represents a pixel point.
  • the first device can start from the 0th pixel point and arrange the pixels in the row in order until the last pixel point in the row ends.
  • the first device (such as the video adaptation module in the first device) can preferentially arrange the R component, then the G component, and finally the B component.
  • the first device (such as the video adaptation module in the first device) can preferentially arrange the Y component, then the Cb component, and finally the Cr component.
  • the first device (such as the video adaptation module in the first device) can preferentially arrange the Y component of the 0th pixel, the Cb component of the 0th pixel, then the Y component of the 1st pixel, the Cr component of the 1st pixel, then the Y component of the 2nd pixel, the Cb component of the 2nd pixel, ..., the Y component of the nth pixel, the Cb component of the nth pixel, and so on.
  • the first device (such as the video adaptation module in the first device) can arrange the two Y components first, and then arrange the Cb or Cr components according to the odd and even rows, that is, arrange the Cb components in the even rows and the Cr components in the odd rows.
  • the payload portion of the valid video message AVP is used to transmit valid video pixel data or compressed video pixel data, and its length may not exceed 508 bytes.
  • the first device such as the video adapter module in the first device
  • the receiving second device may identify and discard the padding data through the number of horizontal active pixels (HActivePixels) in the VFC.
  • RGB video with a resolution of 1280 ⁇ 720, a frame rate of 60 (P60), and a 24-bit resolution (bit resolution, bpp) as an example.
  • the data arrangement of the video is shown in Figure 17, and the row length of the effective video area is 1280 pixels.
  • R[7], G[7], and B[7] are the highest bits, and R[0], G[0], and B[0] are the lowest bits. Only the pixel components of some pixels are shown in Figure 17.
  • Each row of the video frame has 1280 pixels, each pixel is 24 bits, and one row is 3840 bytes.
  • An effective video packet AVP can carry a maximum of 508 bytes of effective video pixel data.
  • pixels (3840 bytes) require 8 effective video packets AVP to transmit.
  • the first AVP (AVP0) transmits the R component from pixel 0 to pixel 169;
  • AVP1 starts from the G component of pixel 169 to the G component of pixel 338; ...
  • AVP7 starts from the G component of pixel 1185 to pixel 1279.
  • the S flag in the message header of the valid video message AVP0 needs to be set to 1, indicating that this valid video message AVP is the first valid video message AVP in the row
  • the E flag of the valid video message AVP7 needs to be set to 1, indicating that this valid video message AVP is the last valid video message AVP in the row
  • the S flag and the E flag in the message header of the intermediate valid video message AVP should be set to 0, indicating that it is an intermediate valid video message AVP.
  • each row of the video has 1280 pixels, each pixel is 30 bits, and one row is 4800 bytes.
  • a valid video message AVP can carry a maximum of 508 bytes of valid video pixel data.
  • pixels (4800 bytes) require 10 valid video messages AVP to transmit, where the first valid video message AVP (AVP0) transmits the lower 4 bits of the G component of pixel 0 to pixel 135; the valid video message AVP1 starts from the upper 6 bits (G135[9:4]) of the G component of pixel 135 to the lower 8 bits B270[7:0] of the B component of pixel 270; ..., the valid video message AVP9 transmits the upper 4 bits (R1219[9:6]) of the R component of pixel 1219 to pixel 1279, a total of 228 bytes.
  • the S flag in the message header of the valid video message AVP0 needs to be set to 1, indicating that this valid video message AVP is the first valid video message AVP in the row, and the E flag of the valid video message AVP needs to be set to 1, indicating that this valid video message AVP is the last valid video message AVP in the row, among which, the S flag and E flag in the message header of the intermediate valid video message AVP (such as the valid video message AVP1) should be set to 0, indicating that it is an intermediate valid video message AVP.
  • the descriptive information message DIP can transmit audio and video configuration and control information, etc. According to the type of data transmitted, the descriptive information message can be divided into audio control DIP message, compression parameter DIP message, video stream information DIP message, video metadata DIP message, manufacturer extension DIP message, etc. Among them, if only descriptive information messages and audio sampling messages are transmitted between the first device and the second device, the descriptive information message can be sent at any time; if there are other high-priority data (such as vertical blanking message VBP or horizontal blanking message HBP) to be transmitted, the descriptive information message DIP can be sent after the high-priority data or message is sent. As shown in Figure 21, a schematic diagram of a descriptive information message format provided in an embodiment of the present application includes a 4-byte message header and a 40-byte message body (payload). The meaning of each field in the descriptive information message header can be shown in Table 7.
  • the first 4 bytes (HB0 to HB3) of the DIP payload of the descriptive information message are the descriptive information header, followed by 32 bytes of data, and finally 4 bytes of CRC check bytes, totaling 40 bytes.
  • the meaning of the fields of the payload of the descriptive information message can be shown in Table 8.
  • the descriptive information message header byte 0 (HB0) is the descriptive information type, and its meaning can be shown in Table 9.
  • the descriptive information message header byte 1 (HB1) is a reserved field.
  • the descriptive information message header bytes 2 and 3 describe the unique attributes of various types of descriptive information.
  • a descriptive information message DIP can carry a maximum of 32 bytes of descriptive information. Usually, when the length of the descriptive information to be transmitted exceeds 32 bytes, it needs to be split into multiple descriptive information messages DIP. At this time, the S flag in the message header of the first descriptive information message (DIP0) needs to be set to 1, and the E flag of the last descriptive information message (DIP2) needs to be set to 1 to identify the first and last descriptive information messages DIP respectively.
  • the audio control DIP message can be used to transmit audio control information, and its payload structure can be as shown in Table 10.
  • the first device can set the audio mute flag (AudioMuteFlag) in the audio control DIP message to 0b to mute the audio. For example, before switching audio (different audio) or audio format (such as sampling rate) and stopping playing audio, the source side should set AudioMuteFlag to 0b and send the corresponding audio control DIP message to avoid the sink side outputting audio with noise, glitches and other user-perceivable errors.
  • DB1-DB3 of the audio control message payload jointly identify the audio sampling frequency, namely AudioSampleFreq.
  • the sampling rate identified by AudioSampleFreq should not deviate from the actual audio sampling rate by more than 0.5%.
  • the first device (source side) should immediately send the audio control DIP message; when the content of the audio control DIP message remains unchanged, the first device (source side) should send the audio control DIP message once every 1280 audio sampling messages.
  • the source side should fill 0x00; the sink side should have the ability to identify and ignore reserved fields, and should also retain the parsing capability to meet expansion requirements.
  • the video stream information DIP message can be used to transmit the functional information of the current video, and the structure of the payload part can be as shown in Table 11.
  • the video metadata DIP message can be used to transmit the metadata information of the video, and its payload structure can be shown in Table 12.
  • the length of the video metadata information is less than 32 bytes, one video metadata DIP can carry it.
  • the S flag and the E flag in the video metadata DIP message header must be set to 1; and the actual metadata length is saved in the HB3 byte.
  • each message transmits 32 bytes (HB3 is fixed to 32), and the last message is filled with 0 when it is less than 32 bytes, and the actual number of valid bytes in the current message is saved in HB3; in addition, the S flag in the first DIP message is set to 1, and the E flag of the last DIP message is set to 1.
  • the content of the audio information frame DIP message shall comply with the definition in CTA-861-G and be carried by the audio information frame DIP message.
  • the message header and payload of the audio information frame DIP message are shown in Table 13 below.
  • DIP messages may also include manufacturer extended DIP messages, manufacturer specific DIP messages, video information frame DIP messages, source product description DIP messages, moving pictures experts group (MPEG) source information frame DIP messages, dynamic range DIP messages, etc.
  • manufacturer extended DIP messages may also include manufacturer extended DIP messages, manufacturer specific DIP messages, video information frame DIP messages, source product description DIP messages, moving pictures experts group (MPEG) source information frame DIP messages, dynamic range DIP messages, etc.
  • MPEG moving pictures experts group
  • An audio sampling message ASP also called an audio data message, can be used to transmit an audio signal (or audio data).
  • FIG. 24 is a schematic diagram of an audio sampling message ASP format provided in an embodiment of the present application.
  • the audio sampling message ASP can be used to transmit the sampling data of the audio stream (pulse code modulation (linear pulse code modulation, LPCM audio) or compressed sampling data (IEC 61937).
  • the UMI audio data is encapsulated in a manner similar to the audio time slot of the IEC60958 sub-frame.
  • One audio time slot corresponds to 4 bytes, and the specific meaning is shown in Table 14 below.
  • the audio to be transmitted should first comply with the IEC 60958 standard package to form an audio sub-frame of IEC 60958.
  • the first device (such as the video adapter module/audio-visual adapter module in the first device) packages the audio sub-frame into an audio time slot, and finally encapsulates the audio time slot into an audio sampling message ASP for transmission.
  • the audio to be transmitted should comply with the mapping rule in Chapter 6 of IEC 61937, that is, map the audio data to the audio subframe specified by IEC 60958, and the subsequent packetization specification is consistent with the above-mentioned uncompressed audio.
  • the video adapter module or the video and audio adapter module encapsulates the audio data packets that follow the IEC 60958 or IEC 61937 protocol into the audio sampling message ASP to realize the transmission of audio data.
  • the ASP message structure can be shown in Figure 24.
  • One audio time slot corresponds to 4 bytes.
  • the data length in the ASP message is variable.
  • the audio data is 128 bytes, that is, the tunnel message header + ASP audio message header + audio data is 136 bytes.
  • the specific byte meanings are shown in Table 15 below:
  • encryption description packet EDP
  • KDP key distribution packet
  • the structures of the encryption description packet EDP and the key distribution packet KDP can refer to the introduction of the above video packets, and will not be described in detail.
  • the encryption parameters such as the encryption algorithm can be carried in the payload of the encryption description packet EDP
  • the encryption key can be carried in the payload of the encryption description packet KDP.
  • the first device can continuously send multiple copies of the vertical blanking message VBP and the encryption description message EDP, and the key distribution message is sent once per frame, but multiple frames must be sent continuously.
  • the encryption description message EDP is sent three times in a row, and after the second device receives the above three messages, it first performs a correctness check, such as a CRC check. If the check fails, it receives the next message until the correct message is received. For example: after the video receiving adapter in the second device receives the encryption description message EDP, it first performs a CRC check. If the check fails, it receives the next encryption description message EDP until the correct encryption description message EDP is received.
  • a correctness check such as a CRC check.
  • the message structure is shown in FIG. 25.
  • the R flag in the message header of the vertical blanking message VBP, the encryption description message EDP, and the key distribution message KDP can also be changed from reserved (Reserved) to retry (Retry), and the transport layer can realize reliable transmission.
  • Reliable transmission rules of the transport layer When the R flag in the message header is 0, the receiving side (such as the second device) transmits the received message backward regardless of whether it is correct. When the R flag is 1, it is necessary to check whether the message is correct. If it is not correct, a retransmission request is initiated to request the previous level to retransmit the message.
  • the first device can include a Retry flag in the message header of the vertical blanking message VBP, the encryption description message EDP, and the key distribution message KDP.
  • the Retry flag is 1, it is used to instruct the receiving end to check whether the message is correct and initiate a retransmission request when the message is incorrect.
  • the second device initiates retransmission of the message.
  • S605 the second device generates a vertical synchronization signal according to the vertical blanking message, and generates a horizontal synchronization signal according to the horizontal blanking message.
  • the second device generates a video data signal and a display enable signal according to the vertical synchronization signal, the horizontal synchronization signal, the pixel clock signal, and the valid video data carried by the valid video message.
  • the second device outputs a video signal, where the video signal includes a vertical synchronization signal, a horizontal synchronization signal, a pixel clock signal, a display enable signal, and a video data signal.
  • the second device (or the video receiving adapter in the second device) can obtain the vertical blanking message, horizontal blanking message and other messages sent by the first device by demultiplexing the video stream from the first device (such as demultiplexing by a stream demultiplexer, etc.).
  • demultiplexing the video stream to obtain multiple types of messages may refer to identifying (or dividing) multiple types of messages in the video stream according to the identifier of the message.
  • demultiplexing the video stream to obtain vertical blanking messages, horizontal blanking messages and valid video messages may refer to identifying (or dividing) the vertical blanking message, horizontal blanking message and valid video message in the video stream according to the identifier of the vertical blanking message (such as 4b'0001), the identifier of the horizontal blanking message (such as 4b'0010) and the identifier of the valid video message (such as 4b'0101).
  • the second device can pre-agree or negotiate with the first device, or obtain them through the VFC in the vertical blanking message VBP sent by the first device, and the embodiment of the present application does not limit this.
  • the second device when the second device (or the video (or audio and video) receiving adapter in the second device) receives the vertical blanking message VBP from the first device, it determines that the vertical synchronization signal changes from a low level to a high level, changes the vertical synchronization polarity (VSyncPol) from a low level to (or pulls up to) a high level, and continues for the number of VSync lines to generate a vertical synchronization signal.
  • VSyncPol vertical synchronization polarity
  • the second device when the second device receives the horizontal blanking message HBP from the first device, after counting HActivePixels+HfrontPorch, it changes the horizontal synchronization polarity (HSyncPol) from a low level to (pull up to) a high level, and continues for the number of HSync pixels to generate a horizontal synchronization signal.
  • HSyncPol horizontal synchronization polarity
  • Each transmission path of the video data signal (or pixel data signal) transmits data corresponding to one pixel in each pixel clock cycle.
  • the video data signal (or pixel data signal) has no transmission of valid pixels in the vertical blanking area and the horizontal blanking area.
  • the display enable signal is high, it indicates that the current pixel data is transmitting valid video pixel data.
  • the second device can configure the local phase locked loop (PLL) according to the pixel clock signal frequency (PixelClockFreq) to generate a local pixel clock signal.
  • the second device can put the valid video data (i.e., the video pixel data carried by the valid video message AVP) carried by the valid video message AVP into the buffer of the "video signal restoration" module.
  • the “Video Signal Restoration” module pulls up the display enable signal according to the timing parameters after HBackPorch (HBlankPixels–HSyncPixels-HFrontPorch) pixel clocks after the falling edge of the horizontal synchronization signal, or after HSync+HBackPorch (HBlankPixels–HFrontPorch) pixel clocks after the rising and falling edge of the horizontal synchronization signal, and outputs the pixel data cached in the buffer through the video signal line along with the pixel clock.
  • the video stream may also include a descriptive information message DIP, which may be used to transmit descriptive information such as audio and video configuration and control information.
  • the second device may also configure and control the audio and video based on the descriptive information such as the audio and video configuration and control information, and may also output an audio signal based on the audio signal (or audio data) of the transmitted audio sampling message ASP.
  • the second device can also obtain the encryption parameters carried by the encryption description message EDP in the video stream and the encryption key carried by the key distribution message KDP, and decrypt the encrypted valid video message AVP’ and the encrypted audio data ASP’ to obtain the decrypted valid video message AVP and the decrypted audio sampling message ASP.
  • content protection such as encryption
  • the valid video message AVP and the audio data ASP are encrypted (such as ADCP encryption)
  • the second device can also obtain the encryption parameters carried by the encryption description message EDP in the video stream and the encryption key carried by the key distribution message KDP, and decrypt the encrypted valid video message AVP’ and the encrypted audio data ASP’ to obtain the decrypted valid video message AVP and the decrypted audio sampling message ASP.
  • the video stream may not include the encryption description message EDP and the key distribution message KDP.
  • the first device can generate a vertical blanking message VBP, a horizontal blanking message HBP, an effective video message AVP, an audio sampling message ASP, and a descriptive information message DIP.
  • the effective video message AVP and the audio sampling message ASP are encrypted.
  • the second device can decrypt the effective video message AVP and the audio sampling message ASP, and parse the decrypted effective video message AVP and the audio sampling message ASP according to the vertical blanking message VBP, the horizontal blanking message HBP, and the descriptive information message DIP to obtain video data and audio data.
  • the first device can also encrypt one or more messages in the vertical blanking message VBP, the horizontal blanking message HBP, the effective video message AVP and the audio sampling message ASP, and the descriptive information DIP as required.
  • the content protection identifier (such as the CP flag) in the message header can be set to 1 or 0 according to the content protection (such as encryption) requirements.
  • the first device (such as the content protection module in the first device) can process according to the CP flag in the message header. If the CP flag is 1, the message is encrypted. If the CP flag is 0, it is not encrypted.
  • the second device can determine whether the vertical blanking message VBP, the horizontal blanking message HBP, the effective video message AVP and the audio sampling message ASP, and the content protection identifier (such as the CP flag) in the descriptive information DIP is 1, and decrypt the message when it is 1.
  • the audio-visual receiving adapter (or video receiving adapter) and the audio-visual transmitting adapter (or video receiving adapter) are located in different devices.
  • a method for pixel clock synchronization processing using a first-input first-output (FIFO) waterline scheme is provided in the embodiment of the present application.
  • the pixel clock recovery of the video receiving adapter is shown in FIG27.
  • the key information pixel clock frequency (PixelClockFreq) used to recover the video stream from the vertical blanking message VBP.
  • An asynchronous FIFO is used to complete the bandwidth synchronization from the receiving domain to the transmitting domain, reconstruct the video clock based on PixelClockFreq, reconstruct the video timing based on other timing information, and fine-tune the pixel clock frequency through the HBP/VBP message position.
  • the audio and video receiving adapter when the video receiving adapter reconstructs the video pixel data stream, in order to avoid the phenomenon that the video data output is interrupted due to the delayed arrival of some messages caused by the jitter in the video data message transmission process, the audio and video receiving adapter should cache a certain amount of audio and video data.
  • the video receiving adapter receives the first valid video message AVP of each line (the S flag in the message is 1), it cannot be output immediately through the video data signal, but needs to go through an accumulation cycle (such as a video delay of t pixel clocks (tVideoDelay)) before sending the cached valid video data out. Except for the first valid video message AVP, other valid video messages AVP messages can be directly output after receiving them.
  • the accumulation period must ensure that the cached video data can basically resist the impact of transmission jitter on the valid video message AVP, so as to avoid the video receiving adapter frequently inserting padding data in the valid video area due to buffer underflow.
  • the accumulation period should also not be too large to avoid introducing too much delay and increasing the cache cost.
  • the accumulation period is affected by the jitter of the transport layer, logic layer, and routing equipment.
  • vertical synchronization represents the vertical synchronization signal (VSYNC)
  • horizontal synchronization represents the horizontal synchronization signal (HSYNC)
  • display enable represents the display enable signal
  • video data represents the video data signal.
  • 1 can represent the vertical synchronization signal vertical blanking area back shoulder delay of t pixel clocks (tVideoDelay-VSYNC-VBackporch);
  • 2 can represent the horizontal synchronization signal horizontal blanking area back shoulder delay of t pixel clocks ((tVideoDelay-HSYNC-HBackporch);
  • 3 can represent the video data delay of t pixel clocks ((tVideoDelay).
  • the video receiving adapter In order to ensure that the buffer can accumulate a certain amount of video data, the video receiving adapter also needs to delay the accumulation period after receiving the vertical blanking message VBP and the horizontal blanking message HBP, that is, delay t (tVideoDelay) pixel clocks before generating the vertical synchronization signal (VSYNC) and the horizontal blanking message HBP.
  • delay t tVideoDelay
  • the horizontal synchronization signal is generated after the second device on the receiving side receives the vertical blanking message VBP, and generates a vertical synchronization signal after a fixed delay (such as tVideoDelay), and the vertical synchronization signal lasts for VSyncLines*HTotal pixel clocks; after the second device on the receiving side receives the horizontal synchronization signal HBP message, it generates a horizontal synchronization signal after a fixed delay (such as tVideoDelay), and the horizontal synchronization signal lasts for HSyncPixels pixel cycles.
  • the second device on the receiving side receives the valid video message AVP, it puts the video pixel data it carries into the buffer of the "video signal restoration" module and outputs it with the display enable signal.
  • the video receiving adapter after the video receiving adapter receives the corresponding message header (such as the vertical blanking message VBP message header), it can start tVideoDelay timing after identifying the message type according to the message header, without waiting for a complete message to be received.
  • tVideoDelay can only start timing after receiving the complete message, and the number of pixel clock cycles corresponding to this message needs to be subtracted at this time.
  • the vertical blanking message VBP needs to subtract 9 pixel clock cycles
  • the horizontal blanking message HBP needs to subtract 8 pixel clock cycles
  • the valid video message AVP needs to calculate the number of pixel clock cycles to be subtracted based on the actual length of the effective load part.
  • a filling method can be used to solve the problem, that is, temporarily lowering the display enable (DE) signal, waiting for data in the buffer, and then raising the display enable signal and continuing to send video data.
  • DE display enable
  • the valid video message AVP3 arrives late, and the buffer of the video receiving adapter underflows. After the buffered pixel number is sent, the display enable signal needs to be temporarily lowered; when the valid video message AVP3 is received, the video data transmission is immediately resumed.
  • the first device and the second device include hardware structures and/or software modules corresponding to the execution of each function.
  • the present application can be implemented in the form of hardware or a combination of hardware and computer software. Whether a function is executed in the form of hardware or computer software driving hardware depends on the specific application scenario and design constraints of the technical solution.
  • Figures 30 and 31 are schematic diagrams of possible signal transmission devices provided by embodiments of the present application. These signal transmission devices can be used to implement the functions of the first device or the second device in the above method embodiments, and thus can also achieve the beneficial effects of the above method embodiments.
  • the signal transmission device can be the first device or the second device, and can also be a module (such as a chip) applied to the first device or the second device.
  • the signal transmission device 3000 includes a processing unit 3010 and an interface unit 3020, wherein the interface unit 3020 can also be a transceiver unit or an input/output interface.
  • the signal transmission device 3000 can be used to implement the functions of the first device or the second device in the method embodiment shown in Fig. 6 above.
  • the processing unit 3010 is used to obtain a video signal, where the video signal includes a vertical synchronization signal, a horizontal synchronization signal, a pixel clock signal, a display enable signal, and a video data signal; obtain valid video data in the video data signal according to the pixel clock signal and the display enable signal; encapsulate a vertical blanking message according to the vertical synchronization signal, encapsulate a horizontal blanking message according to the horizontal synchronization signal, and encapsulate a valid video message according to the valid video data;
  • the interface unit 3020 is used to combine the vertical blanking message, the horizontal blanking message and the effective video message into a video stream and send it.
  • the processing unit 3010 is also used to obtain descriptive information located in the blanking area of the video data signal according to the pixel clock signal and the display enable signal; and to encapsulate the descriptive information message according to the descriptive information; when the interface unit 3020 combines the vertical blanking message, the horizontal blanking message and the valid video message into a video stream and sends it, it is specifically used to combine the vertical blanking message, the horizontal blanking message, the valid video message and the descriptive information message into a video stream and send it.
  • the processing unit 3010 when the processing unit 3010 encapsulates the descriptive information message according to the descriptive information, it is specifically used to encapsulate the descriptive information message according to the descriptive information and the audio signal when the video signal also includes an associated audio signal.
  • the processing unit 3010 is also used to encapsulate the audio sampling message according to the audio signal when the video signal also includes an associated audio signal; when the interface unit 3020 combines the vertical blanking message, the horizontal blanking message, the valid video message and the descriptive information message into a video stream and sends it, it is specifically used to combine the vertical blanking message, the horizontal blanking message, the valid video message, the descriptive information message and the audio sampling message into a video stream and send it.
  • the processing unit 3010 is also used to encrypt the valid video message according to the encryption parameters and the encryption key to obtain the encrypted valid video message; encapsulate the encrypted description message according to the encryption parameters, and encapsulate the key distribution message according to the encryption key; when the interface unit 3020 combines the vertical blanking message, the horizontal blanking message, the valid video message and the descriptive information message into a video stream and sends it, it is specifically used to combine the vertical blanking message, the horizontal blanking message, the encrypted valid video message, the descriptive information message, the encrypted description message and the key distribution message into a video stream and send it.
  • the processing unit 3010 is further used to encrypt the valid video message and the audio sampling message according to the encryption parameters and the encryption key to obtain the encrypted valid video message and the encrypted audio sampling message; encapsulate the encryption description message according to the encryption parameters, and encapsulate the key distribution message according to the encryption key; when the interface unit 3020 combines the vertical blanking message, the horizontal blanking message, the valid video message, the descriptive information message and the audio sampling message into a video stream and sends it, it is specifically used to combine the vertical blanking message, the horizontal blanking message, the encrypted valid video message, the descriptive information message, the encrypted audio sampling message, the encrypted description message and the key distribution message into a video stream and send it.
  • vertical blanking messages, encryption description messages, and key distribution messages are repeatedly sent multiple times in a video stream.
  • the message headers of the vertical blanking message, the encryption description message, and the key distribution message include a retry flag.
  • the Retry flag is 1, it is used to instruct the receiving end to check whether the message is correct and initiate a retransmission request if the message is incorrect.
  • the processing unit 3010 when the processing unit 3010 encapsulates the vertical blanking message according to the vertical synchronization signal, it is specifically used to encapsulate the vertical blanking message according to the vertical synchronization signal and the video frame control information, and the video frame control information includes the pixel clock parameters of the pixel clock signal.
  • the processing unit 3010 when the processing unit 3010 encapsulates the horizontal blanking message according to the horizontal synchronization signal, it is specifically used to encapsulate the horizontal blanking message according to the horizontal synchronization signal and the pixel clock count value of the pixel clock signal.
  • the interface unit 3020 is used to receive a video stream, where the video stream includes a vertical blanking message, a horizontal blanking message and a valid video message;
  • the processing unit 3010 is used to generate a vertical synchronization signal according to a vertical blanking message, and to generate a horizontal synchronization signal according to a horizontal blanking message; to generate a video data signal and a display enable signal according to the vertical synchronization signal, the horizontal synchronization signal, the pixel clock signal and the valid video data carried by the valid video message; and to output a video signal, wherein the video signal includes a vertical synchronization signal, a horizontal synchronization signal, a pixel clock signal, a display enable signal and a video data signal.
  • the video stream also includes a descriptive information message.
  • the processing unit 3010 When the processing unit 3010 generates a video data signal and a display enable signal based on the vertical synchronization signal, the horizontal synchronization signal, the pixel clock signal, and the valid video data carried by the valid video message, it is specifically used to generate a video data signal and a display enable signal based on the vertical synchronization signal, the horizontal synchronization signal, the pixel clock signal, the valid video data carried by the valid video message, and the descriptive information carried by the descriptive information message.
  • the processing unit 3010 is further used to generate an audio signal based on the descriptive information message, wherein the output video signal also includes an audio signal.
  • the video stream also includes an audio sampling message
  • the processing unit 3010 is further configured to generate an audio signal according to the audio sampling message; wherein the output video signal also includes an audio signal.
  • the video stream also includes an encrypted description message and a key distribution message
  • the valid video message is an encrypted valid video message.
  • the processing unit 3010 is also used to obtain encryption parameters carried by the encrypted description message and to obtain an encryption key carried by the key distribution message; the encrypted valid video message is decrypted according to the encryption parameters and the encryption key to obtain a decrypted valid video message.
  • the video stream also includes an encrypted description message and a key distribution message
  • the valid video message and the audio sampling message are encrypted valid video messages and encrypted audio sampling messages.
  • the processing unit 3010 is further used to obtain encryption parameters carried by the encrypted description message and to obtain encryption keys carried by the key distribution message; and to decrypt the encrypted valid video message and the encrypted audio sampling message according to the encryption parameters and the encryption key to obtain decrypted valid video messages and decrypted audio sampling messages.
  • the video stream includes multiple vertical blanking messages, encrypted description messages and key distribution messages that are sent repeatedly and continuously.
  • the interface unit 3020 is also used to select the vertical blanking messages, encrypted description messages and key distribution messages that pass the correctness check from the multiple vertical blanking messages, encrypted description messages and key distribution messages that are sent repeatedly and continuously.
  • the message headers of the vertical blanking message, the encryption description message and the key distribution message include a retry flag.
  • the Retry flag is 1, it is used to instruct the receiving end to check whether the message is correct and to initiate a retransmission request when the message is incorrect.
  • the interface unit 3020 is also used to initiate retransmission of the message when any of the vertical blanking message, the encryption description message and the key distribution message is incorrect.
  • the processing unit 3010 is also used to obtain video frame control information carried by the vertical blanking message, where the video frame control information includes pixel clock parameters; and generate a pixel clock signal based on the video frame control information.
  • the horizontal blanking message also carries a pixel clock count value of the pixel clock signal, and the processing unit 3010 is further used to correct the pixel clock signal according to the pixel clock count value.
  • the present application also provides a signal transmission device 3100, including a processor 3110 and an interface circuit 3120.
  • the processor 3110 and the interface circuit 3120 are coupled to each other.
  • the interface circuit 3120 can be a transceiver, an input-output interface, an input interface, an output interface, a communication interface, etc.
  • the signal transmission device 3100 may also include a memory 3130 for storing instructions executed by the processor 3110 or storing input data required by the processor 3110 to run the instructions or storing data generated after the processor 3110 runs the instructions.
  • the memory 3130 may also be integrated with the processor 3110.
  • the processor 3110 may be used to implement the function of the processing unit 3010
  • the interface circuit 3120 may be used to implement the function of the interface unit 3020 .
  • the processor in the embodiments of the present application may be a central processing unit (CPU), or other general-purpose processors, digital signal processors (DSP), application specific integrated circuits (ASIC), logic circuits, field programmable gate arrays (FPGA) or other programmable logic devices, transistor logic devices, hardware components or any combination thereof.
  • the general-purpose processor may be a microprocessor or any conventional processor.
  • the method steps in the embodiments of the present application can be implemented by hardware, or by a processor executing software instructions.
  • the software instructions can be composed of corresponding software modules, and the software modules can be stored in random access memory, flash memory, read-only memory, programmable read-only memory, erasable programmable read-only memory, electrically erasable programmable read-only memory, register, hard disk, mobile hard disk, CD-ROM or any other form of storage medium well known in the art.
  • An exemplary storage medium is coupled to the processor so that the processor can read information from the storage medium and can write information to the storage medium.
  • the storage medium can also be a component of the processor.
  • the processor and the storage medium can be located in an ASIC.
  • the ASIC can be located in a network device or a terminal device.
  • the processor and the storage medium can also be present in a network device or a terminal device as discrete components.
  • the computer program product includes one or more computer programs or instructions.
  • the computer may be a general-purpose computer, a special-purpose computer, a computer network, a network device, a user device or other programmable device.
  • the computer program or instruction may be stored in a computer-readable storage medium, or transmitted from one computer-readable storage medium to another computer-readable storage medium, for example, the computer program or instruction may be transmitted from one network device, terminal, computer, server or data center to another network device, terminal, computer, server or data center by wired or wireless means.
  • the computer-readable storage medium may be any available medium that a computer can access or a data storage device such as a server or data center that integrates one or more available media.
  • the available medium may be a magnetic medium, for example, a floppy disk, a hard disk, a tape; it may also be an optical medium, for example, a digital video disc; it may also be a semiconductor medium, for example, a solid-state hard disk.
  • the computer-readable storage medium may be a volatile or nonvolatile storage medium, or may include both volatile and nonvolatile types of storage media.

Landscapes

  • Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Multimedia (AREA)
  • Computer Security & Cryptography (AREA)
  • Physics & Mathematics (AREA)
  • Computer Hardware Design (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

本申请涉及音视频处理技术领域,公开了一种信号传输方法及装置,用以降低视频信号的传输时延,降低视频信号发送设备和接收设备的成本和功耗。该方法包括:获取视频信号,视频信号包括垂直同步信号、水平同步信号、像素时钟信号、显示使能信号和视频数据信号;根据像素时钟信号、显示使能信号获取视频数据信号中有效视频数据;根据垂直同步信号封装垂直消隐报文、根据水平同步信号封装水平消隐报文、根据有效视频数据封装有效视频报文;将垂直消隐报文、水平消隐报文和有效视频报文复合为视频流并发送。

Description

一种信号传输方法及装置
相关申请的交叉引用
本申请要求在2022年07月21日提交中国专利局、申请号为202210859501.X、申请名称为“信号传输方法”的中国专利申请的优先权,其全部内容通过引用结合在本申请中;本申请要求在2023年02月21日提交中国专利局、申请号为202310189854.8、申请名称为“一种信号传输方法及装置”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。
技术领域
本申请实施例涉及音视频处理技术领域,尤其涉及一种信号传输方法及装置。
背景技术
目前计算机、便携机、手机、平板电脑等视频源设备与显示器等视频宿设备之间进行视频信号传输时,需要先通过视频源设备中的显示端口(displayport,DP)控制器将视频信号转换为DP流,然后通过通用串行总线(universal serial bus,USB)4/雷电(thunderbolt、TBT)路由器(router)中的DP输入(IN)适配器将DP流转换成USB4/TBT报文,再通过USB4/TBT网络传输到视频宿设备,由视频宿设备中USB4/TBT router中的DP输出(OUT)适配器根据报文恢复成DP流,再经过DP控制器还原成视频信号。
然而,上述信号传输时,视频信号需要经过DP控制器处理后再送往USB/TBT适配器进行处理,增大了视频信号的传输时延,并且视频源设备和视频宿设备均需集成DP控制器等硬件设备,也会带来额外的成本和功耗。
发明内容
本申请实施例提供一种信号传输方法及装置,用以降低视频信号的传输时延,降低视频信号发送设备(如视频源设备)和接收设备(视频宿设备)的成本和功耗。
第一方面,本申请实施例提供一种信号传输方法,可以应用于第一设备、或第一设备的部件(例如处理器、芯片、或芯片***等),或者与第一设备匹配使用的装置等。其中,第一设备可以是机顶盒,游戏机等视频源设备,还可以是进行视频信号等发送的终端设备、计算机,便携机等设备,以下以应用于第一设备为例。该方法包括:获取视频信号,视频信号包括垂直同步信号、水平同步信号、像素时钟信号、显示使能信号和视频数据信号;根据像素时钟信号、显示使能信号获取视频数据信号中有效视频数据;根据垂直同步信号封装垂直消隐报文、根据水平同步信号封装水平消隐报文、根据有效视频数据封装有效视频报文;将垂直消隐报文、水平消隐报文和有效视频报文复合为视频流并发送。可选地,在本申请实施例中,可以将垂直消隐报文、水平消隐报文和有效视频报文通过流复用器(如视频流复用器、音视频流复用器)等复合为视频流。
需要理解的是,在本申请实施例中,将多类报文复合为视频流,可以是指将多类报文通过一个视频流传输,在多类报文中均可以携带该视频流的标识。例如:上述将垂直消隐报文、水平消隐报文和有效视频报文复合为视频流,可以是指将垂直消隐报文、水平消隐报文和有效视频报文通过一个视频流传输,在一个视频流中包括垂直消隐报文、水平消隐报文和有效视频报文,垂直消隐报文、水平消隐报文和有效视频报文中均可以携带该视频流的标识。
采用上述方法,可以直接将视频信号打包成报文进行传输,视频信号无需经过DP控制器处理后再送USB/TBT适配器进行打包处理,能够降低视频信号的传输时延;另外,无需视频信号发送设备(如第一设备)和接收设备集成额外的DP控制器,以及在USB/TBT等适配器集成DP的电气层和逻辑层,能够降低视频信号发送设备和接收设备的成本和功耗。
在一种可能的设计中,发送视频流包括:通过统一多媒体互联接口(unified multimedia interconnection,UMI)的传输层和物理层发送视频流。
在本申请实施例中,可以通过发送设备中的视音频发送适配器(或视频发送适配器)将视频信号打包为UMI报文,并将UMI报文复合得到的视频流通过UMI网络的传输层和物理层(包括电气层和逻辑层)发送给相应的接收设备中的视音频接收适配器(或视频接收适配器),由视音频接收适配器再将UMI报文转换为视频信号,来降低视频信号发送设备和接收设备的成本和功耗。
在一种可能的设计中,该方法还包括:根据像素时钟信号、显示使能信号获取视频数据信号中位于消隐区的描述性信息;根据描述性信息封装描述性信息报文;将垂直消隐报文、水平消隐报文和有效视频报文复合为视频流并发送,包括:将垂直消隐报文、水平消隐报文、有效视频报文和描述性信息报文复合为视频流并发送。
可选地,根据描述性信息封装描述性信息报文,包括:当视频信号还包括随路音频信号时,根据描述性信息和音频信号封装描述性信息报文。
上述设计中,可以将随路数据,如音频和视频的配置及控制信息等,通过描述性信息报文传输,有利于保证视频信号传输的完整性与可靠性。
在一种可能的设计中,该方法还包括:当视频信号还包括随路音频信号时,根据音频信号封装音频采样报文;将垂直消隐报文、水平消隐报文、有效视频报文和描述性信息报文复合为视频流并发送,包括:将垂直消隐报文、水平消隐报文、有效视频报文、描述性信息报文和音频采样报文复合为视频流并发送。
上述设计中,将音频信号和描述性信息分开传输,有利于在接收端存在纯音频处理设备,如功放(音响)等时,接收端对音频信号的快速处理。
在一种可能的设计中,该方法还包括:根据加密参数和加密密钥对有效视频报文和音频采样报文进行加密,得到加密后的有效视频报文和加密后的音频采样报文;根据加密参数封装加密描述报文、根据加密密钥封装密钥分发报文;将垂直消隐报文、水平消隐报文、有效视频报文、描述性信息报文和音频采样报文复合为视频流并发送,包括:将垂直消隐报文、水平消隐报文、加密后的有效视频报文、描述性信息报文、加密后的音频采样报文、加密描述报文和密钥分发报文复合为视频流并发送。
上述设计中,能够实现对音视频数据的保密传输,满足用户在不同场景下的传输需求。
在一种可能的设计中,垂直消隐报文、加密描述报文和密钥分发报文在视频流中连续重复发送多份。
因视频报文中的垂直消隐报文用于传输垂直同步信号,指示一帧视频图像的开始,加密描述报文和密钥分发报文用于对有效视频数据的解密,如果这些报文传输出错,则可能出现无法正确输出视频信号的问题。上述设计中,连续重复发送多份报文,能够提高视频信号传输的可靠性。
在一种可能的设计中,垂直消隐报文、加密描述报文和密钥分发报文的报文头中包含重试(Retry)标志位,Retry标志位为1时,用于指示接收端检查报文是否正确,并在报文不正确时发起重传请求。
上述设计中,可以指示接收端检查垂直消隐报文、加密描述报文和密钥分发报文是否正确,提高信号传输的可靠性。
在一种可能的设计中,根据垂直同步信号封装垂直消隐报文,包括:根据垂直同步信号和视频帧控制信息封装垂直消隐报文,视频帧控制信息包括像素时钟信号的像素时钟参数。
上述设计中,在垂直消隐报文中携带视频帧控制信息,有利于接收端对视频信号的恢复。
在一种可能的设计中,根据垂直同步信号封装垂直消隐报文,包括:根据垂直同步信号和显示使能信号(DE)信号封装垂直消隐报文。
在一种可能的设计中,还可以根据水平消隐区(HBlank)结束信号封装水平消隐报文。可选地,水平消隐报文中还可以包括:垂直消隐区标志(VBlank_tage)。
第二方面,本申请实施例提供一种信号传输方法,可以应用于第二设备、或第二设备的部件(例如处理器、芯片、或芯片***等),或者与第二设备匹配使用的装置等。其中,第二设备可以是显示器,电视机等视频宿设备,还可以是进行视频信号等接收的终端设备、计算机,便携机等设备,以下以应用于第二设备为例。该方法包括:接收视频流,视频流包括垂直消隐报文、水平消隐报文和有效视频报文;根据垂直消隐报文生成垂直同步信号、根据水平消隐报文生成水平同步信号;根据垂直同步信号、水平同步信号、像素时钟信号和有效视频报文携带的有效视频数据,生成视频数据信号和显示使能信号;输出视频信号,视频信号包括垂直同步信号、水平同步信号、像素时钟信号、显示使能信号和视频数据信号。可选地,在本申请实施例中,可以将视频流通过流解复用器(如视频流解复用器、音视频流解复用器)等解复用,得到垂直消隐报文、水平消隐报文和有效视频报文。
需要理解的是,在本申请实施例中,将视频流解复合,得到将多类报文,可以是指根据报文的标识, 对视频流中的多类报文进行识别(或划分)。以视频流中包括垂直消隐报文、水平消隐报文和有效视频报文为例,将视频流解复用,得到垂直消隐报文、水平消隐报文和有效视频报文,可以是指根据垂直消隐报文的标识(如4b'0001)、水平消隐报文的标识(如4b'0010)和有效视频报文的标识(如4b'0101),在视频流中识别(或划分)出垂直消隐报文、水平消隐报文和有效视频报文。
在一种可能的设计中,接收视频流,包括:通过UMI的传输层和物理层接收视频流。
在一种可能的设计中,视频流还包括描述性信息报文,根据垂直同步信号、水平同步信号、像素时钟信号和有效视频报文携带的有效视频数据,生成视频数据信号和显示使能信号,包括:根据垂直同步信号、水平同步信号、像素时钟信号、有效视频报文携带的有效视频数据和描述性信息报文携带的描述性信息,生成视频数据信号和显示使能信号。
在一种可能的设计中,该方法还包括:根据描述性信息报文,生成音频信号,其中输出的视频信号还包括音频信号。
在一种可能的设计中,视频流还包括音频采样报文,该方法还包括:根据音频采样报文生成音频信号;其中,输出的视频信号还包括音频信号。
在一种可能的设计中,视频流还包括加密描述报文和密钥分发报文,有效视频报文为加密后的有效视频报文,该方法还包括:获取加密描述报文携带的加密参数、获取密钥分发报文携带的加密密钥;根据加密参数和加密密钥对加密后的有效视频报文进行解密,得到解密后的有效视频报文。
在一种可能的设计中,视频流还包括加密描述报文和密钥分发报文,有效视频报文和音频采样报文为加密后的有效视频报文和加密后的音频采样报文,该方法还包括:获取加密描述报文携带的加密参数、获取密钥分发报文携带的加密密钥;根据加密参数和加密密钥对加密后的有效视频报文和加密后的音频采样报文进行解密,得到解密后的有效视频报文和解密后的音频采样报文。
在一种可能的设计中,视频流中包括连续重复发送的多份垂直消隐报文、加密描述报文和密钥分发报文,该方法还包括:从连续重复发送的多份垂直消隐报文、加密描述报文和密钥分发报文中选取通过正确性校验的垂直消隐报文、加密描述报文和密钥分发报文。
在一种可能的设计中,垂直消隐报文、加密描述报文和密钥分发报文的报文头中包含重试(Retry)标志位,Retry标志位为1时,用于指示接收端检查报文是否正确,并在报文不正确时发起重传请求,该方法还包括:当垂直消隐报文、加密描述报文和密钥分发报文中任一报文不正确时,对报文发起重传。
在一种可能的设计中,该方法还包括:获取垂直消隐报文携带的视频帧控制信息,视频帧控制信息包括像素时钟参数;根据视频帧控制信息,生成像素时钟信号。
第三方面,本申请实施例提供一种信号传输装置,该装置具有实现上述第一方面中方法的功能,功能可以通过硬件实现,也可以通过硬件执行相应的软件实现。硬件或软件包括一个或多个与上述功能相对应的模块,比如包括接口单元和处理单元。
在一个可能的设计中,该装置可以是芯片或者集成电路。
在一个可能的设计中,该装置包括存储器和处理器,存储器用于存储处理器执行的指令,当指令被处理器执行时,装置可以执行上述第一方面的方法。
在一个可能的设计中,该装置可以为第一设备整机。
第四方面,本申请实施例提供一种信号传输装置,该装置具有实现上述第二方面中方法的功能,功能可以通过硬件实现,也可以通过硬件执行相应的软件实现。硬件或软件包括一个或多个与上述功能相对应的模块,比如包括接口单元和处理单元。
在一个可能的设计中,该装置可以是芯片或者集成电路。
在一个可能的设计中,该装置包括存储器和处理器,存储器用于存储处理器执行的指令,当指令被处理器执行时,装置可以执行上述第二方面的方法。
在一个可能的设计中,该装置可以为第二设备整机。
第五方面,本申请实施例提供一种信号传输装置,该信号传输装置包括接口电路和处理器,处理器和接口电路之间相互耦合。处理器通过逻辑电路或执行指令用于实现上述第一方面的方法。接口电路用于接收来自该信号传输装置之外的其它信号传输装置的信号并传输至处理器或将来自处理器的信号发送给该信号传输装置之外的其它信号传输装置。可以理解的是,接口电路可以为收发器或收发机或收发信机或输入输出接口。
可选的,信号传输装置还可以包括存储器,用于存储处理器执行的指令或存储处理器运行指令所需要的输入数据或存储处理器运行指令后产生的数据。存储器可以是物理上独立的单元,也可以与处理器耦合,或者处理器包括该存储器。
第六方面,本申请实施例提供一种信号传输装置,该信号传输装置包括接口电路和处理器,处理器和接口电路之间相互耦合。处理器通过逻辑电路或执行指令用于实现上述第二方面的方法。接口电路用于接收来自该信号传输装置之外的其它信号传输装置的信号并传输至处理器或将来自处理器的信号发送给该信号传输装置之外的其它信号传输装置。可以理解的是,接口电路可以为收发器或收发机或收发信机或输入输出接口。
可选的,信号传输装置还可以包括存储器,用于存储处理器执行的指令或存储处理器运行指令所需要的输入数据或存储处理器运行指令后产生的数据。存储器可以是物理上独立的单元,也可以与处理器耦合,或者处理器包括该存储器。
第七方面,本申请实施例提供一种通信***,该通信***包括第一设备和第二设备,第一设备可以实现上述第一方面的方法,第二设备可以实现上述第二方面的方法。
第八方面,本申请实施例提供一种计算机可读存储介质,在存储介质中存储有计算机程序或指令,当计算机程序或指令被处理器执行时,可以实现上述第一方面或第二方面的方法。
第九方面,本申请实施例还提供一种计算机程序产品,包括计算机程序或指令,当计算机程序或指令被处理器执行时,可以实现上述第一方面或第二方面的方法。
第十方面,本申请实施例还提供一种芯片***,该芯片***包括:处理器和存储器,处理器与存储器耦合,存储器用于存储程序或指令,当程序或指令被处理器执行时,可以实现上述第一方面或第二方面的方法。
上述第二方面至第十方面所能达到的技术效果请参照上述第一方面所能达到的技术效果,这里不再重复赘述。
附图说明
图1为本申请实施例提供的图像显示示意图;
图2为本申请实施例提供的视频信号生成示意图;
图3为本申请实施例提供视频传输示意图;
图4为本申请实施例提供的信号传输方法示意图之一;
图5为本申请实施例提供的双向多时音频流传输示意图;
图6为本申请实施例提供的信号传输方法示意图之二;
图7为本申请实施例提供的视频信号时序参数示意图;
图8为本申请实施例提供的视频发送适配器逻辑图之一;
图9为本申请实施例提供的视频报文结构示意图之一;
图10为本申请实施例提供的垂直消隐报文格式示意图;
图11为本申请实施例提供的水平消隐报文格式示意图;
图12为本申请实施例提供的有效视频报文格式示意图;
图13为本申请实施例提供的有效视频数据的排布示意图;
图14为本申请实施例提供的RGB格式视频像素分量排布示意图;
图15为本申请实施例提供的YCbCr444格式视频像素分量排布示意图;
图16为本申请实施例提供的YCbCr422格式视频像素分量排布示意图;
图17为本申请实施例提供的视频的数据排布示意图之一;
图18为本申请实施例提供的有效视频报文示意图之一;
图19为本申请实施例提供的视频的数据排布示意图之二;
图20为本申请实施例提供的有效视频报文示意图之二;
图21为本申请实施例提供的描述性信息报文格式示意图;
图22为本申请实施例提供的描述性信息报文封装示意图;
图23为本申请实施例提供的音频采样报文封装示意图;
图24为本申请实施例提供的音频采样报文示意图;
图25为本申请实施例提供的视频报文结构示意图之二;
图26为本申请实施例提供的视频接收适配器逻辑图之一;
图27为本申请实施例提供的时钟恢复示意图;
图28为本申请实施例提供的视频信号传输与恢复示意图;
图29为本申请实施例提供的视频数据填充示意图;
图30为本申请实施例提供的通信装置结构示意图之一;
图31为本申请实施例提供的通信装置结构示意图之二。
具体实施方式
在介绍本申请实施例之前,首先对本申请中的部分用语进行解释说明,以便于本领域技术人员理解。
1)、视频信号时序,视频信号时序源自阴极射线管(cathode ray tube,CRT)显示器/电视机,这种显示器使用电子枪发射电子束,电子束击打屏上的磷光粉,使磷光粉发光,光子束的强度控制像素的明暗,电子束采用光栅扫描(raster-scan)技术实现图像显示。参照图1所示的图像显示示意图,在开始显示一帧图像之前,电子束(electron beam)需要从右下角回到左上角,在这个过程中电子束需要消隐(不显示图像),这段时间称为垂直消隐(vertical retrace)。在显示完一行(line)数据后,电子束需要从当前行的最右侧回到下一行的最左侧,同样在这个过程中也需要消隐,这段时间称为水平消隐(honzontal retrace)。
2)、视频信号,参照图2所示的视频信号生成示意图,视频信号通常包括垂直同步信号(VSYNC)、水平同步信号(HSYNC)、显示使能信号(DE)、像素时钟信号、视频数据信号(或像素数据信号)等几个子信号。垂直同步信号:高电平时表示一帧视频图像的开始;水平同步信号:高电平时表示一行视频图像的开始;显示使能信号:高电平时表示当前像素数据中传输的是有效视频像素数据;视频数据信号(或像素数据信号):像素数据传输通路,通常有24/30/36/48线(视频信号线)组成,任一传输通路(即视频信号线)在每个像素时钟周期传输一个像素对应的数据。
参照图3所示的视频传输示意图,目前在视频源设备与视频宿设备之间进行视频信号传输时,需要先通过视频源设备中的DP控制器将视频信号转换为DP流,然后通过USB4/TBT路由器(router)中的DP输入(IN)适配器将DP流转换成USB4/TBT报文,再通过USB4/TBT网络传输到视频宿设备,由视频宿设备中USB4/TBT router中的DP输出(OUT)适配器根据报文恢复成DP流,再经过DP控制器还原成视频信号。然而,上述视频信号传输时,视频信号需要经过DP控制器处理后再送USB/TBT适配器进行处理,增大了视频信号的传输时延,并且视频源设备和视频宿设备均需集成额外的DP控制器,适配器需要集成DP的电气层和逻辑层会带来额外的成本和功耗。
有鉴于此,本申请提供一种信号传输方法及装置,旨在通过直接将视频信号打包成报文进行传输,以期降低视频信号的传输时延,降低视频源设备和视频宿设备的成本和功耗。
如图4所示,为本申请实施例提供的一种信号传输方法示意图,视音频信号经过视音频发送适配器(或视频发送适配器)可以转换成统一多媒体互联接口(unified multimedia interconnection,UMI)报文,UMI报文经过UMI网络传输给相应的视音频接收适配器(或视频接收适配器),由视音频接收适配器再将UMI报文转换为视音频信号。其中,UMI标准支持双向多视、音频业务流传输,适配层中视音频数据的排布与传输层无关。基于UMI的双向多时音频流传输可以如图5所示,适配层可以包括一个或多个视音频发送适配器和一个或多个视音频接收适配器,视音频信号经过适配层的视音频发送适配器打包成UMI报文后,由UMI报文构成的视音频流通过传输层和物理层(包括电气层和逻辑层)传输被接收端的视音频接收适配器接收,由视音频接收适配器再将UMI报文转换为视音频信号。需要理解的是,在本申请实施例中,视音频信号不局限于通过UMI传输,也可以支持在PCIe等通用高速数据传输接口中使用。下面将结合附图,对本申请实施例进行详细描述。
另外,需要理解的是,本申请实施例提及“第一”、“第二”等序数词是用于对多个对象进行区分,不用于限定多个对象的大小、内容、顺序、时序、优先级或者重要程度等。例如,第一设备和第二设备,并不是表示这两个设备对应的优先级或者重要程度等的不同。
本申请实施例中,对于名词的数目,除非特别说明,表示“单数名词或复数名词”,即"一个或多个”。“至少一个”是指一个或者多个,“多个”是指两个或两个以上。“和/或”,描述关联对象的关联关系,表示可以存在三种关系,例如,A和/或B,可以表示:单独存在A,同时存在A和B,单独存在B的情况,其中A,B可以是单数或者复数。字符“/”一般表示前后关联对象是一种“或”的关系。例如,A/B,表示: A或B。“以下至少一项(个)”或其类似表达,是指的这些项中的任意组合,包括单项(个)或复数项(个)的任意组合。例如,a,b,或c中的至少一项(个),表示:a,b,c,a和b,a和c,b和c,或a和b和c,其中a,b,c可以是单个,也可以是多个。
图6为本申请实施例提供的一种信号传输方法示意图。图6中以第一设备和第二设备作为执行主体为例示意该方法,其中第一设备可以是机顶盒,游戏机等视频源设备、第二设备可以是显示器,电视机等视频宿设备,第一设备还可以是进行视频信号等发送的终端设备、计算机,便携机等等、第二设备还可以是进行视频信号等接收的终端设备、计算机,便携机等等,本申请并不限制该方法的执行主体。例如,图6中的第一设备也可以是第一设备的部件(例如处理器、芯片、或芯片***等),或者和第一设备匹配使用的装置;图6中的第二设备也可以是第二设备的部件(例如处理器、芯片、或芯片***等),或者和第二设备匹配使用的装置。该方法包括:
S601:第一设备获取视频信号,视频信号包括垂直同步信号、水平同步信号、像素时钟信号、显示使能信号和视频数据信号。
S602:第一设备根据像素时钟信号、显示使能信号获取视频数据信号中有效视频数据。
S603:第一设备根据垂直同步信号和水平消隐区(HBlank)结束信号封装垂直消隐报文、根据HBlank结束信号封装水平消隐报文、根据有效视频数据封装有效视频报文。
在一些实现中,水平消隐区结束信号也可以称为水平消隐结束信号。
S604:第一设备将垂直消隐报文、水平消隐报文和有效视频报文复合为视频流并发送,相应地,第二设备接收视频流。
在本申请实施例中,视频信号可以由第一设备产生,例如由第一设备中的视频信号发生器产生,也可以由第一设备从其它设备处获取,本申请对比不做限定。
参照图7所示,为本申请实施例提供的一种视频信号时序参数示例。视频信号包括视频子信号:水平同步信号(HSYNC)、垂直同步信号(VSYNC)、显示使能信号(DE)、像素时钟信号和视频数据信号视频子信号(图7中矩形区域)。视频信号时序参数是指视频子信号之间在时间上的关系,具体各时序参数如表1所示。
表1

视频数据信号(或像素数据信号)的每个传输通路在每个像素时钟周期传输一个像素对应的数据,结合图7所示的视频信号时序参数示例可知,视频数据信号(或像素数据信号)在垂直消隐区和水平消隐区没有有效像素点的传输,显示使能信号高电平时表示当前像素数据中传输的是有效视频像素数据。因此,在本申请实施例中,可以根据像素时钟信号、显示使能信号获取视频数据信号中的有效视频数据(即有效像素数据)。以降低视频信号的传输带宽。
参照图8所示的视频(或视音频)发送适配器逻辑图,第一设备(或第一设备中的视频发送适配器)在获取到视频信号后,可以根据像素时钟信号、显示使能信号获取(或捕获)视频数据信号中有效视频数据,并根据有效视频数据封装有效视频报文(active video packet,AVP),有效视频报文AVP中可以携带有效视频数据。另外,第一设备还可以根据垂直同步信号封装垂直消隐报文(vertical blanking packet,VBP)、根据水平同步信号封装水平消隐报文(horizontal blanking packet,HBP)。例如:在垂直同步信号由低电平变为高电平时,生成垂直消隐报文VBP,指示垂直同步信号的开始,也即指示一帧视频图像的开始;在DE信号由低电平变为高电平时,生成水平消隐报文HBP,指示一行视频图像的开始。
在一些实施中,第一设备还可以根据像素时钟信号、显示使能信号获取(或捕获)视频数据信号中位于消隐区(如垂直消隐区和水平消隐区)的描述性信息;根据描述性信息封装描述性信息报文(descriptive information packet,DIP),描述性信息报文DIP中可以携带有描述性信息。第一设备还可以将描述性信息报文DIP复合(如通过流复用器等复合)在视频流中发送。
另外,为了进一步提高视频数据传输的安全性,第一设备还可以启动内容保护(如加密),将有效视频报文AVP进行加密(如进行高级数据内容保护***技术(advanced digital content protection,ADCP)加密等),并可以根据加密时采用的加密密钥封装密钥分发报文(key distribution packet,KDP)、加密时采用的加密算法/加密方式等加密参数封装加密描述报文(encryption description packet,EDP),将加密后的有效视频报文AVP’、描述性信息报文、携带加密密钥信息的密钥分发报文KDP、携带加密参数加密描述报文EDP、以及垂直消隐报文VBP、水平消隐报文HBP等复合(如通过流复用器等复合)为视频流并发送。例如:视频流可以通过UMI网络发送。如通过UMI的传输层和物理层(包括电气层和逻辑层)发送给接收的第二设备(或第二设备中的视频接收适配器)。
需要理解的是,如果第一设备和第二设备约定或预先协商有加密参数或加密密钥,视频流中还可以不包括加密描述报文EDP和密钥分发报文KDP。
在本申请实施例中,将多类报文复合(如通过流复用器等复合)为视频流,可以是指将多类报文通过一个视频流传输,在多类报文中均可以携带该视频流的标识。例如:上述将垂直消隐报文、水平消隐报文和有效视频报文复合为视频流,可以是指将垂直消隐报文、水平消隐报文和有效视频报文通过一个视频流传输,在一个视频流中包括垂直消隐报文、水平消隐报文和有效视频报文,垂直消隐报文、水平消隐报文和有效视频报文中均可以携带该视频流的标识。
在一些实施例中,对于音频信号(或音频数据),第一设备可以将音频信号(或音频数据)和描述性信息封装在一起,如封装在描述性信息报文DIP中。当然,为了便于接收端对音频信号的处理,第一设备可以将音频信号(或音频数据)与描述性信息分别传输。参照图8所示的视频发送适配器逻辑图,第一设备可以仅根据描述性信息封装描述性信息报文DIP,并可以根据音频信号单独封装音频采样报文(audio sample packet,ASP),在音频采样报文ASP中携带音频信号(或音频信号对应的音频数据)的信息。可选地,第一设备还可以根据传输需求,对音频采样报文ASP进行加密,传输加密后的音频采样报文ASP’。
另外,第一设备也可以按照需求,对有效视频报文AVP和音频采样报文ASP一项或多项报文进行加密。作为一种示例:第一设备在封装有效视频报文AVP和音频采样报文ASP等时,可以根据内容保护(如加密)需求,将报文头中的内容保护标识(如CP标志)置1或置0,第一设备(如第一设备中的内容保护模块)可以根据报文头中的CP标志进行处理,如果CP标志为1,则对该报文进行加密,如果CP标志为0,则不加密。
在一种可能的实现中,视频流中的视频报文(如垂直消隐报文、水平消隐报文、有效视频报文等)结构可以如图9所示,视频报文可以包括4字节(共32比特(bits)的报文头和有效载荷(payload)部分。其中,视频报文报文头各字段含义可以如表2所示。
表2

视频报文按照报文类型(或功能)可以被划分为垂直消隐报文VBP、水平消隐报文HBP、描述性信息报文DIP、有效视频报文AVP、加密描述报文EDP、密钥分发报文KDP等等。
(1)、垂直消隐报文(VBP),垂直消隐报文用于传输VBS信号,即垂直消隐开始信号(垂直消隐信号由低电平变为高电平)。垂直消隐报文VBP标志一帧的开始,垂直消隐报文VBP格式可以如图10所示,包括报文头4字节,和消息体28字节,总计32字节。当第一设备(如第一设备中的视频发送适配器)接收到VBS信号后,立即生成垂直消隐报文VBP。垂直消隐报文VBP的报文头各字段含义可以如表3所示。
表3
垂直消隐报文VBP的消息体(即有效载荷部分)可以携带用于描述视频帧相关信息的帧级控制信息(video frame control,VFC),在VFC中可以携带当前视频帧的时序、色彩空间和色彩深度等信息。垂直消隐报文的消息体(即VFC)信息可以如表4所示,包括水平有效像素点个数(HActivePixels)、水平消隐区像素点个数(HBlankPixels)、水平消隐区前肩像素点个数(HFrontPorch)、HSync像素点个数(HSyncPixels)、视频帧有效行个数(VActiveLines)、垂直消隐区行数(VBlankLines)、垂直消隐区前肩行数(VFrontPorch)、VSync行数(VSyncLines)等信息。
表4


(2)、水平消隐报文HBP,水平消隐报文HBP可以用于传输行起始信号,当第一设备(如第一设备中的视频发送适配器)-在HBlank结束时,立即生成水平消隐报文HBP。水平消隐报文HBP的格式可以如图11所示,其中水平消隐报文可以只有报文头,其报文头各字段含义可以如表5所示。
表5

(3)、有效视频报文AVP,有效视频报文AVP用于传输视频信号的有效视频数据(也可以称为有效视频像素数据)。有效视频报文AVP的格式可以如图12所示,有效视频报文可以由报文头(4字节)和负载(用于承载有效视频数据,也即有效视频像素数据)两部分组成,垂直消隐报文的报文头各字段含义可以如表6所示。
表6
对于有效视频数据中各像素分量可以采用如下排布规则:对于连续的视频像素点,第一设备(或第一设备的视频适配模块)可以优先排布位于首位的像素数据,然后按照接收顺序,排布位于首位之后的像素数据。作为一种示例:视频信号为红绿蓝(red、green、blue,RGB)视频时,有效视频报文携带的有效视频数据的排布可以参照图13所示,Pixel表示像素点,第一设备对于每一行的像素点,可以从第0个像素点开始,按照顺序依次排列该行中的像素点,直到该行的最后1个像素点结束。
其中,对于RGB格式视频像素分量排布,参照图14所示RGB格式视频像素分量排布示意图,第一设备(如第一设备中的视频适配模块)可以优先排布R分量,再排布G分量,最后排布B分量。
对于YCbCr444格式视频像素分量排布,其中Y是指亮度分量,Cb指蓝色色度分量,而Cr指红色色度分量,参照图15所示YCbCr444格式视频像素分量排布示意图,第一设备(如第一设备中的视频适配模块)可以优先排布Y分量,再排布Cb分量,最后排布Cr分量。
对于YCbCr422格式视频,参照图16所示YCbCr422格式视频像素分量排布示意图,第一设备(如第一设备中的视频适配模块)可以优先排布第0像素点的Y分量,第0像素点的Cb分量,然后是第1像素点的Y分量,第1像素点的Cr分量,再然后是第2像素点的Y分量,第2像素点的Cb分量,……,第n像素点的Y分量,第n像素点的Cb分量等等。
对于YCbCr420格式视频,第一设备(如第一设备中的视频适配模块)可以优先排布2个Y分量,再根据奇偶行排布Cb或Cr分量,即偶数行排布Cb分量,奇数行再排布Cr分量。
有效视频报文AVP的负载部分用于传输有效视频像素数据或压缩视频像素数据,其长度可以不超过508字节。当行结尾处的有效视频数据无法4字节对齐时,第一设备(如第一设备中的视频适配模块)可以填充数据“0”进行结构整型,以确保有效视频报文AVP数据负载部分4字节对齐。接收端第二设备(如第二设备中的视频适配模块)可以通过VFC中的水平有效像素点个数(HActivePixels)识别并丢弃填充数据。
作为一种示例,以分辨率为1280×720帧率为60(P60)、24位分辨率(bit resolution,bpp)的RGB视频为例,视频的数据排布如图17所示,有效视频区的行长度为1280个像素点(pixel)。像素分量中R[7]、G[7]、B[7]是最高位,R[0]、G[0]、B[0]是最低位,其中图17中仅示出了部分像素点的像素分量。视频帧每一行有1280个像素,每个像素24bit,一行为3840字节。一个有效视频报文AVP最大可以携带508字节的有效视频像素数据,1280个像素(3840字节)需要8个有效视频报文AVP传输。其中,第1个AVP(AVP0)传输像素0到像素169的R分量;AVP1从像素169的G分量开始,到像素338的G分量;……,AVP7传输像素1185的G分量开始,到像素1279。参照图18所示的有效视频报文AVP示意图,对于有效视频报文AVP0报文头中S标志需置1,标识此有效视频报文AVP为该行的首个有效视频报文AVP,有效视频报文AVP7的E标志需置1,标识此有效视频报文AVP为本行的最后一个有效视频报文AVP,其中,中间有效视频报文AVP(如有效视频报文AVP1)报文头中S标志和E标识应置0,标识为中间有效视频报文AVP。
作为又一种示例,以1280×720P60、30bpp RGB视频为例,视频的数据排布如图19所示,效视频区的行长度为1280个像素点。像素分量中R[9]、G[9]、B[9]是最高位,R[0]、G[0]、B[0]是最低位。其中图19中仅示出了部分像素点的像素分量。视频每一行有1280个像素,每个像素30bit,一行为4800字节。一个有效视频报文AVP最大可以携带508字节的有效视频像素数据,1280个像素(4800字节)需要10个有效视频报文AVP传输,其中第1个有效视频报文AVP(AVP0)传输像素0到像素135的G分量的低4位;有效视频报文AVP1从像素135的G分量高6位(G135[9:4])开始,到像素270的B分量的低8位B270[7:0];……,有效视频报文AVP9传输像素1219的R分量高4位(R1219[9:6])开始,到像素1279,共228个字节。参照图20所示的有效视频报文AVP示意图,对于有效视频报文AVP0报文头中S标志需置1,标识此有效视频报文AVP为该行的首个有效视频报文AVP,有效视频报文AVP的E标志需置1,标识此有效视频报文AVP为本行的最后一个有效视频报文AVP,其中,中间有效视频报文AVP报文(如有效视频报文AVP1)报文头中S标志和E标识应置0,标识为中间有效视频报文AVP。
(4)、描述性信息报文DIP,描述性信息报文DIP可以传输音频和视频的配置及控制信息等。描述性信息报文按传输的数据类型可以分为音频控制DIP报文,压缩参数DIP报文、视频流信息DIP报文、视频元数据DIP报文、厂商扩展DIP报文等等。其中,若第一设备和第二设备之间仅传输描述性信息报文和音频采样报文,则描述性信息报文可以在任意时刻发送;如果还有其他高优先级数据(如垂直消隐报文VBP或水平消隐报文HBP)传输,描述性信息报文DIP可以在高优先级数据或报文发送完成后发送。如图21所示为本申请实施例提供的一种描述性信息报文格式示意图,包括报文头4字节,和40字节的消息体(有效载荷)。描述性信息报文报文头各字段含义可以如表7所示。
表7

描述性信息报文DIP负载部分的前4字节(HB0~HB3)为描述性信息头,随后是32字节的数据,最后是4字节的CRC校验字节,总计40字节。描述性信息报文负载部分字段含义可以如表8所示。
表8
其中,描述性信息报文头字节0(HB0)为描述性信息类型,其含义可以如表9所示。描述性信息报文头字节1(HB1)为保留字段。描述性信息报文头字节2和字节3描述各类描述性信息的特有属性。
表9
参照如图22所示描述性信息报文DIP封装示意图,一个描述性信息报文DIP最大可以携带32字节的描述性信息。通常当待传输的描述性信息长度超过32字节时,需要拆分成多个描述性信息报文DIP,此时需将首个描述性信息报文(DIP0)的报文头中的S标志置1,将最后一个描述性信息报文(DIP2)的E标志置1来分别标识第一个和最后一个描述性信息报文DIP。
对于音频控制DIP报文,可以用于传输音频控制信息,其负载部分结构可以如表10所示。第一设备(源侧)可将音频控制DIP报文中的音频静音标志(AudioMuteFlag)设置为0b,对音频进行静音处理。例如在切换音频(不同音频)或音频格式(如采样率)以及停止播放音频前,源侧应将AudioMuteFlag设置为0b,并发送对应的音频控制DIP报文,避免宿侧输出带有噪声、毛刺等用户可察觉错误的音频。音频控制报文载荷的DB1-DB3共同标识音频采样频率,即AudioSampleFreq。AudioSampleFreq标识的采样率与实际音频采样率偏差应不超过0.5%。当音频控制报文内容变更时,第一设备(源侧)应立刻发送音频控制DIP报文;当音频控制DIP报文内容不变时,第一设备(源侧)应每1280个音频采样报文发送一次音频控制DIP报文。对于保留域,源侧应填充0x00;宿侧应具备识别、忽略保留域的能力,也应保留解析能力,保证扩展需求。
表10
对于视频流信息DIP报文,可以用于传输当前视频的功能信息,其负载部分结构可以如表11所示。
表11

视频元数据DIP报文,可以用于传输视频的元数据信息,其负载部分结构可以如表12所示。当视频元数据信息长度小于32字节,1个视频元数据DIP即可承载,此时视频元数据DIP报文头中S标和E标均需置1;同时HB3字节中保存实际的元数据长度。当视频元数据长度大于32字节时,需要封装为多个视频元数据DIP报文,每个报文传输32字节(HB3固定为32),最后一个报文不足32字节时填充0,同时在HB3中保存当前报文中实际有效的字节数;此外第1个DIP报文中的S标志置1,最后1个DIP报文的E标志置1。
表12

音频信息帧DIP报文音频信息帧的内容应符合CTA-861-G中的定义,通过音频信息帧DIP报文承载。音频信息帧DIP报文的报文头和载荷分别如下表13所示。当音频信息帧内容变更时,第一设备(源侧)应立刻发送音频信息帧DIP报文;当音频信息帧内容不变时,第一设备(源侧)应每1280个音频数据DIP报文发送一次音频信息帧报文。
表13

DIP报文还可以包括厂商扩展DIP报文、厂商特定DIP报文、视频信息帧DIP报文、源产品描述DIP报文、动态图像专家组(moving pictures experts group,MPEG)源信息帧DIP报文、动态范围DIP报文等等,具体结构可以参照上述DIP报文的描述,不再进行赘述。
(5)、音频采样报文ASP,也可以称为音频数据报文,可以用于传输音频信号(或音频数据)。如图24所示为本申请实施例提供的一种音频采样报文ASP格式示意图。
对于音频采样报文ASP,可以用于传输音频流的采样数据(脉冲编码调制(linear pulse code modulation,LPCM音频)或压缩后的采样数据(IEC 61937)。如图23所示,UMI音频数据使用类似于IEC60958子帧(Sub-Frame)的音频时隙的方式封装,一个音频时隙对应4字节,具体含义如下表14所示。如图23所示,对于非压缩音频,首先待传输的音频应遵守IEC 60958规范封包,构成IEC 60958的音频子帧。然后遵守“音频采样数据结构”规范,第一设备(如第一设备中视频适配模块/视音频适配模块)将音频子帧封包为音频时隙,最终将音频时隙封装为音频采样报文ASP进行传输。
对于压缩音频,待传输的音频应符合IEC 61937第6章映射规范(mapping rule),即将音频数据映射到IEC 60958规定的音频子帧,后续封包规范与上述非压缩音频一致。
表14
视频适配模块或视音频适配模块将遵循IEC 60958或IEC 61937协议的音频数据封包到音频采样报文ASP中,实现音频数据的传输。ASP报文结构可以如图24所示,一个音频时隙对应4字节,ASP报文中的数据长度可变,在32声道下,音频数据为128字节,即隧道报文头+ASP音频报文头+音频数据为136字节,具体的字节含义如下表15所示:
表15
(6)、加密描述报文(encryption description packet,EDP)和密钥分发报文(key distribution packet,KDP),其中加密描述报文EDP和密钥分发报文KDP的结构可以参考上述视频报文处的介绍,不再进行赘述。对于加密描述报文EDP,在加密描述报文EDP的有效载荷中可以承载加密算法等加密参数,对于密钥分发报文KDP在加密描述报文KDP的有效载荷中可以承载加密密钥。
在一些实施中,因视频报文中的垂直消隐报文VBP用于传输垂直同步信号,指示一帧视频图像的开始,加密描述报文EDP和密钥分发报文KDP用于对有效视频数据等的解密,如果这些报文传输出错,则可能出现无法接收侧的第二设备无法正确输出视频信号的问题,为保障这些报文的传输,第一设备对于垂直消隐报文VBP、加密描述报文EDP可以连续重复发送多份,密钥分发报文一帧发送一次,但要连续发送多帧。如加密描述报文EDP连续发送三份,第二设备接收到上述三种报文后,先进行正确性校验,如CRC校验,如校验未通过,则接收下一个报文,直到接收到正确的报文为止。例如:第二设备中的视频接收适配器接收到加密描述报文EDP后,先进行CRC校验,如校验未通过,则接收下一个加密描述报文EDP,直到接收到正确的加密描述报文EDP为止。
在另一些实施中,如图25所示的报文结构。也可以将垂直消隐报文VBP、加密描述报文EDP和密钥分发报文KDP报文头中R标志由保留(Reserved)变为重试(Retry),由传输层来实现可靠传输。传输层的可靠传输规则:当报文头中的R标志为0时,则接收侧(如第二设备)不管收到的报文是否正确都向后传输。而当由到R标志为1时,则必须检查该报文是否正确,如果不正确,则发起重传请求,要求前一级重传该报文。也即,第一设备可以令垂直消隐报文VBP、加密描述报文EDP和密钥分发报文KDP的报文头中包含Retry标志位,Retry标志位为1时,用于指示接收端检查报文是否正确,并在报文不正确时发起重传请求。第二设备当垂直消隐报文VBP、加密描述报文EDP和密钥分发报文KDP中任一报文不正确时,对报文发起重传。
重新回到图6,S605:第二设备根据垂直消隐报文生成垂直同步信号、根据水平消隐报文生成水平同步信号。
S606:第二设备根据垂直同步信号、水平同步信号、像素时钟信号和有效视频报文携带的有效视频数据,生成视频数据信号和显示使能信号。
S607:第二设备输出视频信号,视频信号包括垂直同步信号、水平同步信号、像素时钟信号、显示使能信号和视频数据信号。
在本申请实施例中,第二设备(或第二设备中的视频接收适配器)可以通过对来自第一设备的视频流解复合(如通过流解复用器等解复合),得到第一设备发送的垂直消隐报文、水平消隐报文等报文。需要理解的是,在本申请实施例中,将视频流解复合,得到将多类报文,可以是指根据报文的标识,对视频流中的多类报文进行识别(或划分)。以视频流中包括垂直消隐报文、水平消隐报文和有效视频报文为例,将视频流解复用,得到垂直消隐报文、水平消隐报文和有效视频报文,可以是指根据垂直消隐报文的标识(如4b'0001)、水平消隐报文的标识(如4b'0010)和有效视频报文的标识(如4b'0101),在视频流中识别(或划分)出垂直消隐报文、水平消隐报文和有效视频报文。
对于如表1所示VSync行数(VSyncLines)、HSync像素点个数(HSyncPixels)、像素时钟信号频点(PixelClockFreq)等视频信号时序参数第二设备可以和第一设备预先约定或协商,也可以通过第一设备发送的垂直消隐报文VBP中的VFC获取,本申请实施例对此不作限定。参照图26所示的视频接收适配器逻辑图,第二设备(或第二设备中的视频(或视音频)接收适配器)接收到来自第一设备的垂直消隐报文VBP时,确定垂直同步信号由低电平变为高电平,将垂直同步极性(VSyncPol)由低电平变为(或拉升为)高电平,并持续VSync行数,生成垂直同步信号。类似的,第二设备接收到来自第一设备的水平消隐报文HBP时,计数HActivePixels+HfrontPorch后,将水平同步极性(HSyncPol)由低电平变为(拉升为)高电平,并持续HSync像素点个数,生成水平同步信号。
视频数据信号(或像素数据信号)的每个传输通路在每个像素时钟周期传输一个像素对应的数据,结合图7所示的视频信号时序参数示例可知,视频数据信号(或像素数据信号)在垂直消隐区和水平消隐区没有有效像素点的传输,显示使能信号高电平时表示当前像素数据中传输的是有效视频像素数据。第二设备可以根据像素时钟信号频点(PixelClockFreq)配置本端锁相环(phase locked loop,PLL),生成本地像素时钟信号。第二设备接收到有效视频报文AVP后,可以将有效视频报文AVP携带的有效视频数据(即有效视频报文AVP携带的视频像素数据)放入“视频信号还原”模块的缓冲区。“视频信号还原”模块根据时序参数,在水平同步信号下降沿之后的HBackPorch(HBlankPixels–HSyncPixels-HFrontPorch)个像素时钟后,或者在水平同步信号上降沿之后的HSync+HBackPorch(HBlankPixels–HFrontPorch)个像素时钟后将显示使能信号拉高,同时将缓冲区中缓存的像素数据随像素时钟通过视频信号线输出。
在一些实施中,视频流中还可以包括描述性信息报文DIP,描述性信息报文DIP可以用于传输音频和视频的配置及控制信息等描述性信息,第二设备还可以根据音频和视频的配置及控制信息等描述性信息,进行音频和视频的配置及控制,并且还可以根据传输的音频采样报文ASP的音频信号(或音频数据),输出音频信号。
在一些实施中,参照图8所示的视频发送适配器逻辑图,如果第一设备启动内容保护(如加密),将有效视频报文AVP和音频数据ASP进行加密(如进行ADCP加密),第二设备在对有效视频报文AVP和音频数据ASP进行处理前,还可以获取视频流中的加密描述报文EDP携带的加密参数、密钥分发报文KDP携带的加密密钥,对加密后的有效视频报文AVP’和加密后的音频数据ASP’进行解密,得到解密后的有效视频报文AVP和解密后的音频采样报文ASP。
需要理解的是,如果第一设备和第二设备约定或预先协商有加密参数或加密密钥,视频流中还可以不包括加密描述报文EDP和密钥分发报文KDP。
在一些实施例中,可以按照图8所示的视频发送适配器逻辑图,第一设备生成垂直消隐报文VBP、水平消隐报文HBP、有效视频报文AVP和音频采样报文ASP,以及描述性信息报文DIP。并且对有效视频报文AVP和音频采样报文ASP进行加密处理。第二设备可以对有效视频报文AVP和音频采样报文ASP进行解密,并根据垂直消隐报文VBP、水平消隐报文HBP以及描述性信息报文DIP解析解密后的有效视频报文AVP和音频采样报文ASP,得到视频数据和音频数据。
另外,第一设备也可以按照需求,对垂直消隐报文VBP、水平消隐报文HBP、有效视频报文AVP和音频采样报文ASP,以及描述性信息DIP中的一项或多项报文进行加密。可以根据内容保护(如加密)需求,将报文头中的内容保护标识(如CP标志)置1或置0,第一设备(如第一设备中的内容保护模块)可以根据报文头中的CP标志进行处理,如果CP标志为1,则对该报文进行加密,如果CP标志为0,则不加密。第二设备可以判断垂直消隐报文VBP、水平消隐报文HBP、有效视频报文AVP和音频采样报文ASP,以及描述性信息DIP中的内容保护标识(如CP标志)是否为1,在为1的时候对报文进行解密。
视音频接收适配器(或视频接收适配器)和视音频发送适配器(或视频接收适配器)位于不同的设备中,为避免视音频接收适配器因时钟漂移、展频、突发等导致像素时钟和视频发送适配器的像素时钟随时间的推进而产生偏差,本申请实施例中提供了一种使用先进先出(first Iinput first output,FIFO)水线方案进行像素时钟同步处理的方法,视频接收适配器像素时钟恢复如图27所示。接收到视频信息流时,我们从垂直消隐报文VBP中解析出用于恢复视频流的关键信息像素时钟频率(PixelClockFreq)。使用一个异步FIFO完成接收域到发送域的带宽同步,基于PixelClockFreq重建视频时钟,基于其他时序信息重建视频时序并通过HBP/VBP报文位置微调像素时钟频点。
在一些实施中,视频接收适配器在重建视频像素数据流时,为避免因视频数据报文传输过程中的抖动导致部分报文延迟到达,从而出现视频数据输出中断的现象,视音频接收适配器应缓存一定量的视音频数据。为了确保缓冲区能累积一定的视频数据,以抵抗传输抖动,在视频接收适配器接收到每行的首个有效视频报文AVP时(报文中S标志为1)后,不能立即通过视频数据信号输出,而是需要经过一个累积周期(如视频延迟t个像素时钟(tVideoDelay))后,才将缓存的有效视频数据向外发送。除首个有效视频报文AVP外,其他的有效视频报文AVP报文接收到后可以直接输出。
其中,累积周期要确保缓存的视频数据基本能抵抗传输抖动对有效视频报文AVP影响,以避免视频接收适配器因缓冲区下溢而需要频繁地在有效视频区***填充数据。累积周期同样不宜过大,以避免引入过多的时延,并增加缓存成本。其中,累计周期受传输层、逻辑层以及路由设备的抖动影响。
另外,参照图28所示的视频信号传输与恢复示意图,其中垂直同步表示垂直同步信号(VSYNC)、水平同步表示水平同步信号(HSYNC)、显示使能表示显示使能信号、视频数据表示视频数据信号。①可以表示垂直同步信号垂直消隐区后肩延迟t个像素时钟(tVideoDelay-VSYNC-VBackporch);②可以表示水平同步信号水平消隐区后肩延迟t个像素时钟((tVideoDelay-HSYNC-HBackporch);③可以表示视频数据延迟t个像素时钟((tVideoDelay)。为了确保缓冲区能累积一定的视频数据,视频接收适配器在接收到垂直消隐报文VBP和水平消隐报文HBP后,也需要延迟累积周期,即延迟t(tVideoDelay)个像素时钟后再生成垂直同步信号(VSYNC)和水平同步信号(HSYNC),比如:也即接收侧的第二设备接收到垂直消隐报文VBP后,延迟一个固定时间(如tVideoDelay)后产生垂直同步信号,并将垂直同步信号持续VSyncLines*HTotal个像素时钟;接收侧的第二设备接收水平同步信号HBP报文后,延迟一个固定时间(如tVideoDelay)后产生水平同步信号,并将水平同步信号持续HSyncPixels个像素周期。接收侧的第二设备接收到有效视频报文AVP后,将其携带视频像素数据放入“视频信号还原”模块的缓冲区,并随显示使能信号进行输出。
需要理解的是视频接收适配器在接收到相应的报文头(如垂直消隐报文VBP报文头),根据报文头识别出报文类型后即可开始tVideoDelay计时,而不用等待到一个完整的报文接收结束。对于采用帧转发的***,即传输层需要完整接收一帧后才将整帧数据从传输层缓冲区复制到适配器缓冲区的***,则tVideoDelay只能从接收到完整的报文后开始计时,此时需要减去本报文对应的像素时钟周期数。如以RGB/YUV444 8bpc为例,垂直消隐报文VBP需减去9个像素时钟周期,水平消隐报文HBP减去8个像素时钟周期,有效视频报文AVP则需要根据实际的有效负载部分的长度计算需减去的像素时钟周期数。
在一些实施中,当视频接收适配器的缓冲区下溢时,可以采用填充的方式解决,即临时拉低显示使能(DE)信号,等待缓冲区有数据时再拉高显示使能信号并继续发送视频数据。如图29所示,有效视频报文AVP3延迟到达,视频接收适配器的缓冲区出现下溢,在缓存的像素数发送完成后,需要临时拉低显示使能信号;当接收到有效视频报文AVP3后立即恢复视频数据发送。
可以理解的是,为了实现上述实施例中功能,第一设备和第二设备包括了执行各个功能相应的硬件结构和/或软件模块。本领域技术人员应该很容易意识到,结合本申请中所公开的实施例描述的各示例的单元及方法步骤,本申请能够以硬件或硬件和计算机软件相结合的形式来实现。某个功能究竟以硬件还是计算机软件驱动硬件的方式来执行,取决于技术方案的特定应用场景和设计约束条件。
图30和图31为本申请的实施例提供的可能的信号传输装置的结构示意图。这些信号传输装置可以用于实现上述方法实施例中第一设备或第二设备的功能,因此也能实现上述方法实施例所具备的有益效果。在一种可能的实现中,该信号传输装置可以是第一设备或第二设备,还可以是应用于第一设备或第二设备的模块(如芯片)。
如图30所示,信号传输装置3000包括处理单元3010和接口单元3020,其中接口单元3020还可以为收发单元或输入输出接口。信号传输装置3000可用于实现上述图6中所示的方法实施例中第一设备或第二设备的功能。
当信号传输装置3000用于实现图6所示的方法实施例中第一设备的功能时:
处理单元3010,用于获取视频信号,视频信号包括垂直同步信号、水平同步信号、像素时钟信号、显示使能信号和视频数据信号;根据像素时钟信号、显示使能信号获取视频数据信号中有效视频数据;根据垂直同步信号封装垂直消隐报文、根据水平同步信号封装水平消隐报文、根据有效视频数据封装有效视频报文;
接口单元3020,用于将垂直消隐报文、水平消隐报文和有效视频报文复合为视频流并发送。
在一种可能的设计中,处理单元3010,还用于根据像素时钟信号、显示使能信号获取视频数据信号中位于消隐区的描述性信息;根据描述性信息封装描述性信息报文;接口单元3020将垂直消隐报文、水平消隐报文和有效视频报文复合为视频流并发送时,具体用于将垂直消隐报文、水平消隐报文、有效视频报文和描述性信息报文复合为视频流并发送。
在一种可能的设计中,处理单元3010根据描述性信息封装描述性信息报文时,具体用于当视频信号还包括随路音频信号时,根据描述性信息和音频信号封装描述性信息报文。
在一种可能的设计中,处理单元3010,还用于当视频信号还包括随路音频信号时,根据音频信号封装音频采样报文;接口单元3020将垂直消隐报文、水平消隐报文、有效视频报文和描述性信息报文复合为视频流并发送时,具体用于将垂直消隐报文、水平消隐报文、有效视频报文、描述性信息报文和音频采样报文复合为视频流并发送。
在一种可能的设计中,处理单元3010,还用于根据加密参数和加密密钥对有效视频报文进行加密,得到加密后的有效视频报文;根据加密参数封装加密描述报文、根据加密密钥封装密钥分发报文;接口单元3020将垂直消隐报文、水平消隐报文、有效视频报文和描述性信息报文复合为视频流并发送时,具体用于将垂直消隐报文、水平消隐报文、加密后的有效视频报文、描述性信息报文、加密描述报文和密钥分发报文复合为视频流并发送。
在一种可能的设计中,处理单元3010,还用于根据加密参数和加密密钥对有效视频报文和音频采样报文进行加密,得到加密后的有效视频报文和加密后的音频采样报文;根据加密参数封装加密描述报文、根据加密密钥封装密钥分发报文;接口单元3020将垂直消隐报文、水平消隐报文、有效视频报文、描述性信息报文和音频采样报文复合为视频流并发送时,具体用于将垂直消隐报文、水平消隐报文、加密后的有效视频报文、描述性信息报文、加密后的音频采样报文、加密描述报文和密钥分发报文复合为视频流并发送。
在一种可能的设计中,垂直消隐报文、加密描述报文和密钥分发报文在视频流中连续重复发送多份。
在一种可能的设计中,垂直消隐报文、加密描述报文和密钥分发报文的报文头中包含重试Retry标志位,Retry标志位为1时,用于指示接收端检查报文是否正确,并在报文不正确时发起重传请求。
在一种可能的设计中,处理单元3010根据垂直同步信号封装垂直消隐报文时,具体用于根据垂直同步信号和视频帧控制信息封装垂直消隐报文,视频帧控制信息包括像素时钟信号的像素时钟参数。
在一种可能的设计中,处理单元3010根据水平同步信号封装水平消隐报文时,具体用于根据水平同步信号和像素时钟信号的像素时钟计数值封装水平消隐报文。
当信号传输装置3000用于实现图6所示的方法实施例中第二设备的功能时:
接口单元3020,用于接收视频流,视频流包括垂直消隐报文、水平消隐报文和有效视频报文;
处理单元3010,用于根据垂直消隐报文生成垂直同步信号、根据水平消隐报文生成水平同步信号;根据垂直同步信号、水平同步信号、像素时钟信号和有效视频报文携带的有效视频数据,生成视频数据信号和显示使能信号;以及输出视频信号,视频信号包括垂直同步信号、水平同步信号、像素时钟信号、显示使能信号和视频数据信号。
在一种可能的设计中,视频流还包括描述性信息报文,处理单元3010根据垂直同步信号、水平同步信号、像素时钟信号和有效视频报文携带的有效视频数据,生成视频数据信号和显示使能信号时,具体用于根据垂直同步信号、水平同步信号、像素时钟信号、有效视频报文携带的有效视频数据和描述性信息报文携带的描述性信息,生成视频数据信号和显示使能信号。
在一种可能的设计中,处理单元3010,还用于根据描述性信息报文,生成音频信号,其中输出的视频信号还包括音频信号。
在一种可能的设计中,视频流还包括音频采样报文,处理单元3010,还用于根据音频采样报文生成音频信号;其中,输出的视频信号还包括音频信号。
在一种可能的设计中,视频流还包括加密描述报文和密钥分发报文,有效视频报文为加密后的有效视频报文,处理单元3010,还用于获取加密描述报文携带的加密参数、获取密钥分发报文携带的加密密钥;根据加密参数和加密密钥对加密后的有效视频报文进行解密,得到解密后的有效视频报文。
在一种可能的设计中,视频流还包括加密描述报文和密钥分发报文,有效视频报文和音频采样报文为加密后的有效视频报文和加密后的音频采样报文,处理单元3010,还用于获取加密描述报文携带的加密参数、获取密钥分发报文携带的加密密钥;根据加密参数和加密密钥对加密后的有效视频报文和加密后的音频采样报文进行解密,得到解密后的有效视频报文和解密后的音频采样报文。
在一种可能的设计中,视频流中包括连续重复发送的多份垂直消隐报文、加密描述报文和密钥分发报文,接口单元3020,还用于从连续重复发送的多份垂直消隐报文、加密描述报文和密钥分发报文中选取通过正确性校验的垂直消隐报文、加密描述报文和密钥分发报文。
在一种可能的设计中,垂直消隐报文、加密描述报文和密钥分发报文的报文头中包含重试Retry标志位,Retry标志位为1时,用于指示接收端检查报文是否正确,并在报文不正确时发起重传请求,接口单元3020,还用于当垂直消隐报文、加密描述报文和密钥分发报文中任一报文不正确时,对报文发起重传。
在一种可能的设计中,处理单元3010,还用于获取垂直消隐报文携带的视频帧控制信息,视频帧控制信息包括像素时钟参数;根据视频帧控制信息,生成像素时钟信号。
在一种可能的设计中,水平消隐报文还携带像素时钟信号的像素时钟计数值,处理单元3010,还用于根据像素时钟计数值对像素时钟信号进行修正。
如图31所示,本申请还提供一种信号传输装置3100,包括处理器3110和接口电路3120。处理器3110和接口电路3120之间相互耦合。可以理解的是,接口电路3120可以为收发器、输入输出接口、输入接口、输出接口、通信接口等。可选的,信号传输装置3100还可以包括存储器3130,用于存储处理器3110执行的指令或存储处理器3110运行指令所需要的输入数据或存储处理器3110运行指令后产生的数据。可选的,存储器3130还可以和处理器3110集成在一起。
当信号传输装置3100用于实现图6所示的方法时,处理器3110可以用于实现上述处理单元3010的功能,接口电路3120可以用于实现上述接口单元3020的功能。
可以理解的是,本申请的实施例中的处理器可以是中央处理单元(central processing unit,CPU),还可以是其它通用处理器、数字信号处理器(digital signal processor,DSP)、专用集成电路(application specific integrated circuit,ASIC)、逻辑电路、现场可编程门阵列(field programmable gate array,FPGA)或者其它可编程逻辑器件、晶体管逻辑器件,硬件部件或者其任意组合。通用处理器可以是微处理器,也可以是任何常规的处理器。
本申请的实施例中的方法步骤可以通过硬件的方式来实现,也可以由处理器执行软件指令的方式来实现。软件指令可以由相应的软件模块组成,软件模块可以被存放于随机存取存储器、闪存、只读存储器、可编程只读存储器、可擦除可编程只读存储器、电可擦除可编程只读存储器、寄存器、硬盘、移动硬盘、CD-ROM或者本领域熟知的任何其它形式的存储介质中。一种示例性的存储介质耦合至处理器,从而使处理器能够从该存储介质读取信息,且可向该存储介质写入信息。当然,存储介质也可以是处理器的组成部分。处理器和存储介质可以位于ASIC中。另外,该ASIC可以位于网络设备或终端设备中。当然,处理器和存储介质也可以作为分立组件存在于网络设备或终端设备中。
在上述实施例中,可以全部或部分地通过软件、硬件、固件或者其任意组合来实现。当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。所述计算机程序产品包括一个或多个计算机程序或指令。在计算机上加载和执行所述计算机程序或指令时,全部或部分地执行本申请实施例所述的流程或功能。所述计算机可以是通用计算机、专用计算机、计算机网络、网络设备、用户设备或者其它可编程装置。所述计算机程序或指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,所述计算机程序或指令可以从一个网络设备、终端、计算机、服务器或数据中心通过有线或无线方式向另一个网络设备、终端、计算机、服务器或数据中心进行传输。所述计算机可读存储介质可以是计算机能够存取的任何可用介质或者是集成一个或多个可用介质的服务器、数据中心等数据存储设备。所述可用介质可以是磁性介质,例如,软盘、硬盘、磁带;也可以是光介质,例如,数字视频光盘;还可以是半导体介质,例如,固态硬盘。该计算机可读存储介质可以是易失性或非易失性存储介质,或可包括易失性和非易失性两种类型的存储介质。
在本申请的各个实施例中,如果没有特殊说明以及逻辑冲突,不同的实施例之间的术语和/或描述具有一致性、且可以相互引用,不同的实施例中的技术特征根据其内在的逻辑关系可以组合形成新的实施例。
另外,需要理解,在本申请实施例中,“示例的”一词用于表示作例子、例证或说明。本申请中被描述为“示例”的任何实施例或设计方案不应被解释为比其它实施例或设计方案更优选或更具优势。确切而言,使用示例的一词旨在以具体方式呈现概念。
可以理解的是,在本申请的实施例中涉及的各种数字编号仅为描述方便进行的区分,并不用来限制本申请的实施例的范围。上述各过程的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定。

Claims (32)

  1. 一种信号传输方法,其特征在于,包括:
    获取视频信号,所述视频信号包括垂直同步信号、水平同步信号、像素时钟信号、显示使能信号和视频数据信号;
    根据所述像素时钟信号、显示使能信号获取所述视频数据信号中有效视频数据;
    根据所述垂直同步信号封装垂直消隐报文、根据所述水平同步信号封装水平消隐报文、根据所述有效视频数据封装有效视频报文;
    将所述垂直消隐报文、所述水平消隐报文和所述有效视频报文复合为视频流并发送。
  2. 如权利要求1所述的方法,其特征在于,所述方法还包括:
    根据所述像素时钟信号、显示使能信号获取所述视频数据信号中位于消隐区的描述性信息;
    根据所述描述性信息封装描述性信息报文;
    所述将所述垂直消隐报文、所述水平消隐报文和所述有效视频报文复合为视频流并发送,包括:
    将所述垂直消隐报文、所述水平消隐报文、所述有效视频报文和所述描述性信息报文复合为视频流并发送。
  3. 如权利要求2所述的方法,其特征在于,所述根据所述描述性信息封装描述性信息报文,包括:
    当所述视频信号还包括随路音频信号时,根据所述描述性信息和所述音频信号封装所述描述性信息报文。
  4. 如权利要求2所述的方法,其特征在于,所述方法还包括:
    当所述视频信号还包括随路音频信号时,根据所述音频信号封装音频采样报文;
    所述将所述垂直消隐报文、所述水平消隐报文、所述有效视频报文和所述描述性信息报文复合为视频流并发送,包括:
    将所述垂直消隐报文、所述水平消隐报文、所述有效视频报文、所述描述性信息报文和所述音频采样报文复合为视频流并发送。
  5. 如权利要求2或3所述的方法,其特征在于,所述方法还包括:
    根据加密参数和加密密钥对所述有效视频报文进行加密,得到加密后的有效视频报文;
    根据所述加密参数封装加密描述报文、根据所述加密密钥封装密钥分发报文;
    所述将所述垂直消隐报文、所述水平消隐报文、所述有效视频报文和所述描述性信息报文复合为视频流并发送,包括:
    将所述垂直消隐报文、所述水平消隐报文、所述加密后的有效视频报文、所述描述性信息报文、所述加密描述报文和所述密钥分发报文复合为视频流并发送。
  6. 如权利要求4所述的方法,其特征在于,所述方法还包括:
    根据加密参数和加密密钥对所述有效视频报文和所述音频采样报文进行加密,得到加密后的有效视频报文和加密后的音频采样报文;
    根据所述加密参数封装加密描述报文、根据所述加密密钥封装密钥分发报文;
    所述将所述垂直消隐报文、所述水平消隐报文、所述有效视频报文、所述描述性信息报文和所述音频采样报文复合为视频流并发送,包括:
    将所述垂直消隐报文、所述水平消隐报文、所述加密后的有效视频报文、所述描述性信息报文、所述加密后的音频采样报文、所述加密描述报文和所述密钥分发报文复合为视频流并发送。
  7. 如权利要求5或6所述的方法,其特征在于,所述垂直消隐报文、所述加密描述报文和所述密钥分发报文在所述视频流中连续重复发送多份。
  8. 如权利要求5或6所述的方法,其特征在于,所述垂直消隐报文、所述加密描述报文和所述密钥分发报文的报文头中包含重试Retry标志位,所述Retry标志位为1时,用于指示接收端检查报文是否正确,并在报文不正确时发起重传请求。
  9. 如权利要求1-8中任一项所述的方法,其特征在于,所述根据所述垂直同步信号封装垂直消隐报文,包括:
    根据所述垂直同步信号和视频帧控制信息封装垂直消隐报文,所述视频帧控制信息包括所述像素时钟信号的像素时钟参数。
  10. 如权利要求1-9中任一项所述的方法,其特征在于,所述根据所述水平同步信号封装水平消隐报文,包括:
    根据所述水平同步信号和所述像素时钟信号的像素时钟计数值封装所述水平消隐报文。
  11. 如权利要求1-10中任一项所述的方法,其特征在于,将所述垂直消隐报文、所述水平消隐报文和所述有效视频报文复合为视频流,包括:
    将所述垂直消隐报文、所述水平消隐报文和所述有效视频报文通过流复用器复合为视频流。
  12. 如权利要求1-11中任一项所述的方法,其特征在于,发送所述视频流包括:
    通过统一多媒体互联接口UMI的传输层和物理层发送所述视频流。
  13. 一种信号传输方法,其特征在于,包括:
    接收视频流,所述视频流包括垂直消隐报文、水平消隐报文和有效视频报文;
    根据所述垂直消隐报文生成垂直同步信号、根据所述水平消隐报文生成水平同步信号;
    根据所述垂直同步信号、所述水平同步信号、像素时钟信号和所述有效视频报文携带的有效视频数据,生成视频数据信号和显示使能信号;
    输出视频信号,所述视频信号包括所述垂直同步信号、所述水平同步信号、所述像素时钟信号、所述显示使能信号和所述视频数据信号。
  14. 如权利要求13所述的方法,其特征在于,所述视频流还包括描述性信息报文,所述根据所述垂直同步信号、所述水平同步信号、像素时钟信号和所述有效视频报文携带的有效视频数据,生成视频数据信号和显示使能信号,包括:
    根据所述垂直同步信号、所述水平同步信号、像素时钟信号、所述有效视频报文携带的有效视频数据和所述描述性信息报文携带的描述性信息,生成视频数据信号和显示使能信号。
  15. 如权利要求14所述的方法,其特征在于,所述方法还包括:
    根据所述描述性信息报文,生成音频信号,其中输出的所述视频信号还包括所述音频信号。
  16. 如权利要求14所述的方法,其特征在于,所述视频流还包括音频采样报文,所述方法还包括:
    根据所述音频采样报文生成音频信号;其中,输出的所述视频信号还包括所述音频信号。
  17. 如权利要求14或15所述的方法,其特征在于,所述视频流还包括加密描述报文和密钥分发报文,所述有效视频报文为加密后的有效视频报文,所述方法还包括:
    获取所述加密描述报文携带的加密参数、获取所述密钥分发报文携带的加密密钥;
    根据所述加密参数和所述加密密钥对所述加密后的有效视频报文进行解密,得到解密后的有效视频报文。
  18. 如权利要求16所述的方法,其特征在于,所述视频流还包括加密描述报文和密钥分发报文,所述有效视频报文和所述音频采样报文为加密后的有效视频报文和加密后的音频采样报文,所述方法还包括:
    获取所述加密描述报文携带的加密参数、获取所述密钥分发报文携带的加密密钥;
    根据所述加密参数和所述加密密钥对所述加密后的有效视频报文和加密后的音频采样报文进行解密,得到解密后的有效视频报文和解密后的音频采样报文。
  19. 如权利要求17或18所述的方法,其特征在于,所述视频流中包括连续重复发送的多份垂直消隐报文、加密描述报文和密钥分发报文,所述方法还包括:
    从所述连续重复发送的多份垂直消隐报文、加密描述报文和密钥分发报文中选取通过正确性校验的垂直消隐报文、加密描述报文和密钥分发报文。
  20. 如权利要求17或18所述的方法,其特征在于,所述垂直消隐报文、加密描述报文和密钥分发报文的报文头中包含重试Retry标志位,所述Retry标志位为1时,用于指示接收端检查报文是否正确,并在报文不正确时发起重传请求,所述方法还包括:
    当所述垂直消隐报文、所述加密描述报文和所述密钥分发报文中任一报文不正确时,对所述报文发起重传。
  21. 如权利要求13-20中任一项所述的方法,其特征在于,所述方法还包括:
    获取所述垂直消隐报文携带的视频帧控制信息,所述视频帧控制信息包括像素时钟参数;
    根据所述视频帧控制信息,生成所述像素时钟信号。
  22. 如权利要求13-21中任一项所述的方法,其特征在于,所述水平消隐报文还携带所述像素时钟信号的像素时钟计数值,所述方法还包括:
    根据所述像素时钟计数值对所述像素时钟信号进行修正。
  23. 如权利要求13-22中任一项所述的方法,其特征在于,接收所述视频流之后,所述方法还包括:
    将所述视频流通过流解复用器解复用,得到所述垂直消隐报文、水平消隐报文和有效视频报文。
  24. 如权利要求13-23中任一项所述的方法,其特征在于,所述接收视频流,包括:
    通过统一多媒体互联接口UMI的传输层和物理层接收所述视频流。
  25. 一种信号传输装置,其特征在于,包括接口单元和处理单元;
    接口单元,用于接收和发送数据;
    处理单元,用于通过所述接口单元,执行如权利要求1-12中任一项所述的方法。
  26. 一种信号传输装置,其特征在于,包括接口单元和处理单元;
    接口单元,用于接收和发送数据;
    处理单元,用于通过所述接口单元,执行如权利要求13-24中任一项所述的方法。
  27. 一种信号传输装置,其特征在于,包括处理器和接口电路,所述接口电路用于接收来自所述信号传输装置之外的其它信号传输装置的信号并传输至所述处理器,或将来自所述处理器的信号发送给所述信号传输装置之外的其它信号传输装置,所述处理器通过逻辑电路或执行指令用于实现如权利要求1-12中任一项所述的方法。
  28. 一种信号传输装置,其特征在于,包括处理器和接口电路,所述接口电路用于接收来自所述信号传输装置之外的其它信号传输装置的信号并传输至所述处理器,或将来自所述处理器的信号发送给所述信号传输装置之外的其它信号传输装置,所述处理器通过逻辑电路或执行指令用于实现如权利要求13-24中任一项所述的方法。
  29. 一种计算机程序产品,其特征在于,包含指令,当所述指令被处理器执行,使得如权利要求1-24中任一项所述的方法被实现。
  30. 一种芯片***,其特征在于,所述芯片***包括处理器,所述处理器用于与存储器耦合,所述存储器用于存储程序或指令,当所述程序或指令被所述处理器执行时,实现如权利要求1-24中任一项所述的方法。
  31. 一种计算机可读存储介质,其特征在于,所述存储介质中存储有计算机程序或指令,当所述计算机程序或指令被处理器执行时,使得如权利要求1-24中任一项所述的方法被实现。
  32. 一种信号传输***,其特征在于,包括第一设备和第二设备;
    所述第一设备用于执行如权利要求1-12中任一项所述的方法;
    所述第二设备用于执行如权利要求13-24中任一项所述的方法。
PCT/CN2023/107128 2022-07-21 2023-07-13 一种信号传输方法及装置 WO2024017125A1 (zh)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
CN202210859501.X 2022-07-21
CN202210859501 2022-07-21
CN202310189854.8 2023-02-21
CN202310189854.8A CN117440063A (zh) 2022-07-21 2023-02-21 一种信号传输方法及装置

Publications (2)

Publication Number Publication Date
WO2024017125A1 WO2024017125A1 (zh) 2024-01-25
WO2024017125A9 true WO2024017125A9 (zh) 2024-05-02

Family

ID=89545071

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2023/107128 WO2024017125A1 (zh) 2022-07-21 2023-07-13 一种信号传输方法及装置

Country Status (2)

Country Link
CN (1) CN117440063A (zh)
WO (1) WO2024017125A1 (zh)

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8397272B2 (en) * 2008-08-05 2013-03-12 Analogix Semiconductor, Inc. Multi-stream digital display interface
CN101686399B (zh) * 2008-09-28 2012-09-05 中兴通讯股份有限公司 会议电视***芯片间传输视频流的装置及方法
JP5610199B2 (ja) * 2010-06-09 2014-10-22 ソニー株式会社 受信装置、受信方法、及び通信システム
JP6500199B2 (ja) * 2017-01-13 2019-04-17 株式会社アクセル 画像処理装置、および画像処理システム
JP2021150790A (ja) * 2020-03-18 2021-09-27 ソニーグループ株式会社 送信装置、送信方法および受信装置
US11936927B2 (en) * 2020-09-23 2024-03-19 Realtek Semiconductor Corporation Transmission control system of multi-media signal, transmitter control circuit and receiver control circuit

Also Published As

Publication number Publication date
WO2024017125A1 (zh) 2024-01-25
CN117440063A (zh) 2024-01-23

Similar Documents

Publication Publication Date Title
US10999554B2 (en) Communication device and communication method
US8397272B2 (en) Multi-stream digital display interface
ES2384354T3 (es) Sistema de comunicación, método de transmisión de señales de vídeo, transmisor, método de transmisión, receptor y método de recepción
US9185385B2 (en) Transmitting apparatus, stereo image data transmitting method, receiving apparatus, and stereo image data receiving method
US10255875B2 (en) Transmission device, transmission method, reception device, reception method, and transmission/reception system
US8918569B2 (en) Streaming audio visual content simultaneously to different topologies on a wireless adapter
US9160959B2 (en) Transmission device, transmission method, reception device, reception method, and transmission/reception system
US9270929B2 (en) Formatting audio-video information compliant with first transmission format to second transmission format in integrated circuit for offloading physical layer logic for first transmission format to separate integrated circuit
US9769417B1 (en) Metadata transfer in audio video systems
US20140267902A1 (en) Transmission device and reception device for baseband video data, and transmission/reception system
US20120151537A1 (en) Method and system for asynchronous and isochronous data transmission in a high speed video network
JP6449318B2 (ja) Hdmiインタフェースを介して補助データフレームを同期送信するためのトランスミッタ、レシーバ、及び、システム
JP2014510426A (ja) パケット通信ネットワークを介して伝送されるコンテンツをストリーミングするためのクロックリカバリ機構
WO2024017125A9 (zh) 一种信号传输方法及装置
CN113348654B (zh) 发送装置、发送方法、接收装置、接收方法和发送/接收装置
CN115037847A (zh) 一种基于数据包交换的视频数据交换方法及***
TW201444372A (zh) 用於傳達具有非壓縮視頻的邊帶資料的方法、設備及系統
TWI748447B (zh) 影音介面之控制訊號傳輸電路及控制訊號接收電路
TWI814611B (zh) Hdmi設備和省電方法
CN113691744A (zh) 影音接口的控制信号传输电路及控制信号接收电路
KR20090024611A (ko) 수평 귀선 소거 기간을 사용하여 상향 채널을 확보한비대칭 점 대 점 영상 신호 송수신 장치 및 그 방법

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23842190

Country of ref document: EP

Kind code of ref document: A1