WO2011140718A1 - Method for eliminating subtitles of a video program, and associated video display system - Google Patents

Method for eliminating subtitles of a video program, and associated video display system Download PDF

Info

Publication number
WO2011140718A1
WO2011140718A1 PCT/CN2010/072783 CN2010072783W WO2011140718A1 WO 2011140718 A1 WO2011140718 A1 WO 2011140718A1 CN 2010072783 W CN2010072783 W CN 2010072783W WO 2011140718 A1 WO2011140718 A1 WO 2011140718A1
Authority
WO
WIPO (PCT)
Prior art keywords
subtitle
color
predetermined region
eliminate
exists
Prior art date
Application number
PCT/CN2010/072783
Other languages
French (fr)
Inventor
Yan-wei YUAN
Original Assignee
Mediatek Singapore Pte. Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mediatek Singapore Pte. Ltd. filed Critical Mediatek Singapore Pte. Ltd.
Priority to PCT/CN2010/072783 priority Critical patent/WO2011140718A1/en
Priority to US12/918,816 priority patent/US20120249879A1/en
Priority to CN2010800183892A priority patent/CN102511047A/en
Priority to TW099132999A priority patent/TWI408957B/en
Publication of WO2011140718A1 publication Critical patent/WO2011140718A1/en

Links

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/102Programmed access in sequence to addressed parts of tracks of operating record carriers
    • G11B27/105Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/62Text, e.g. of license plates, overlay texts or captions on TV images
    • G06V20/635Overlay text, e.g. embedded captions in a TV program
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • G11B27/034Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/20Disc-shaped record carriers
    • G11B2220/25Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
    • G11B2220/2537Optical discs
    • G11B2220/2562DVDs [digital versatile discs]; Digital video discs; MMCDs; HDCDs

Definitions

  • VIDEO PROGRAM AND ASSOCIATED VIDEO
  • the present invention relates to subtitle elimination of a video program, and more particularly, to a method for eliminating subtitles of a video program, and to an associated video display system.
  • a conventional video display system such as a conventional Digital Versatile Disc (DVD) player can enable/disable subtitle display or select subtitles of a specific language, for being displayed on a screen, given that subtitle data is typically stored separately.
  • DVD Digital Versatile Disc
  • the conventional digital TV or the conventional digital TV receiver is also capable of enabling/disabling subtitle display since a subtitle data stream is typically available.
  • some problems may occur, and it seems unlikely that the related art can handle the situation properly.
  • An exemplary embodiment of a method for eliminating subtitles of a video program is provided, where each of the subtitles is originally stored as a portion of an image of the video program.
  • the method comprises: detecting whether a sub- region of a specific color exists within a predetermined region on the image, in order to determine whether a subtitle exists; and when it is detected that the subtitle exists, changing at least one color within the predetermined region to eliminate the subtitle.
  • An exemplary embodiment of an associated video display system comprises a processing circuit arranged to eliminate subtitles of a video program, wherein each of the subtitles is originally stored as a portion of an image of the video program.
  • the processing circuit comprises a detection module and an elimination module.
  • the detection module is arranged to detect whether a sub-region of a specific color exists within a predetermined region on the image, in order to determine whether a subtitle exists. Additionally, when it is detected that the subtitle exists, the elimination module changes at least one color within the predetermined region to eliminate the subtitle.
  • An exemplary embodiment of an associated video display system comprises a processing circuit arranged to eliminate subtitles of a video program, wherein each of the subtitles is originally stored as a portion of an image of the video program.
  • the processing circuit comprises an elimination module arranged to change at least one color within a predetermined region on the image to eliminate any subtitle.
  • the processing circuit further comprises a detection module arranged to selectively detect whether a sub-region of a specific color exists within the predetermined region, in order to determine whether the subtitle exists, and in a situation where the detection of the detection module is disabled, the elimination module changes the at least one color within the predetermined region to eliminate any subtitle.
  • FIG. 1 is a diagram of a video display system according to a first embodiment of the present invention.
  • FIG. 2 is a flowchart of a method for eliminating subtitles of a video program according to one embodiment of the present invention.
  • FIGS. 3A-3B illustrate some implementation details of the processing circuit shown in FIG. 1 according to an embodiment of the present invention.
  • FIG. 4 is a diagram of a video display system according to a second embodiment of the present invention.
  • FIG. 1 illustrates a diagram of a video display system 100 according to a first embodiment of the present invention.
  • the video display system 100 comprises a demultiplexer 1 10, a buffer 1 15, a video decoding circuit 120, and a processing circuit 130, where the processing circuit 130 comprises a detection module 132 and an elimination module 134.
  • the buffer 1 15 can be positioned outside the video decoding circuit 120. This is for illustrative purposes only, and is not meant to be a limitation of the present invention.
  • the buffer 1 15 can be integrated into the video decoding circuit 120.
  • the buffer 1 15 can be integrated into another component within the video display system 100.
  • the video display system 100 of this embodiment can be implemented as a digital television (TV) or a digital TV receiver, and comprises a digital tuner (not shown) for receiving broadcasting signals to generate a data stream such as a TV data stream S IN of a video program.
  • TV digital television
  • the video display system 100 can be implemented as an analog TV or an analog TV receiver with the digital tuner mentioned above being replaced with an analog tuner, which is utilized for receiving broadcasting signals to generate a video data signal instead of the TV data stream S IN .
  • the video decoding circuit 120 can be replaced by a pre-processing circuit in response to the differences between this variation and the first embodiment, where the demultiplexer 1 10 and a front stage within the pre-processing circuit can be implemented as analog components.
  • the video display system 100 of this variation may further comprise some other circuits arranged to generate an analog output signal instead of the output signal S 0UT shown in FIG. 1.
  • the digital TV or the digital TV receiver mentioned above can be taken as an example of the video display system 100.
  • the video display system 100 can be implemented as an optical storage device such as a Digital Versatile Disc (DVD) player.
  • the demultiplexer 1 10 is arranged to demultiplex the TV data stream S IN into a video data stream Sv and an audio data stream S A (not shown).
  • the video decoding circuit 120 decodes the video data stream Sv to generate one or more images of the video program, where the buffer 1 15 is arranged to temporarily store the images of the video program.
  • the processing circuit 130 is arranged to eliminate subtitles of the video program, and more particularly, the subtitles that are originally embedded in the images, where each of the subtitles is originally stored as a portion of an image of the video program. As a result, the processing circuit 130 generates the output signal S OUT that carries the images without subtitles being respectively embedded therein. More specifically, the detection module 132 is arranged to detect whether a sub-region of a specific color exists within a predetermined region on the image, in order to determine whether a subtitle exists. When it is detected that the subtitle exists, the elimination module 134 can change at least one color within the predetermined region to eliminate the subtitle.
  • the detection module 132 can perform the detection all the way, and the elimination module 134 may operate in response to the detection of the detection module 132.
  • the detection module 132 is arranged to selectively detect whether a sub-region of a specific color exists within the predetermined region, in order to determine whether the subtitle exists.
  • the detection of the detection module 132 can be enabled/disabled based upon default settings or user settings.
  • the elimination module 134 may operate in response to the detection of the detection module 132.
  • the elimination module 134 can still change the at least one color within the predetermined region to eliminate any subtitle.
  • the elimination module 134 can blur the predetermined region to eliminate the subtitle.
  • the elimination module 134 can fill the predetermined region with a predetermined color to eliminate the subtitle, where the predetermined color may represents a subtitle background color of the predetermined region.
  • the video display system 100 can properly eliminate any subtitle originally stored as a portion of the image of the video program.
  • the user can utilize a remote controller of the video display system 100 to disable subtitle display, and it really works, giving the user a good viewing experience.
  • FIG. 2 is a flowchart of a method 910 for eliminating subtitles of a video program according to one embodiment of the present invention.
  • the method 910 shown in FIG. 2 can be applied to the video display system 100 shown in FIG. 1. The method is described as follows.
  • the detection module 132 detects whether a sub-region of a specific color exists within a predetermined region on the image, in order to determine whether a subtitle exists.
  • the specific color may represent the subtitle background color of the predetermined region, and the specific color is a predetermined color such as that mentioned above.
  • the specific color in a second mode of the detection module 132, may represent a color of a stroke of the subtitle, and the sub- region comprises at least one bar, such as one or more bars respectively corresponding to one or more strokes of the subtitle.
  • the elimination module 134 changes at least one color within the predetermined region to eliminate the subtitle.
  • the elimination module 134 may operate in response to the detection corresponding to the first mode of the detection module 132.
  • the elimination module 134 may operate in response to the detection corresponding to the second mode of the detection module 132.
  • the elimination module 134 fills the predetermined region with the specific color to eliminate the subtitle.
  • the specific color such as the subtitle background color is black and the predetermined region is a rectangular region at the bottom of the image, and the detection module 132 can detect whether four sub-regions around the four corners of the rectangular region are black to determine whether the subtitle exists in the predetermined region.
  • the detection module 132 determines that the subtitle exists in the predetermined region. Then, the elimination module 134 fills the predetermined region with black, in order to eliminate the subtitle.
  • the elimination module 134 changes at least one color of the subtitle to be the aforementioned specific color (e.g. the subtitle background color such as black) to eliminate the subtitle, rather than forcibly filling the whole of the predetermined region with the specific color.
  • the sub- region comprises at least one bar, such as one or more bars respectively corresponding to one or more strokes of at least one character/word of the subtitle, where the bars represent the sub-regions that are occupied by the strokes.
  • the elimination module 134 fills the sub-region with a color of at least one pixel outside the at least one bar to eliminate the subtitle.
  • the detection module 132 can detect whether there are yellow bars around the center of the rectangular region to determine whether the subtitle exists in the predetermined region.
  • the detection module 132 determines that the subtitle exists in the predetermined region.
  • the elimination module 134 fills the predetermined region with the colors of the pixels outside the bars, and more particularly, with the respective colors of the neighboring pixels outside the bars, in order to eliminate the subtitle. This is for illustrative purposes only, and is not meant to be a limitation of the present invention.
  • the elimination module 134 when it is detected that the subtitle exists, the elimination module 134 fills the sub-region with a color mixed from at least one color of a plurality of pixels respectively positioned at different sides of the aforementioned at least one bar to eliminate the subtitle.
  • the detection module 132 of this embodiment may detect the specific color, and further notify the elimination module 134 of what the specific color is, so the elimination module 134 may operate accordingly. For example, regarding the first mode, in a situation where it is detected that the four sub-region around the four corners of the rectangular region are of a same color, the detection module 132 can determine the specific color mentioned in Step 912 to be this color. In another example, regarding the second mode, in a situation where it is detected that the aforementioned at least one bar of a same color exists, the detection module 132 can determine the specific color mentioned in Step 912 to be this color in the at least one bar.
  • the first and the second modes of the detection module 132 are involved. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, more than two modes of the detection module 132 can be implemented when needed.
  • the detection module 132 may compare the aforementioned at least one bar with one or more predetermined stroke patterns, in order to determine whether the subtitle exists in the predetermined region, where the predetermined stroke patterns may represent at least one portion of a plurality of characters (e.g. a portion or all of the plurality of characters) or represent at least one portion of a single character (e.g. a portion or the whole of the single character).
  • the detection module 132 performs operations that are combined from those of the second and the third modes, in order to achieve a better detection result, where the processing circuit 130 can be equipped with better computation ability.
  • the predetermined region can be rectangular and can be positioned at the bottom of the image. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, the predetermined region can be non-rectangular. According to another variation of this embodiment, the predetermined region can be positioned somewhere else within the image, rather than the bottom of the image.
  • the subtitle elimination operations associated to Step 914 can be automatically triggered by the detection of the detection module 132, based upon user settings and/or default settings. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, the subtitle elimination operations associated to Step 914 can be triggered manually, based upon user settings and/or default settings.
  • FIGS. 3A-3B illustrate some implementation details of the processing circuit shown in FIG. 1 according to an embodiment of the present invention.
  • the detection module 132 comprises a comparison unit 132C (labeled “Comparison” in FIG. 3 A), an area detection unit 132A (labeled “Area detection” in FIG. 3 A), and an IIR- filter 132F
  • the elimination module 134 comprises a Luma-data IIR-filter 134F and two multiplexers 134B and 134R (labeled "MUX" in FIG. 3B).
  • the notations Reg(l) and Reg(2) shown in FIG. 3 A and the notation Reg(3)[7:0] shown in FIG. 3B represent register values, and the notations Ydata_in[7:0] and Ydata_out[7:0] represent the luminance component of the input data of the processing circuit 130 and the luminance component of the output data of the processing circuit 130, respectively.
  • the notations V_cnt[10:0] and H_cnt[10:0] respectively represent vertical and horizontal locations within the image, where "V ent” stands for vertical count, and "H cnt” stands for horizontal count.
  • the notations V_cnt_cap[10:0] and H_cnt_cap[10:0] represent vertical and horizontal locations within the predetermined region, respectively.
  • the comparison unit 132C compares the vertical location V_cnt[10:0] with the register value Reg(l) to generate an indication signal Start cap, and utilize the indication signal Start cap to notify the area detection unit 132A of whether to start/stop performing area detection.
  • the area detection unit 132A can perform the area detection when the vertical location V_cnt[10:0] reaches the register value Reg(l).
  • the register value Reg(2) defines the ratio of the height of the predetermined region to the height of the image.
  • the area detection unit 132A detects the predetermined region, and inputs the luminance component Ydata_in[7:0] into the IIR- filter 132F when the current value of the vertical location V_cnt[10:0] and the current value of the horizontal location H_cnt[10:0] indicate that a current pixel under consideration is within the predetermined region.
  • the IIR- filter 132F performs filtering on the luminance component Ydata_in[7:0] to compare the luminance component Ydata_in[7:0] with a pre-set value that represents the luminance component of the aforementioned predetermined color such as the subtitle background color, in order to prevent from determining the predetermined region as an incorrect area covering non-subtitle video contents of the image. For example, in a situation where the subtitle background color is black, the IIR-filter 132F may compare the luminance component Ydata_in[7:0] with zero.
  • the vertical location V_cnt_cap[10:0] and the horizontal location H_cnt_cap[10:0] restricts the predetermined region to be a correct area that does not cover any non-subtitle video content of the image.
  • the Luma-data IIR-filter 134F performs filtering on the luminance component Ydata_in[7:0].
  • the elimination module 134 replaces the original color within the predetermined region by the aforementioned predetermined color such as the subtitle background color, where the predetermined region is restricted by the vertical location V_cnt_cap[10:0] and the horizontal location H_cnt_cap[10:0].
  • a control unit within the elimination module 134 is arranged to generate a replacement enabling signal Replace en and a calculated background color BGC cal.
  • the multiplexer 134R dynamically selects the luminance component Ydata_in[7:0] or the predetermined color such as the output of the multiplexer 134B as the luminance component Ydata_out[7:0].
  • the replacement enabling signal Replace en can be in an enabling state when the pixel under consideration is within the predetermined region, and can be in a disabling state when the pixel under consideration is outside the predetermined region.
  • the replacement enabling signal Replace en may dynamically switch between the enabling state and the disabling state when the pixel under consideration is within the predetermined region, and can be in a disabling state when the pixel under consideration is outside the predetermined region, where the replacement enabling signal Replace en can be sensitive to the stroke of the subtitle.
  • the multiplexer 134R selects the pre-set color defined by the register value Reg(3) or selects the calculated background color BGC cal, where the calculated background color BGC cal represents the background color automatically detected from the image. Similar descriptions for this embodiment are not repeated in detail.
  • FIG. 4 is a diagram of a video display system 200 according to a second embodiment of the present invention. The differences between the first and the second embodiments are described as follows.
  • the processing circuit 130 mentioned above is replaced by a processing circuit 230 executing program code 230C, where the program code 230C comprises program modules such as a detection module 232 and an elimination module 234 respectively corresponding to the detection module 132 and the elimination module 134.
  • the processing circuit 230 executing the detection module 232 typically performs the same operations as those of the detection module 132
  • the processing circuit 230 executing the elimination module 234 typically performs the same operations as those of the elimination module 134, where the detection module 232 and the elimination module 234 can be regarded as the associated software/firmware representatives of the detection module 132 and the elimination module 134, respectively. Similar descriptions for this embodiment are not repeated in detail.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Processing Of Color Television Signals (AREA)
  • Controls And Circuits For Display Device (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

A method for eliminating subtitles of a video program is provided, where each of the subtitles is originally stored as a portion of an image of the video program. The method includes: detecting whether a sub-region of a specific color exists within a predetermined region on the image, in order to determine whether a subtitle exists; and when it is detected that the subtitle exists, changing at least one color within the predetermined region to eliminate the subtitle. An associated video display system is also provided.

Description

METHOD FOR ELIMINATING SUBTITLES OF A
VIDEO PROGRAM, AND ASSOCIATED VIDEO
DISPLAY SYSTEM
TECHNICAL FIELD
[0001] The present invention relates to subtitle elimination of a video program, and more particularly, to a method for eliminating subtitles of a video program, and to an associated video display system.
BACKGROUND
[0002] According to the related art, a conventional video display system such as a conventional Digital Versatile Disc (DVD) player can enable/disable subtitle display or select subtitles of a specific language, for being displayed on a screen, given that subtitle data is typically stored separately. Taking a conventional digital television (TV) or a conventional digital TV receiver as another example of the conventional video display system, the conventional digital TV or the conventional digital TV receiver is also capable of enabling/disabling subtitle display since a subtitle data stream is typically available. However, in a situation where a subtitle is originally stored as a portion of an image of a video program, some problems may occur, and it seems unlikely that the related art can handle the situation properly.
[0003] For example, when a user is viewing a TV program that is played back with a language that is not his/her own native language, the user may rely on subtitles of the TV program to understand the conversations in the TV program. Sometimes the subtitles are not clearly displayed. Although the TV program can be broadcasted digitally, when the subtitles are originally stored with an image format, the display quality of the subtitles may still be unqualified due to various reasons. As a result, the user may try to utilize a remote controller of the conventional digital TV or the conventional digital TV receiver to disable subtitle display, but it does not work, giving the user a bad viewing experience.
[0004] In another example, when a user is viewing a TV program that is played back with a language that is not his/her own native language, the user may try to understand the conversations in the TV program without relying on subtitles of the TV program, in order to learn the language during viewing the TV program. Although the TV program can be broadcasted digitally, when the subtitles are originally stored with an image format, the user still cannot utilize the remote controller of the conventional digital TV or the conventional digital TV receiver to disable subtitle display, giving the user a bad viewing experience.
[0005] Please note that the conventional video display system does not serve the user well. Thus, a novel method is required for eliminating a subtitle originally stored as a portion of an image of a video program.
SUMMARY
[0006] It is therefore an objective of the claimed invention to provide a method for eliminating subtitles of a video program, and to provide an associated video display system, in order to solve the above-mentioned problems.
[0007] An exemplary embodiment of a method for eliminating subtitles of a video program is provided, where each of the subtitles is originally stored as a portion of an image of the video program. The method comprises: detecting whether a sub- region of a specific color exists within a predetermined region on the image, in order to determine whether a subtitle exists; and when it is detected that the subtitle exists, changing at least one color within the predetermined region to eliminate the subtitle.
[0008] An exemplary embodiment of an associated video display system comprises a processing circuit arranged to eliminate subtitles of a video program, wherein each of the subtitles is originally stored as a portion of an image of the video program. The processing circuit comprises a detection module and an elimination module. In addition, the detection module is arranged to detect whether a sub-region of a specific color exists within a predetermined region on the image, in order to determine whether a subtitle exists. Additionally, when it is detected that the subtitle exists, the elimination module changes at least one color within the predetermined region to eliminate the subtitle.
[0009] An exemplary embodiment of an associated video display system comprises a processing circuit arranged to eliminate subtitles of a video program, wherein each of the subtitles is originally stored as a portion of an image of the video program. The processing circuit comprises an elimination module arranged to change at least one color within a predetermined region on the image to eliminate any subtitle. In particular, the processing circuit further comprises a detection module arranged to selectively detect whether a sub-region of a specific color exists within the predetermined region, in order to determine whether the subtitle exists, and in a situation where the detection of the detection module is disabled, the elimination module changes the at least one color within the predetermined region to eliminate any subtitle.
[0010] These and other objectives of the present invention will no doubt become obvious to those of ordinary skill in the art after reading the following detailed description of the preferred embodiment that is illustrated in the various figures and drawings.
BRIEF DESCRIPTION OF DRAWINGS
[0011] FIG. 1 is a diagram of a video display system according to a first embodiment of the present invention.
[0012] FIG. 2 is a flowchart of a method for eliminating subtitles of a video program according to one embodiment of the present invention.
[0013] FIGS. 3A-3B illustrate some implementation details of the processing circuit shown in FIG. 1 according to an embodiment of the present invention.
[0014] FIG. 4 is a diagram of a video display system according to a second embodiment of the present invention.
DETAILED DESCRIPTION
[0015] Certain terms are used throughout the following description and claims, which refer to particular components. As one skilled in the art will appreciate, electronic equipment manufacturers may refer to a component by different names. This document does not intend to distinguish between components that differ in name but not in function. In the following description and in the claims, the terms "include" and "comprise" are used in an open-ended fashion, and thus should be interpreted to mean "include, but not limited to Also, the term "couple" is intended to mean either an indirect or direct electrical connection. Accordingly, if one device is coupled to another device, that connection may be through a direct electrical connection, or through an indirect electrical connection via other devices and connections.
[0016] Please refer to FIG. 1 , which illustrates a diagram of a video display system 100 according to a first embodiment of the present invention. As shown in FIG. 1 , the video display system 100 comprises a demultiplexer 1 10, a buffer 1 15, a video decoding circuit 120, and a processing circuit 130, where the processing circuit 130 comprises a detection module 132 and an elimination module 134. In practice, the buffer 1 15 can be positioned outside the video decoding circuit 120. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, the buffer 1 15 can be integrated into the video decoding circuit 120. According to another variation of this embodiment, the buffer 1 15 can be integrated into another component within the video display system 100.
[0017] In addition, the video display system 100 of this embodiment can be implemented as a digital television (TV) or a digital TV receiver, and comprises a digital tuner (not shown) for receiving broadcasting signals to generate a data stream such as a TV data stream SIN of a video program. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, the video display system 100 can be implemented as an analog TV or an analog TV receiver with the digital tuner mentioned above being replaced with an analog tuner, which is utilized for receiving broadcasting signals to generate a video data signal instead of the TV data stream SIN. In this variation, the video decoding circuit 120 can be replaced by a pre-processing circuit in response to the differences between this variation and the first embodiment, where the demultiplexer 1 10 and a front stage within the pre-processing circuit can be implemented as analog components. For example, the video display system 100 of this variation may further comprise some other circuits arranged to generate an analog output signal instead of the output signal S0UT shown in FIG. 1.
[0018] Please note that, according to this embodiment, the digital TV or the digital TV receiver mentioned above can be taken as an example of the video display system 100. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, the video display system 100 can be implemented as an optical storage device such as a Digital Versatile Disc (DVD) player. [0019] In this embodiment, the demultiplexer 1 10 is arranged to demultiplex the TV data stream SIN into a video data stream Sv and an audio data stream SA (not shown). The video decoding circuit 120 decodes the video data stream Sv to generate one or more images of the video program, where the buffer 1 15 is arranged to temporarily store the images of the video program. In addition, the processing circuit 130 is arranged to eliminate subtitles of the video program, and more particularly, the subtitles that are originally embedded in the images, where each of the subtitles is originally stored as a portion of an image of the video program. As a result, the processing circuit 130 generates the output signal SOUT that carries the images without subtitles being respectively embedded therein. More specifically, the detection module 132 is arranged to detect whether a sub-region of a specific color exists within a predetermined region on the image, in order to determine whether a subtitle exists. When it is detected that the subtitle exists, the elimination module 134 can change at least one color within the predetermined region to eliminate the subtitle. For example, the detection module 132 can perform the detection all the way, and the elimination module 134 may operate in response to the detection of the detection module 132. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, the detection module 132 is arranged to selectively detect whether a sub-region of a specific color exists within the predetermined region, in order to determine whether the subtitle exists.
[0020] More specifically, in this variation, the detection of the detection module 132 can be enabled/disabled based upon default settings or user settings. In a situation where the detection of the detection module is enabled, the elimination module 134 may operate in response to the detection of the detection module 132. In a situation where the detection of the detection module is disabled, the elimination module 134 can still change the at least one color within the predetermined region to eliminate any subtitle. For example, the elimination module 134 can blur the predetermined region to eliminate the subtitle. In another example, the elimination module 134 can fill the predetermined region with a predetermined color to eliminate the subtitle, where the predetermined color may represents a subtitle background color of the predetermined region.
[0021] Based upon the architecture of the first embodiment or any of its variations disclosed above, the video display system 100 can properly eliminate any subtitle originally stored as a portion of the image of the video program. In a situation where eliminating the subtitles is required, the user can utilize a remote controller of the video display system 100 to disable subtitle display, and it really works, giving the user a good viewing experience. Some implementation details are further described according to FIG. 2.
[0022] FIG. 2 is a flowchart of a method 910 for eliminating subtitles of a video program according to one embodiment of the present invention. The method 910 shown in FIG. 2 can be applied to the video display system 100 shown in FIG. 1. The method is described as follows.
[0023] In Step 912, the detection module 132 detects whether a sub-region of a specific color exists within a predetermined region on the image, in order to determine whether a subtitle exists. For example, in a first mode of the detection module 132, the specific color may represent the subtitle background color of the predetermined region, and the specific color is a predetermined color such as that mentioned above. In another example, in a second mode of the detection module 132, the specific color may represent a color of a stroke of the subtitle, and the sub- region comprises at least one bar, such as one or more bars respectively corresponding to one or more strokes of the subtitle.
[0024] In Step 914, when it is detected that the subtitle exists, the elimination module 134 changes at least one color within the predetermined region to eliminate the subtitle. For example, the elimination module 134 may operate in response to the detection corresponding to the first mode of the detection module 132. In another example, the elimination module 134 may operate in response to the detection corresponding to the second mode of the detection module 132.
[0025] Regarding the first mode of the detection module 132, the implementation details thereof are described as follows. According to this embodiment, in a situation where the specific color represents the subtitle background color of the predetermined region (e.g. the subtitle background color is typically black), when it is detected that the subtitle exists, the elimination module 134 fills the predetermined region with the specific color to eliminate the subtitle. For example, suppose that the specific color such as the subtitle background color is black and the predetermined region is a rectangular region at the bottom of the image, and the detection module 132 can detect whether four sub-regions around the four corners of the rectangular region are black to determine whether the subtitle exists in the predetermined region. When it is detected that the four sub-regions around the four corners of the rectangular region are black, the detection module 132 determines that the subtitle exists in the predetermined region. Then, the elimination module 134 fills the predetermined region with black, in order to eliminate the subtitle. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, when it is detected that the subtitle exists, the elimination module 134 changes at least one color of the subtitle to be the aforementioned specific color (e.g. the subtitle background color such as black) to eliminate the subtitle, rather than forcibly filling the whole of the predetermined region with the specific color.
[0026] Regarding the second mode of the detection module 132, the implementation details thereof are described as follows. As mentioned, in the second mode, the sub- region comprises at least one bar, such as one or more bars respectively corresponding to one or more strokes of at least one character/word of the subtitle, where the bars represent the sub-regions that are occupied by the strokes. According to this embodiment, in a situation where the specific color represents the color of the stroke of the subtitle, when it is detected that the subtitle exists, the elimination module 134 fills the sub-region with a color of at least one pixel outside the at least one bar to eliminate the subtitle. For example, suppose that the specific color such as the stroke color is yellow and the predetermined region is a rectangular region at the bottom of the image, and the detection module 132 can detect whether there are yellow bars around the center of the rectangular region to determine whether the subtitle exists in the predetermined region. When it is detected that there are yellow bars around the center of the rectangular region, the detection module 132 determines that the subtitle exists in the predetermined region. Then, the elimination module 134 fills the predetermined region with the colors of the pixels outside the bars, and more particularly, with the respective colors of the neighboring pixels outside the bars, in order to eliminate the subtitle. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, when it is detected that the subtitle exists, the elimination module 134 fills the sub-region with a color mixed from at least one color of a plurality of pixels respectively positioned at different sides of the aforementioned at least one bar to eliminate the subtitle. [0027] Please note that the specific color mentioned in Step 912 can be unknown at first. In practice, the detection module 132 of this embodiment may detect the specific color, and further notify the elimination module 134 of what the specific color is, so the elimination module 134 may operate accordingly. For example, regarding the first mode, in a situation where it is detected that the four sub-region around the four corners of the rectangular region are of a same color, the detection module 132 can determine the specific color mentioned in Step 912 to be this color. In another example, regarding the second mode, in a situation where it is detected that the aforementioned at least one bar of a same color exists, the detection module 132 can determine the specific color mentioned in Step 912 to be this color in the at least one bar.
[0028] According to this embodiment, the first and the second modes of the detection module 132 are involved. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, more than two modes of the detection module 132 can be implemented when needed. For example, in a third mode of the detection module 132, the detection module 132 may compare the aforementioned at least one bar with one or more predetermined stroke patterns, in order to determine whether the subtitle exists in the predetermined region, where the predetermined stroke patterns may represent at least one portion of a plurality of characters (e.g. a portion or all of the plurality of characters) or represent at least one portion of a single character (e.g. a portion or the whole of the single character). According to another variation of this embodiment, in a fourth mode of the detection module 132, the detection module 132 performs operations that are combined from those of the second and the third modes, in order to achieve a better detection result, where the processing circuit 130 can be equipped with better computation ability.
[0029] According to this embodiment, the predetermined region can be rectangular and can be positioned at the bottom of the image. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, the predetermined region can be non-rectangular. According to another variation of this embodiment, the predetermined region can be positioned somewhere else within the image, rather than the bottom of the image.
[0030] According to this embodiment, the subtitle elimination operations associated to Step 914 can be automatically triggered by the detection of the detection module 132, based upon user settings and/or default settings. This is for illustrative purposes only, and is not meant to be a limitation of the present invention. According to a variation of this embodiment, the subtitle elimination operations associated to Step 914 can be triggered manually, based upon user settings and/or default settings.
[0031] FIGS. 3A-3B illustrate some implementation details of the processing circuit shown in FIG. 1 according to an embodiment of the present invention. In this embodiment, the detection module 132 comprises a comparison unit 132C (labeled "Comparison" in FIG. 3 A), an area detection unit 132A (labeled "Area detection" in FIG. 3 A), and an IIR- filter 132F, and the elimination module 134 comprises a Luma-data IIR-filter 134F and two multiplexers 134B and 134R (labeled "MUX" in FIG. 3B).
[0032] The notations Reg(l) and Reg(2) shown in FIG. 3 A and the notation Reg(3)[7:0] shown in FIG. 3B represent register values, and the notations Ydata_in[7:0] and Ydata_out[7:0] represent the luminance component of the input data of the processing circuit 130 and the luminance component of the output data of the processing circuit 130, respectively. In addition, the notations V_cnt[10:0] and H_cnt[10:0] respectively represent vertical and horizontal locations within the image, where "V ent" stands for vertical count, and "H cnt" stands for horizontal count. Additionally, the notations V_cnt_cap[10:0] and H_cnt_cap[10:0] represent vertical and horizontal locations within the predetermined region, respectively.
[0033] Referring to FIG. 3 A, the comparison unit 132C compares the vertical location V_cnt[10:0] with the register value Reg(l) to generate an indication signal Start cap, and utilize the indication signal Start cap to notify the area detection unit 132A of whether to start/stop performing area detection. For example, the area detection unit 132A can perform the area detection when the vertical location V_cnt[10:0] reaches the register value Reg(l). Please note that there is an automatic area detection enabling signal Auto AD en arranged to notify the area detection unit 132A of whether to perform the area detection in an automatic mode or a manual mode. In addition, the register value Reg(2) defines the ratio of the height of the predetermined region to the height of the image. For example, when the register value Reg(2) is equal to 1/4, it restricts the predetermined region to have the height equivalent to 1/4 of the height of the image. In another example, when the register value Reg(2) is equal to 1/6, it restricts the predetermined region to have the height equivalent to 1/6 of the height of the image. As a result, the area detection unit 132A detects the predetermined region, and inputs the luminance component Ydata_in[7:0] into the IIR- filter 132F when the current value of the vertical location V_cnt[10:0] and the current value of the horizontal location H_cnt[10:0] indicate that a current pixel under consideration is within the predetermined region. Additionally, the IIR- filter 132F performs filtering on the luminance component Ydata_in[7:0] to compare the luminance component Ydata_in[7:0] with a pre-set value that represents the luminance component of the aforementioned predetermined color such as the subtitle background color, in order to prevent from determining the predetermined region as an incorrect area covering non-subtitle video contents of the image. For example, in a situation where the subtitle background color is black, the IIR-filter 132F may compare the luminance component Ydata_in[7:0] with zero. Based upon the architecture disclosed above, the vertical location V_cnt_cap[10:0] and the horizontal location H_cnt_cap[10:0] restricts the predetermined region to be a correct area that does not cover any non-subtitle video content of the image.
[0034] Referring to FIG. 3B, the Luma-data IIR-filter 134F performs filtering on the luminance component Ydata_in[7:0]. By utilizing the Luma-data IIR-filter 134F and the multiplexers 134R and 134B, the elimination module 134 replaces the original color within the predetermined region by the aforementioned predetermined color such as the subtitle background color, where the predetermined region is restricted by the vertical location V_cnt_cap[10:0] and the horizontal location H_cnt_cap[10:0]. Please note that a control unit (not shown in FIG. 3B) within the elimination module 134 is arranged to generate a replacement enabling signal Replace en and a calculated background color BGC cal. Under control of the replacement enabling signal Replace en, the multiplexer 134R dynamically selects the luminance component Ydata_in[7:0] or the predetermined color such as the output of the multiplexer 134B as the luminance component Ydata_out[7:0]. For example, the replacement enabling signal Replace en can be in an enabling state when the pixel under consideration is within the predetermined region, and can be in a disabling state when the pixel under consideration is outside the predetermined region. In another example, the replacement enabling signal Replace en may dynamically switch between the enabling state and the disabling state when the pixel under consideration is within the predetermined region, and can be in a disabling state when the pixel under consideration is outside the predetermined region, where the replacement enabling signal Replace en can be sensitive to the stroke of the subtitle. In addition, under control of an automatic background color enabling signal Auto BGC en, the multiplexer 134R selects the pre-set color defined by the register value Reg(3) or selects the calculated background color BGC cal, where the calculated background color BGC cal represents the background color automatically detected from the image. Similar descriptions for this embodiment are not repeated in detail.
[0035] FIG. 4 is a diagram of a video display system 200 according to a second embodiment of the present invention. The differences between the first and the second embodiments are described as follows.
[0036] The processing circuit 130 mentioned above is replaced by a processing circuit 230 executing program code 230C, where the program code 230C comprises program modules such as a detection module 232 and an elimination module 234 respectively corresponding to the detection module 132 and the elimination module 134. In practice, the processing circuit 230 executing the detection module 232 typically performs the same operations as those of the detection module 132, and the processing circuit 230 executing the elimination module 234 typically performs the same operations as those of the elimination module 134, where the detection module 232 and the elimination module 234 can be regarded as the associated software/firmware representatives of the detection module 132 and the elimination module 134, respectively. Similar descriptions for this embodiment are not repeated in detail.
[0037] It is an advantage of the present invention that, based upon the architecture of the embodiments/variations disclosed above, the goal of eliminating any subtitle originally stored as a portion of the image of the video program can be achieved. In a situation where eliminating the subtitles is required, the user can utilize the remote controller of the video display system or a button positioned on the video display system to disable subtitle display with ease, and the related art problems can no longer be an issue.
[0038] Those skilled in the art will readily observe that numerous modifications and alterations of the device and method may be made while retaining the teachings of the invention. Accordingly, the above disclosure should be construed as limited only by the metes and bounds of the appended claims.

Claims

1. A method for eliminating subtitles of a video program, each of the subtitles being originally stored as a portion of an image of the video program, the method comprising:
detecting whether a sub-region of a specific color exists within a predetermined region on the image, in order to determine whether a subtitle exists; and
when it is detected that the subtitle exists, changing at least one color within the predetermined region to eliminate the subtitle.
2. The method of claim 1, wherein the specific color represents a subtitle background color of the predetermined region; and the specific color is a predetermined color.
3. The method of claim 2, wherein the step of changing the at least one color within the predetermined region to eliminate the subtitle further comprises:
when it is detected that the subtitle exists, filling the predetermined region with the specific color to eliminate the subtitle.
4. The method of claim 2, wherein the step of changing the at least one color within the predetermined region to eliminate the subtitle further comprises:
when it is detected that the subtitle exists, changing at least one color of the subtitle to be the specific color to eliminate the subtitle.
5. The method of claim 2, wherein the predetermined color is black.
6. The method of claim 1, wherein the specific color represents a color of a stroke of the subtitle; and the sub-region comprises at least one bar.
7. The method of claim 6, wherein the step of changing the at least one color within the predetermined region to eliminate the subtitle further comprises:
when it is detected that the subtitle exists, filling the sub-region with a color of at least one pixel outside the at least one bar to eliminate the subtitle.
8. The method of claim 6, wherein the step of changing the at least one color within the predetermined region to eliminate the subtitle further comprises:
when it is detected that the subtitle exists, filling the sub-region with a color mixed from at least one color of a plurality of pixels respectively positioned at different sides of the bar to eliminate the subtitle.
9. The method of claim 1, wherein the predetermined region is positioned at the bottom of the image.
10. The method of claim 1, wherein the predetermined region is rectangular.
11. A video display system, comprising:
a processing circuit arranged to eliminate subtitles of a video program, wherein each of the subtitles is originally stored as a portion of an image of the video program, and the processing circuit comprises:
a detection module arranged to detect whether a sub-region of a specific color exists within a predetermined region on the image, in order to determine whether a subtitle exists; and
an elimination module, wherein when it is detected that the subtitle exists, the elimination module changes at least one color within the predetermined region to eliminate the subtitle.
12. The video display system of claim 11, wherein the specific color represents a subtitle background color of the predetermined region; and the specific color is a predetermined color.
13. The video display system of claim 12, wherein when it is detected that the subtitle exists, the elimination module fills the predetermined region with the specific color to eliminate the subtitle.
14. The video display system of claim 12, wherein when it is detected that the subtitle exists, the elimination module changes at least one color of the subtitle to be the specific color to eliminate the subtitle.
15. The video display system of claim 12, wherein the predetermined color is black.
16. The video display system of claim 11, wherein the specific color represents a color of a stroke of the subtitle; and the sub-region comprises at least one bar.
17. The video display system of claim 16, wherein when it is detected that the subtitle exists, the elimination module fills the sub-region with a color of at least one pixel outside the at least one bar to eliminate the subtitle.
18. The video display system of claim 16, wherein when it is detected that the subtitle exists, the elimination module fills the sub-region with a color mixed from at least one color of a plurality of pixels respectively positioned at different sides of the bar to eliminate the subtitle.
19. The video display system of claim 11, wherein the predetermined region is positioned at the bottom of the image.
20. The video display system of claim 11, wherein the predetermined region is rectangular.
21. A video display system, comprising:
a processing circuit arranged to eliminate subtitles of a video program, wherein each of the subtitles is originally stored as a portion of an image of the video program, and the processing circuit comprises:
an elimination module arranged to change at least one color within a predetermined region on the image to eliminate any subtitle.
22. The video display system of claim 21, wherein the processing circuit further comprises:
a detection module arranged to selectively detect whether a sub-region of a specific color exists within the predetermined region, in order to determine whether the subtitle exists;
wherein in a situation where the detection of the detection module is disabled, the elimination module changes the at least one color within the predetermined region to eliminate any subtitle.
23. The video display system of claim 21, wherein the elimination module blurs the predetermined region to eliminate the subtitle.
24. The video display system of claim 21, wherein the elimination module fills the predetermined region with a predetermined color to eliminate the subtitle.
PCT/CN2010/072783 2010-05-14 2010-05-14 Method for eliminating subtitles of a video program, and associated video display system WO2011140718A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
PCT/CN2010/072783 WO2011140718A1 (en) 2010-05-14 2010-05-14 Method for eliminating subtitles of a video program, and associated video display system
US12/918,816 US20120249879A1 (en) 2010-05-14 2010-05-14 Method for eliminating subtitles of a video program, and associated video display system
CN2010800183892A CN102511047A (en) 2010-05-14 2010-05-14 Method for eliminating subtitles of a video program, and associated video display system
TW099132999A TWI408957B (en) 2010-05-14 2010-09-29 Method for eliminating subtitles of a video program, and associated video display system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2010/072783 WO2011140718A1 (en) 2010-05-14 2010-05-14 Method for eliminating subtitles of a video program, and associated video display system

Publications (1)

Publication Number Publication Date
WO2011140718A1 true WO2011140718A1 (en) 2011-11-17

Family

ID=44913849

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2010/072783 WO2011140718A1 (en) 2010-05-14 2010-05-14 Method for eliminating subtitles of a video program, and associated video display system

Country Status (4)

Country Link
US (1) US20120249879A1 (en)
CN (1) CN102511047A (en)
TW (1) TWI408957B (en)
WO (1) WO2011140718A1 (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102802074B (en) * 2012-08-14 2015-04-08 海信集团有限公司 Method for extracting and displaying text messages from television signal and television
CN105898322A (en) * 2015-07-24 2016-08-24 乐视云计算有限公司 Video watermark removing method and device
CN109472260B (en) * 2018-10-31 2021-07-27 成都索贝数码科技股份有限公司 Method for removing station caption and subtitle in image based on deep neural network
US11216684B1 (en) * 2020-02-04 2022-01-04 Amazon Technologies, Inc. Detection and replacement of burned-in subtitles

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060088291A1 (en) * 2004-10-22 2006-04-27 Jiunn-Shyang Wang Method and device of automatic detection and modification of subtitle position
JP2006203719A (en) * 2005-01-24 2006-08-03 Sharp Corp Digital broadcast receiver
CN101115151A (en) * 2007-07-10 2008-01-30 北京大学 Method for extracting video subtitling

Family Cites Families (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH08275205A (en) * 1995-04-03 1996-10-18 Sony Corp Method and device for data coding/decoding and coded data recording medium
EP0765082A3 (en) * 1995-09-25 1999-04-07 Sony Corporation Subtitle signal encoding/decoding
US5805153A (en) * 1995-11-28 1998-09-08 Sun Microsystems, Inc. Method and system for resizing the subtitles of a video
US6201879B1 (en) * 1996-02-09 2001-03-13 Massachusetts Institute Of Technology Method and apparatus for logo hiding in images
EP0913993A1 (en) * 1997-10-28 1999-05-06 Deutsche Thomson-Brandt Gmbh Method and apparatus for automatic format detection in digital video picture
JP3653450B2 (en) * 2000-07-17 2005-05-25 三洋電機株式会社 Motion detection device
US6839094B2 (en) * 2000-12-14 2005-01-04 Rgb Systems, Inc. Method and apparatus for eliminating motion artifacts from video
US7206029B2 (en) * 2000-12-15 2007-04-17 Koninklijke Philips Electronics N.V. Picture-in-picture repositioning and/or resizing based on video content analysis
CA2330854A1 (en) * 2001-01-11 2002-07-11 Jaldi Semiconductor Corp. A system and method for detecting a non-video source in video signals
JP4197958B2 (en) * 2001-05-15 2008-12-17 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Subtitle detection in video signal
JP2003037792A (en) * 2001-07-25 2003-02-07 Toshiba Corp Data reproducing device and data reproducing method
EP1286546A1 (en) * 2001-08-02 2003-02-26 Pace Micro Technology PLC Television system allowing teletext windows repositioning
EP1408684A1 (en) * 2002-10-03 2004-04-14 STMicroelectronics S.A. Method and system for displaying video with automatic cropping
CN1237485C (en) * 2002-10-22 2006-01-18 中国科学院计算技术研究所 Method for covering face of news interviewee using quick face detection
JP4170808B2 (en) * 2003-03-31 2008-10-22 株式会社東芝 Information display device, information display method, and program
JP4552426B2 (en) * 2003-11-28 2010-09-29 カシオ計算機株式会社 Display control apparatus and display control processing program
JP4612406B2 (en) * 2004-02-09 2011-01-12 株式会社日立製作所 Liquid crystal display device
US20060045346A1 (en) * 2004-08-26 2006-03-02 Hui Zhou Method and apparatus for locating and extracting captions in a digital image
US7911536B2 (en) * 2004-09-23 2011-03-22 Intel Corporation Screen filled display of digital video content
US7489833B2 (en) * 2004-10-06 2009-02-10 Panasonic Corporation Transmitting device, reconstruction device, transmitting method and reconstruction method for broadcasts with hidden subtitles
TWI245562B (en) * 2004-11-12 2005-12-11 Via Tech Inc Apparatus for detecting the scrolling of the caption and its method
KR20060084599A (en) * 2005-01-20 2006-07-25 엘지전자 주식회사 Method and apparatus for displaying text of (an) image display device
WO2007013238A1 (en) * 2005-07-27 2007-02-01 Pioneer Corporation Video processing device and vide processing method
DE102005059765A1 (en) * 2005-12-14 2007-06-21 Patent-Treuhand-Gesellschaft für elektrische Glühlampen mbH Display device with a plurality of pixels and method for displaying images
US7672539B2 (en) * 2005-12-15 2010-03-02 General Instrument Corporation Method and apparatus for scaling selected areas of a graphics display
JP4253327B2 (en) * 2006-03-24 2009-04-08 株式会社東芝 Subtitle detection apparatus, subtitle detection method, and pull-down signal detection apparatus
JP4247638B2 (en) * 2006-04-06 2009-04-02 ソニー株式会社 Recording / reproducing apparatus and recording / reproducing method
US20070297755A1 (en) * 2006-05-31 2007-12-27 Russell Holt Personalized cutlist creation and sharing system
US20080150966A1 (en) * 2006-12-21 2008-06-26 General Instrument Corporation Method and Apparatus for Scaling Graphics Images Using Multiple Surfaces
CN105263012A (en) * 2007-03-16 2016-01-20 汤姆森许可贸易公司 System and method for combining text with three-dimensional content
CN101102419B (en) * 2007-07-10 2010-06-09 北京大学 A method for caption area of positioning video
JP5061774B2 (en) * 2007-08-02 2012-10-31 ソニー株式会社 Video signal generator
US9602757B2 (en) * 2007-09-04 2017-03-21 Apple Inc. Display of video subtitles
JP4856041B2 (en) * 2007-10-10 2012-01-18 パナソニック株式会社 Video / audio recording and playback device
US8121409B2 (en) * 2008-02-26 2012-02-21 Cyberlink Corp. Method for handling static text and logos in stabilized images
TW200941242A (en) * 2008-03-19 2009-10-01 Giga Byte Tech Co Ltd Caption coloring method using pixel as a coloring unit
JP4518194B2 (en) * 2008-06-10 2010-08-04 ソニー株式会社 Generating apparatus, generating method, and program
JP2010074772A (en) * 2008-09-22 2010-04-02 Sony Corp Video display, and video display method
CN101420556A (en) * 2008-11-19 2009-04-29 康佳集团股份有限公司 Television subtitle shielding method and system
WO2010073177A1 (en) * 2008-12-24 2010-07-01 Nxp B.V. Image processing
JP4620163B2 (en) * 2009-06-30 2011-01-26 株式会社東芝 Still subtitle detection apparatus, video device for displaying image including still subtitle, and method for processing image including still subtitle
US8196164B1 (en) * 2011-10-17 2012-06-05 Google Inc. Detecting advertisements using subtitle repetition

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060088291A1 (en) * 2004-10-22 2006-04-27 Jiunn-Shyang Wang Method and device of automatic detection and modification of subtitle position
JP2006203719A (en) * 2005-01-24 2006-08-03 Sharp Corp Digital broadcast receiver
CN101115151A (en) * 2007-07-10 2008-01-30 北京大学 Method for extracting video subtitling

Also Published As

Publication number Publication date
US20120249879A1 (en) 2012-10-04
TWI408957B (en) 2013-09-11
CN102511047A (en) 2012-06-20
TW201141207A (en) 2011-11-16

Similar Documents

Publication Publication Date Title
US8421922B2 (en) Display device, frame rate conversion device, and display method
JP3664251B2 (en) Video output device
US20120249879A1 (en) Method for eliminating subtitles of a video program, and associated video display system
US8432495B2 (en) Video processor and video processing method
US8325251B2 (en) Imaging apparatus, function control method, and function control program
US7345709B2 (en) Method and apparatus for displaying component video signals
CN1998227A (en) Device and method for indicating the detected degree of motion in video
EP1768397A2 (en) Video Processing Apparatus and Method
JP2008147826A (en) Black stripe region detection circuit
JP2008096806A (en) Video display device
US20090310019A1 (en) Image data processing apparatus and method, and reception apparatus
US20040263688A1 (en) Television receiver and control method thereof
JP4424371B2 (en) Display device
KR100607264B1 (en) An image display device and method for setting screen ratio of the same
JP2007013335A (en) Video display device
JPH1198423A (en) Display device and display method
JP4611787B2 (en) Television receiver and vertical position adjustment method
US20080316190A1 (en) Video Outputting Apparatus and Mounting Method
US7495706B2 (en) Video signal setting device for performing output setting to a display device
JP2005260796A (en) Video signal processor
KR100930560B1 (en) Method and apparatus for displaying channel number and logo of digital TV system
KR100739137B1 (en) Method to adjust image signal for digital television
KR101397036B1 (en) Method for outputting images and display unit enabling of the method
KR20070093481A (en) Method for displaying and switching external input in digital video device
US20110116553A1 (en) Image processing device and image processing method

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201080018389.2

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 12918816

Country of ref document: US

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10851225

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 18/02/2013)

122 Ep: pct application non-entry in european phase

Ref document number: 10851225

Country of ref document: EP

Kind code of ref document: A1