WO2021208255A1 - Video clip marking method and device, and handheld camera - Google Patents
Video clip marking method and device, and handheld camera Download PDFInfo
- Publication number
- WO2021208255A1 WO2021208255A1 PCT/CN2020/099832 CN2020099832W WO2021208255A1 WO 2021208255 A1 WO2021208255 A1 WO 2021208255A1 CN 2020099832 W CN2020099832 W CN 2020099832W WO 2021208255 A1 WO2021208255 A1 WO 2021208255A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image frame
- target image
- information
- mark
- category
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 65
- 230000014509 gene expression Effects 0.000 claims description 18
- 238000012545 processing Methods 0.000 description 41
- 230000006870 function Effects 0.000 description 23
- 241000282472 Canis lupus familiaris Species 0.000 description 17
- 230000008569 process Effects 0.000 description 12
- 238000010586 diagram Methods 0.000 description 10
- 241000282326 Felis catus Species 0.000 description 9
- 230000006872 improvement Effects 0.000 description 9
- 238000004590 computer program Methods 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 5
- 241000282414 Homo sapiens Species 0.000 description 4
- 238000011161 development Methods 0.000 description 3
- 230000013011 mating Effects 0.000 description 3
- 241001465754 Metazoa Species 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 244000025254 Cannabis sativa Species 0.000 description 1
- 241000196324 Embryophyta Species 0.000 description 1
- WHXSMMKQMYFTQS-UHFFFAOYSA-N Lithium Chemical compound [Li] WHXSMMKQMYFTQS-UHFFFAOYSA-N 0.000 description 1
- 241000353135 Psenopsis anomala Species 0.000 description 1
- XUIMIQQOPSSXEZ-UHFFFAOYSA-N Silicon Chemical compound [Si] XUIMIQQOPSSXEZ-UHFFFAOYSA-N 0.000 description 1
- 239000006227 byproduct Substances 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000008878 coupling Effects 0.000 description 1
- 238000010168 coupling process Methods 0.000 description 1
- 238000005859 coupling reaction Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000008921 facial expression Effects 0.000 description 1
- 239000012634 fragment Substances 0.000 description 1
- 229910052744 lithium Inorganic materials 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 239000000047 product Substances 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 239000010979 ruby Substances 0.000 description 1
- 229910001750 ruby Inorganic materials 0.000 description 1
- 239000000523 sample Substances 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 229910052710 silicon Inorganic materials 0.000 description 1
- 239000010703 silicon Substances 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7837—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/73—Querying
- G06F16/732—Query formulation
- G06F16/7328—Query by example, e.g. a complete video frame or video sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/7867—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title and artist information, manually generated time, location and usage information, user ratings
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/695—Control of camera direction for changing a field of view, e.g. pan, tilt or based on tracking of objects
Definitions
- the embodiments of the present application relate to the field of image processing technologies, and in particular, to a video clip marking method, device, and handheld camera.
- the description information used to describe the video clip can be generated, so that the video clip or part of the image in the video clip can be subsequently determined according to the description information corresponding to the video clip.
- Frame search, clustering and other processing
- one of the technical problems solved by the embodiments of the present invention is to provide a video segment marking method, device, and handheld camera to overcome the inconsistency of the description information recording method generated by multiple image recognition algorithms in the prior art , which is not conducive to the defects of subsequent data processing and storage.
- the embodiment of the application provides a video segment marking method, including:
- the mark description information of the video segment is obtained, wherein the mark description information includes information recorded based on bits, and the length of the bits is T*N, where T means For the number of object categories in the target image frame, N is an integer greater than or equal to 4.
- the attribute information includes identification mark information for identifying an identification mark of at least one object category corresponding to the target image frame; correspondingly, the video clip is obtained according to the attribute information corresponding to the target image frame
- the tag description information includes:
- the mark description information of the video clip is obtained.
- the identification mark information includes at least one of the following information:
- Object category information used to identify the identification mark of the object category corresponding to the target image frame; scene category information used to identify the identification mark of the scene object category corresponding to the target image frame; used to identify the target image Face category information of the recognition mark of the face object category corresponding to the frame.
- the face category information includes at least one of the following sub-information: expression sub-attribute information used to identify the recognition mark of the expression category corresponding to the target image frame; and used to identify the target image frame The orientation sub-attribute information of the identification mark of the corresponding orientation category; the gender sub-attribute information of the identification mark used to identify the gender corresponding to the target image frame.
- the method further includes: obtaining the mark recording information of the video clip according to the identification mark information corresponding to the target image frame, wherein the mark recording information is used to record in the target image frame At least one identification mark corresponding to the second target category.
- the attribute information includes time information used to identify the time stamp corresponding to the target image frame
- the method further includes: obtaining the time information of the video clip according to the time information corresponding to the target image frame.
- the first time description information and/or the second time description information wherein the first time description information is used to record the time stamp corresponding to the target image frame including at least one target mark, and the second time description information is used To record the start time stamp and the end time stamp of the video segment.
- the N is equal to 8.
- An embodiment of the present application also provides a video clip marking device, including: a memory, a processor, and a video collector, where the video collector is used to collect a target to be tracked in a target area; the memory is used to store program code; The processor calls the program code, and when the program code is executed, it is used to perform the following operations:
- the mark description information of the video segment is obtained, wherein the mark description information includes information recorded based on bits, and the length of the bits is T*N, where T means For the number of object categories in the target image frame, N is an integer greater than or equal to 4.
- An embodiment of the present application also provides a handheld camera, including the video clip marking device according to the foregoing, and is characterized in that it further includes a carrier, which is fixedly connected to the video collector and is used to carry the video. At least part of the collector.
- the carrier includes but is not limited to a handheld pan/tilt.
- the handheld PTZ is a handheld three-axis PTZ.
- the video capture device includes, but is not limited to, a handheld three-axis pan/tilt camera.
- the attribute information corresponding to at least one target image frame in the continuous image frame is obtained by recognizing the continuous image frames in the video segment; then the tag description of the video segment is obtained according to the attribute information corresponding to the target image frame Information, where the mark description information includes information recorded based on bits, the length of the bits is T*N, T represents the number of object categories in the target image frame, and N is an integer greater than or equal to 4. Therefore, the embodiment of the present invention can not only record the recognition results of consecutive image frames in the video segment by different image recognition algorithms in a unified manner, but also greatly save storage space.
- FIG. 1 is a schematic flowchart of a method for marking video clips provided in Embodiment 1 of this application;
- FIG. 2 is a schematic flowchart of a method for marking video clips provided in Embodiment 2 of the present application;
- FIG. 3 is a schematic flowchart of a method for marking video clips provided in Embodiment 3 of the present application.
- FIG. 4 is a schematic structural diagram of a video segment marking device provided in Embodiment 4 of this application.
- FIG. 5 is a schematic structural diagram of a handheld pan/tilt head provided by Embodiment 5 of the application; FIG. 5
- FIG. 6 is a schematic structural diagram of a handheld PTZ connected with a mobile phone according to Embodiment 5 of the application;
- FIG. 7 is a schematic structural diagram of a handheld pan/tilt head provided in Embodiment 5 of this application.
- Embodiment 1 of the present application provides a video segment marking method, as shown in FIG. 1.
- FIG. 1 is a schematic flowchart of a video segment marking provided by an embodiment of this application, including:
- Step S101 Recognizing continuous image frames in a video clip, and obtaining attribute information corresponding to at least one target image frame in the continuous image frames.
- the video clip includes multiple consecutive image frames, and the number of consecutive image frames in the video clip is not limited.
- one long video can be divided into multiple short video segments, and the number of consecutive image frames included in each video segment can be a fixed value or a non-fixed value.
- one or more image recognition algorithms may be used to recognize consecutive image frames in the video segment.
- the type of image recognition algorithm selected is not limited, and it can be selected according to the video processing requirements or the hardware configuration to perform the processing in practical applications.
- the target image frame is part or all of the continuous image frames in the video clip.
- the target image frame After at least one image recognition algorithm recognizes the continuous image frame, it can generate attribute information for identifying the recognition result of the target image frame.
- the type of information included in the attribute information and the way of identifying the information are not limited, and it mainly depends on the image recognition algorithm that recognizes the target image frame.
- an image recognition algorithm for identifying object categories can be used to obtain attribute information for identifying whether the target image frame includes objects such as people, cats, dogs, etc.
- an image recognition algorithm for identifying scene categories can be used to obtain an image for identifying the target Whether the frame includes the attribute information of the sky, sea, grass and other scene objects.
- Step S102 Obtain tag description information of the video clip according to the attribute information corresponding to the target image frame.
- the mark description information is used to record the description content of the image recognition result of the target image frame, so that subsequent video processing operations such as similarity comparison and clustering between video clips can be performed according to the mark description information.
- the tag description information describes the image recognition result of the target image frame.
- the tag description information can be used to describe how many cats have appeared in the video clip in total, or used to describe the magnitude of the cats that have appeared in the video clip, and so on.
- the mark description information includes information recorded based on bits, the length of the bits is T*N, T represents the number of object categories in the target image frame, and N is an integer greater than or equal to 4.
- the value of T can be determined according to the subsequent video processing requirements and/or the image recognition results of continuous image frames in the video clip; the value of N can be determined according to the subsequent video processing requirements and/or the hardware storage space for data processing Sure.
- the bit length is 3N.
- N 4 bits
- the bit length of each category is 4 bits
- the three categories of human, cat, and dog need to use a total of 12 bits for recording.
- the bit is the smallest storage unit of the computer, and the value of the bit is represented by 0 or 1.
- the rule of recording based on the bit in this embodiment Not limited, in actual applications, the bit recording rules can be set according to subsequent video processing requirements and/or video clip content.
- the tag description information is used to record the number of faces included in all target image frames of the video clip, when the value of N is set to 4, you can use 0001 to record a total of 0 faces, and use 0010 to record a total of 1 For faces, use 0100 to record a total of 2 faces, and use 1000 to record a total of more than 3 faces.
- N when the value of N is set to 5, you can use 00000 to record a total of 0 faces, use 00001 to record a total of 1 face, use 00010 to record a total of 2 faces, and use 00011 to record a total of 3 faces. A total of 4 faces, etc. are recorded using 00100.
- bit-based information to record on the one hand, the results of image processing by different image recognition algorithms can be recorded in a unified manner, which is convenient for subsequent video processing operations; on the other hand, it can also greatly save storage space.
- N is equal to 8.
- the embodiments of the present invention first identify the continuous image frames in the video clip to obtain the attribute information corresponding to at least one target image frame in the continuous image frames; then, according to the attribute information corresponding to the target image frame, Obtain the mark description information of the video segment, where the mark description information includes information recorded based on bits, the length of the bits is T*N, T represents the number of object categories in the target image frame, and N is an integer greater than or equal to 4 . Therefore, the embodiment of the present invention can not only record the recognition results of consecutive image frames in the video segment by multiple image recognition algorithms in a unified manner, but also greatly save data storage space.
- FIG. 2 is a schematic flowchart of a video segment marking provided by an embodiment of the application, including:
- Step S201 Recognizing continuous image frames in a video clip to obtain attribute information corresponding to at least one target image frame in the continuous image frames, where the attribute information includes identification mark information.
- a variety of different image recognition algorithms can be used to recognize consecutive image frames in a video clip, and the target image frame or the objects included in the target image frame are classified according to multiple angles to obtain at least one The identification mark information corresponding to the target image frame.
- the identification mark information is used to identify the identification mark of at least one object category corresponding to the target image frame.
- One target image frame or one object in the target image frame may correspond to the identification mark of one or more object categories.
- a variety of different identification marks can be included.
- the identification mark information can be included.
- Use “DOG” to identify the corresponding identification marks of the three dogs in the "object category” and use "01", “02", and “03” to identify the corresponding identification marks of the three dogs in the "dog category”.
- the identification mark information may include at least one of the following information: object category information used to identify the identification mark of the object object category corresponding to the target image frame; and the identification mark used to identify the scene object category corresponding to the target image frame Scene category information; face category information used to identify the recognition mark of the face object category corresponding to the target image frame.
- the object category is to classify the objects included in the target image frame, and the angle of the classification and the corresponding identification mark can be determined according to the video processing requirements or the adopted image recognition algorithm.
- the identification mark corresponding to the object category can be used to identify objects of different animal categories such as “people”, “cats”, and “dogs”, and it can also be used to identify different object categories such as “animals”, “plants”, and “daily necessities”. Objects.
- the scene object category is to classify the scene objects included in the target image frame.
- the angle of the classification and the corresponding identification identifier can be determined according to the video processing requirements or the image recognition algorithm used.
- the identification mark corresponding to the scene object category can be used to identify scene objects in different weather categories such as “rainy”, “sunny”, and “cloudy”, and it can also be used to identify different backgrounds such as “grassland”, “sky”, and “sea”.
- the category of scene objects can be used to identify scene objects in different weather categories such as "rainy”, “sunny”, and “cloudy”, and it can also be used to identify different backgrounds such as “grassland”, “sky”, and “sea”.
- the face object category is to classify the face objects included in the target image frame.
- the angle of the classification and the corresponding identification identifier can be determined according to the video processing requirements or the image recognition algorithm used.
- the recognition mark corresponding to the face object category can be used to identify face objects of different age groups such as “elderly”, “middle-aged”, “child”, etc., and can also be used to identify “round face” and “square face”. , "Melon seed face” and other face objects with different face shapes.
- the target image frame can be targeted
- the face objects of the face objects are recognized and identified in more categories.
- the face object category information includes at least one of the following sub-information: expression sub-attribute information used to identify the recognition mark of the expression category corresponding to the target image frame; recognition used to identify the orientation category corresponding to the target image frame The orientation sub-attribute information of the mark; the gender sub-attribute information of the identification mark used to identify the gender corresponding to the target image frame.
- the expression category is to classify the human faces included in the target image frame according to expressions.
- the recognition mark corresponding to the expression category can be used to identify facial expressions such as "laughing”, “cry”, and "in a daze”.
- the orientation category is to classify the faces included in the target image frame according to the face orientation.
- the identification mark corresponding to the orientation category can be used to identify the face orientations such as "front”, “back”, and "side”.
- Gender is to classify the faces included in the target image frame according to gender.
- the identification mark corresponding to the gender can be used to identify "male”, “female”, and "uncertain”.
- Step S202 Determine the number of identification marks corresponding to at least one first target category in the target image frame according to the identification mark information corresponding to the target image frame.
- the tag description information obtained subsequently may only describe and record more important object categories.
- at least one of all object categories may be determined as the first target category, so that all the identification marks corresponding to the first target category in the target image frame can be determined according to the identification mark information corresponding to the target image frame. quantity.
- the target image frame A, target image frame B, and target image frame C in the video clip all include the corresponding "dog” identification mark
- the identification mark "01" "And “02” mark the two dogs appearing in the target image frame A
- a dog appears in the video clip three dogs are marked with the identification marks "01", "02", and "03” in the video clip. That is, the number of all the identification marks corresponding to "dogs" in the video clip is 3.
- Step S203 Obtain mark description information of the video clip according to the number of identification marks corresponding to at least one first target category in the target image frame.
- the recording method of the mark description information of the video segment is the same as that in step S102 in the first embodiment, and the details are not described herein again in this embodiment.
- it may further include: obtaining mark recording information of the video segment according to the identification mark information corresponding to the target image frame, where , The mark recording information is used to record the identification mark corresponding to at least one second target category in the target image frame.
- the second target category can be the same or different from the aforementioned first target category; in addition, the identification mark corresponding to the second target category can be all the identification marks corresponding to the second target category, or it can be the part corresponding to the second target category.
- the identification mark can be selected reasonably according to the subsequent video processing requirements in practical applications.
- the target image frame includes three identification marks of "laugh”, “cry”, and “in a daze” corresponding to the expression category, and the identification record information can only be used to record "laugh”.
- the two identification marks of "" and “cry” can also be used to record the three identification marks of "laughing", “cry”, and "in a daze”.
- the same recording method can be used to record all the identifications corresponding to the second target category. mark.
- an int type ID may be used to record the identification mark corresponding to the second target category, where each ID corresponds to an identification mark.
- the embodiment of the present invention can obtain video clip description information used to record the number of identification marks corresponding to at least one first target category; and by selecting subsequent videos Dealing with commonly used object categories to identify continuous image frames in video clips can reduce data processing and storage; by adopting a unified way to record mark and record information, it is convenient for subsequent management and use of data.
- FIG. 3 is a schematic flowchart of a video segment marking provided by an embodiment of the application, including:
- Step S301 Recognizing continuous image frames in the video segment to obtain attribute information corresponding to at least one target image frame in the continuous image frames, where the attribute information includes identification mark information and time information.
- the continuous image frames in the video clip all include the corresponding time stamp, in order to describe the time-related information of the video clip, when the continuous image frames in the video clip are identified, they can be used to identify Time information of the timestamp corresponding to the target image frame.
- Step S302 Obtain the tag description information of the video segment according to the identification tag information corresponding to the target image frame, and obtain the first time description information and/or the second time description information according to the time information corresponding to the target image frame.
- the first time description information is used to record the time stamp corresponding to the target image frame including at least one target mark, so that the object or target image frame identified by the target mark can be determined in the video clip according to the first time description information. Time of appearance. According to the first time description information, subsequent video processing operations such as clustering and screening of target image frames or video fragments including target tags can be performed more conveniently.
- the user may focus on the appearance of a cat.
- the cat can be identified using a preset target mark in the target image frame; By obtaining the timestamp corresponding to at least one target image frame including the target mark, the total appearance time of the cat in the video segment can be determined; thus, the first time describing the appearance time of the cat in the video segment can be finally generated Description.
- the first time description information can be recorded using an array structure, where the numbers stored in the array are used to identify the timestamp corresponding to the target image frame including at least one target mark.
- the target mark is usually used to mark objects required for subsequent video processing or objects that the user pays more attention to.
- the target mark is one or more of the identification marks corresponding to at least one object category, which can be preset according to video description requirements.
- the second time description information is used to record the start time stamp and the end time stamp of the video segment, so that the start and end time of the video segment can be determined subsequently based on the second time description information.
- the start timestamp of the video segment is the timestamp corresponding to the first one of the continuous image frames of the video segment
- the end timestamp of the video segment is the timestamp corresponding to the last one of the continuous image frames of the video segment.
- the second time description information may be recorded using a series of numbers that identify the start time stamp and the end time stamp.
- the embodiment of the present invention obtains the first time description information and/or the second time description information according to the time information corresponding to the target image frame, and can describe and record the time-related information of the video clip.
- the information describing the video segment may include multiple types of tag description information, first time description information, and/or second time description information, which can better meet subsequent video processing requirements.
- FIG. 4 is a video processing device 40 provided in the fourth embodiment of the application, including: a memory 401, a processor 402, and a video collector 403. Tracking target; the memory 401 is used to store program code; the processor 402 calls the program code, and when the program code is executed, it is used to perform the following operations:
- the mark description information of the video segment is obtained, wherein the mark description information includes information recorded based on bits, and the length of the bits is T*N, where T means For the number of object categories in the target image frame, N is an integer greater than or equal to 4.
- the attribute information includes identification mark information for identifying the identification mark of at least one object category corresponding to the target image frame; correspondingly, the attribute information corresponding to the target image frame is obtained.
- the tag description information of the video clip includes:
- the mark description information of the video clip is obtained.
- the identification mark information includes at least one of the following information:
- Object category information used to identify the identification mark of the object category corresponding to the target image frame; scene category information used to identify the identification mark of the scene object category corresponding to the target image frame; used to identify the target image Face category information of the recognition mark of the face object category corresponding to the frame.
- the face category information includes at least one of the following sub-information:
- Expression sub-attribute information used to identify the recognition tag of the expression category corresponding to the target image frame; orientation sub-attribute information used to identify the recognition tag of the orientation category corresponding to the target image frame; used to identify the target image frame The gender sub-attribute information of the corresponding gender identification mark.
- the processor calls the program code, and when the program code is executed, it is further configured to perform the following operations: obtain the video clip according to the identification mark information corresponding to the target image frame The mark recording information of the, wherein the mark recording information is used to record the identification mark corresponding to at least one second target category in the target image frame.
- the attribute information includes time information used to identify the time stamp corresponding to the target image frame; the processor calls the program code, and when the program code is executed, it is also used to execute the following Operation: Obtain the first time description information and/or the second time description information of the video clip according to the time information corresponding to the target image frame; wherein the first time description information used for recording includes at least one A timestamp corresponding to the target image frame of the target mark, and the second time description information is used to record the start timestamp and the end timestamp of the video segment.
- the N is equal to 8.
- a handheld camera including the video processing device described in the fourth embodiment, further includes: a carrier, which is fixedly connected to the video collector, and is configured to carry at least a part of the video collector.
- the carrier includes, but is not limited to, a handheld pan/tilt.
- the handheld pan/tilt is a handheld three-axis pan/tilt.
- the video capture device includes, but is not limited to, a handheld three-axis pan-tilt camera.
- the handheld pan/tilt head 1 of the embodiment of the present invention includes a handle 11 and a photographing device 12 loaded on the handle 11.
- the photographing device 12 may include a three-axis pan/tilt camera , In other embodiments, it includes a pan-tilt camera with two axes or more than three axes.
- the handle 11 is provided with a display screen 13 for displaying the shooting content of the shooting device 12.
- the invention does not limit the type of the display screen 13.
- the display screen 13 By setting the display screen 13 on the handle 11 of the handheld PTZ 1, the display screen can display the shooting content of the shooting device 12, so that the user can quickly browse the pictures or videos shot by the shooting device 12 through the display screen 13, thereby improving The interaction and fun of the handheld PTZ 1 with the user meets the diverse needs of the user.
- the handle 11 is further provided with an operating function unit for controlling the camera 12, and by operating the operating function unit, the operation of the camera 12 can be controlled, for example, the opening and closing of the camera 12 can be controlled. Turning off and controlling the shooting of the shooting device 12, controlling the posture change of the pan-tilt part of the shooting device 12, etc., so that the user can quickly operate the shooting device 12.
- the operation function part may be in the form of a button, a knob or a touch screen.
- the operating function unit includes a photographing button 14 for controlling the photographing of the photographing device 12, a power/function button 15 for controlling the opening and closing of the photographing device 12 and other functions, as well as controlling the pan/tilt.
- the universal key 16 may also include other control buttons, such as image storage buttons, image playback control buttons, etc., which can be set according to actual needs.
- the operation function part and the display screen 13 are arranged on the same side of the handle 11.
- the operation function part and the display screen 13 shown in FIG. Engineering, and at the same time make the overall appearance and layout of the handheld PTZ 1 more reasonable and beautiful.
- the side of the handle 11 is provided with a function operation key A, which is used to facilitate the user to quickly and intelligently form a sheet with one key.
- a function operation key A which is used to facilitate the user to quickly and intelligently form a sheet with one key.
- the handle 11 is further provided with a card slot 17 for inserting a storage element.
- the card slot 17 is provided on the side of the handle 11 adjacent to the display screen 13, and a memory card is inserted into the card slot 17 to store the images taken by the camera 12 in the memory card. .
- arranging the card slot 17 on the side does not affect the use of other functions, and the user experience is better.
- a power supply battery for supplying power to the handle 11 and the imaging device 12 may be provided inside the handle 11.
- the power supply battery can be a lithium battery with large capacity and small size to realize the miniaturized design of the handheld pan/tilt 1.
- the handle 11 is also provided with a charging interface/USB interface 18.
- the charging interface/USB interface 18 is provided at the bottom of the handle 11 to facilitate connection with an external power source or storage device, so as to charge the power supply battery or perform data transmission.
- the handle 11 is further provided with a sound pickup hole 19 for receiving audio signals, and the sound pickup hole 19 communicates with a microphone inside.
- the sound pickup hole 19 may include one or more. It also includes an indicator light 20 for displaying status. The user can realize audio interaction with the display screen 13 through the sound pickup hole 19.
- the indicator light 20 can serve as a reminder, and the user can obtain the power status of the handheld PTZ 1 and the current execution function status through the indicator light 20.
- the sound pickup hole 19 and the indicator light 20 can also be arranged on the front of the handle 11, which is more in line with the user's usage habits and operation convenience.
- the imaging device 12 includes a pan-tilt support and a camera mounted on the pan-tilt support.
- the imager may be a camera, or an image pickup element composed of a lens and an image sensor (such as CMOS or CCD), etc., which can be specifically selected according to needs.
- the camera may be integrated on the pan-tilt support, so that the photographing device 12 is a pan-tilt camera; it may also be an external photographing device, which can be detachably connected or clamped to be mounted on the pan-tilt support.
- the pan/tilt support is a three-axis pan/tilt support
- the photographing device 12 is a three-axis pan/tilt camera.
- the three-axis pan/tilt head bracket includes a yaw axis assembly 22, a roll axis assembly 23 movably connected to the yaw axis assembly 22, and a pitch axis assembly 24 movably connected to the roll axis assembly 23.
- the camera is mounted on the pitch axis assembly 24.
- the yaw axis assembly 22 drives the camera 12 to rotate in the yaw direction.
- the pan/tilt support can also be a two-axis pan/tilt, a four-axis pan/tilt, etc., which can be specifically selected according to needs.
- a mounting portion is further provided, the mounting portion is provided at one end of the connecting arm connected to the roll shaft assembly, and the yaw shaft assembly may be set in the handle, and the yaw shaft assembly drives The camera 12 rotates in the yaw direction together.
- the handle 11 is provided with an adapter 26 for coupling with a mobile device 2 (such as a mobile phone), and the adapter 26 and the handle 11 can be Disconnect the connection.
- the adapter 26 protrudes from the side of the handle for connecting to the mobile device 2.
- the adapter 26 is connected to the mobile device 2, the handheld platform 1 and The adapter 26 is docked and used to be supported at the end of the mobile device 2.
- the handle 11 is provided with an adapter 26 for connecting with the mobile device 2 to connect the handle 11 and the mobile device 2 to each other.
- the handle 11 can be used as a base of the mobile device 2.
- the user can hold the other end of the mobile device 2 Let's pick up and operate the handheld PTZ 1 together, the connection is convenient and fast, and the product is beautiful.
- a communication connection between the handheld pan-tilt 1 and the mobile device 2 can be realized, and the camera 12 and the mobile device 2 can transmit data.
- the adapter 26 and the handle 11 are detachably connected, that is, the adapter 26 and the handle 11 can be mechanically connected or removed. Further, the adapter 26 is provided with an electrical contact portion, and the handle 11 is provided with an electrical contact matching portion that matches with the electrical contact portion.
- the adapter 26 can be removed from the handle 11.
- the adapter 26 is installed on the handle 11 to complete the mechanical connection between the adapter 26 and the handle 11, and at the same time through the electrical contact part and the electrical contact mating part. The connection ensures the electrical connection between the two, so as to realize the data transmission between the camera 12 and the mobile device 2 through the adapter 26.
- a receiving groove 27 is provided on the side of the handle 11, and the adapter 26 is slidably clamped in the receiving groove 27. After the adapter 26 is installed in the receiving slot 27, the adapter 26 partially protrudes from the receiving slot 27, and the portion of the adapter 26 protruding from the receiving slot 27 is used to connect with the mobile device 2.
- the adapter 26 when the adapter 26 is inserted into the receiving groove 27 from the adapter 26, the adapter part is flush with the receiving groove 27, and then The adapter 26 is stored in the receiving groove 27 of the handle 11.
- the adapter 26 can be inserted into the receiving groove 27 from the adapter part, so that the adapter 26 protrudes from the receiving groove 27, So that the mobile device 2 and the handle 11 are connected to each other
- the adapter 26 can be taken out of the receiving slot 27 of the handle 11, and then inserted into the receiving slot from the adapter 26 in the reverse direction 27, the adapter 26 is further stored in the handle 11.
- the adapter 26 is flush with the receiving groove 27 of the handle 11. After the adapter 26 is stored in the handle 11, the surface of the handle 11 can be ensured to be flat, and the adapter 26 is stored in the handle 11 to make it easier to carry.
- the receiving groove 27 is semi-opened on one side surface of the handle 11, which makes it easier for the adapter 26 to be slidably connected to the receiving groove 27.
- the adapter 26 can also be detachably connected to the receiving slot 27 of the handle 11 by means of a snap connection, a plug connection, or the like.
- the receiving groove 27 is provided on the side of the handle 11.
- the receiving groove 27 is clamped and covered by the cover 28, which is convenient for the user to operate, and does not affect the front and sides of the handle. The overall appearance.
- the electrical contact part and the electrical contact mating part may be electrically connected in a contact contact manner.
- the electrical contact portion can be selected as a telescopic probe, can also be selected as an electrical plug-in interface, or can be selected as an electrical contact.
- the electrical contact portion and the electrical contact mating portion can also be directly connected to each other in a surface-to-surface contact manner.
- a method for marking video clips characterized in that it comprises:
- the mark description information of the video segment is obtained, wherein the mark description information includes information recorded based on bits, and the length of the bits is T*N, where T means For the number of object categories in the target image frame, N is an integer greater than or equal to 4.
- the video segment marking method wherein the attribute information includes identification mark information for identifying identification marks of at least one object category corresponding to the target image frame; correspondingly, according to the target
- the attribute information corresponding to the image frame, and obtaining the mark description information of the video segment includes:
- the mark description information of the video clip is obtained.
- identification marking information includes at least one of the following information:
- Object category information used to identify the identification mark of the object category corresponding to the target image frame
- Scene category information used to identify the identification mark of the scene object category corresponding to the target image frame
- Face category information used to identify the recognition mark of the face object category corresponding to the target image frame.
- A4 The video clip marking method according to A3, wherein the face category information includes at least one of the following sub-information:
- the gender sub-attribute information of the identification mark used to identify the gender corresponding to the target image frame is not limited to the gender sub-attribute information of the identification mark.
- A5. The video segment marking method according to A2, wherein the method further includes:
- the mark record information of the video clip is obtained, wherein the mark record information is used to record the identification corresponding to at least one second target category in the target image frame mark.
- A6 The video segment marking method according to A2, wherein the attribute information includes time information used to identify a timestamp corresponding to the target image frame, and the method further includes:
- the time information corresponding to the target image frame obtain the first time description information and/or the second time description information of the video clip; wherein the first time description information is used for recording including at least one target mark
- the time stamp corresponding to the target image frame, and the second time description information is used to record the start time stamp and the end time stamp of the video segment.
- a video segment marking device characterized by comprising: a memory, a processor, and a video collector, the video collector is used to collect a target to be tracked in a target area; the memory is used to store program code; the processing The program code is called, and when the program code is executed, it is used to perform the following operations:
- the mark description information of the video segment is obtained, wherein the mark description information includes information recorded based on bits, and the length of the bits is T*N, where T means For the number of object categories in the target image frame, N is an integer greater than or equal to 4.
- the video clip marking device wherein the attribute information includes identification mark information for identifying identification marks of at least one object category corresponding to the target image frame; correspondingly, according to the target image
- the attribute information corresponding to the frame, and obtaining the mark description information of the video segment includes:
- the mark description information of the video clip is obtained.
- identification marking information includes at least one of the following information:
- Object category information used to identify the identification mark of the object category corresponding to the target image frame
- Scene category information used to identify the identification mark of the scene object category corresponding to the target image frame
- Face category information used to identify the recognition mark of the face object category corresponding to the target image frame.
- A11 The video clip marking device according to A10, wherein the face category information includes at least one of the following sub-information:
- the gender sub-attribute information of the identification mark used to identify the gender corresponding to the target image frame is not limited to the gender sub-attribute information of the identification mark.
- A12 The video clip marking device according to A9, wherein the processor calls the program code, and when the program code is executed, it is further configured to perform the following operations:
- the mark record information of the video clip is obtained, wherein the mark record information is used to record the identification corresponding to at least one second target category in the target image frame mark.
- the video clip marking device according to A9, wherein the attribute information includes time information used to identify the timestamp corresponding to the target image frame; the processor calls the program code, when the program code When executed, it is also used to perform the following operations:
- the time information corresponding to the target image frame obtain the first time description information and/or the second time description information of the video clip; wherein the first time description information is used for recording including at least one target mark
- the time stamp corresponding to the target image frame, and the second time description information is used to record the start time stamp and the end time stamp of the video segment.
- A14 The video segment marking device according to A8, wherein the N is equal to 8.
- a handheld camera characterized by comprising the video clip marking device according to any one of A8-A14, characterized by further comprising: a carrier, which is fixedly connected to the video collector , Used to carry at least a part of the video collector.
- A16 The handheld camera according to A15, wherein the carrier includes but is not limited to a handheld pan/tilt.
- A17 The handheld camera according to A16, wherein the handheld PTZ is a handheld three-axis PTZ.
- a programmable logic device for example, a Field Programmable Gate Array (Field Programmable Gate Array, FPGA)
- PLD Programmable Logic Device
- FPGA Field Programmable Gate Array
- HDL Hardware Description Language
- ABEL Advanced Boolean Expression Language
- AHDL Altera Hardware Description Language
- HDCal JHDL
- Lava Lava
- Lola MyHDL
- PALASM RHDL
- VHDL Very-High-Speed Integrated Circuit Hardware Description Language
- Verilog Verilog
- the controller can be implemented in any suitable manner.
- the controller can take the form of, for example, a microprocessor or a processor and a computer-readable medium storing computer-readable program codes (such as software or firmware) executable by the (micro)processor. , Logic gates, switches, application specific integrated circuits (ASICs), programmable logic controllers and embedded microcontrollers. Examples of controllers include but are not limited to the following microcontrollers: ARC625D, Atmel AT91SAM, Microchip PIC18F26K20 and Silicon Labs C8051F320, the memory controller can also be implemented as part of the memory control logic.
- controllers in addition to implementing the controller in a purely computer-readable program code manner, it is entirely possible to program the method steps to make the controller use logic gates, switches, application-specific integrated circuits, programmable logic controllers, and embedded logic.
- the same function can be realized in the form of a microcontroller or the like. Therefore, such a controller can be regarded as a hardware component, and the devices included in it for realizing various functions can also be regarded as a structure within the hardware component. Or even, the device for realizing various functions can be regarded as both a software module for realizing the method and a structure within a hardware component.
- a typical implementation device is a computer.
- the computer may be, for example, a personal computer, a laptop computer, a cell phone, a camera phone, a smart phone, a personal digital assistant, a media player, a navigation device, an email device, a game console, a tablet computer, a wearable device, or Any combination of these devices.
- These computer program instructions can also be stored in a computer-readable memory that can guide a computer or other programmable data processing equipment to work in a specific manner, so that the instructions stored in the computer-readable memory produce an article of manufacture including the instruction device.
- the device implements the functions specified in one process or multiple processes in the flowchart and/or one block or multiple blocks in the block diagram.
- These computer program instructions can also be loaded on a computer or other programmable data processing equipment, so that a series of operation steps are executed on the computer or other programmable equipment to produce computer-implemented processing, so as to execute on the computer or other programmable equipment.
- the instructions provide steps for implementing the functions specified in one process or multiple processes in the flowchart and/or one block or multiple blocks in the block diagram.
- this application can be provided as a method, a system, or a computer program product. Therefore, this application may adopt the form of a complete hardware embodiment, a complete software embodiment, or an embodiment combining software and hardware. Moreover, this application may adopt the form of a computer program product implemented on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) containing computer-usable program codes.
- a computer-usable storage media including but not limited to disk storage, CD-ROM, optical storage, etc.
- This application may be described in the general context of computer-executable instructions executed by a computer, such as a program module.
- program modules include routines, programs, objects, components, data structures, etc. that perform specific transactions or implement specific abstract data types.
- This application can also be practiced in distributed computing environments. In these distributed computing environments, remote processing devices connected through a communication network execute transactions.
- program modules can be located in local and remote computer storage media including storage devices.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Library & Information Science (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Television Signal Processing For Recording (AREA)
- Studio Devices (AREA)
Abstract
Description
Claims (10)
- 一种视频片段标记方法,其特征在于,包括:A video segment marking method, characterized in that it comprises:对视频片段中的连续图像帧进行识别,获得所述连续图像帧中至少一个目标图像帧所对应的属性信息;Recognizing continuous image frames in the video segment to obtain attribute information corresponding to at least one target image frame in the continuous image frames;根据所述目标图像帧对应的属性信息,获得所述视频片段的标记描述信息,其中,所述标记描述信息包括基于比特位来记录的信息,所述比特位的长度为T*N,T表示所述目标图像帧中的对象类别数量,N为大于或者等于4的整数。According to the attribute information corresponding to the target image frame, the mark description information of the video segment is obtained, wherein the mark description information includes information recorded based on bits, and the length of the bits is T*N, where T means For the number of object categories in the target image frame, N is an integer greater than or equal to 4.
- 根据权利要求1所述的视频片段标记方法,其特征在于,所述属性信息包括用于标识所述目标图像帧对应的至少一对象类别的识别标记的识别标记信息;对应的,根据所述目标图像帧对应的属性信息,获得所述视频片段的标记描述信息包括:The video segment marking method according to claim 1, wherein the attribute information includes identification mark information used to identify an identification mark of at least one object category corresponding to the target image frame; correspondingly, according to the target The attribute information corresponding to the image frame, and obtaining the mark description information of the video segment includes:根据所述目标图像帧对应的所述识别标记信息,确定所述目标图像帧中的至少一第一目标类别对应的所述识别标记的数量;Determine the number of the identification marks corresponding to at least one first target category in the target image frame according to the identification mark information corresponding to the target image frame;根据所述目标图像帧中的至少一第一目标类别对应的所述识别标记的数量,获得所述视频片段的标记描述信息。According to the number of the identification marks corresponding to at least one first target category in the target image frame, the mark description information of the video clip is obtained.
- 根据权利要求2所述的视频片段标记方法,其特征在于,所述识别标记信息包括下述信息中的至少其一:The video segment marking method according to claim 2, wherein the identification marking information includes at least one of the following information:用于标识所述目标图像帧对应的物体对象类别的识别标记的物体类别信息;Object category information used to identify the identification mark of the object category corresponding to the target image frame;用于标识所述目标图像帧对应的场景对象类别的识别标记的场景类别信息;Scene category information used to identify the identification mark of the scene object category corresponding to the target image frame;用于标识所述目目标图像帧对应的人脸对象类别的识别标记的人脸类别信息。Face category information used to identify the recognition mark of the face object category corresponding to the target image frame.
- 根据权利要求3所述的视频片段标记方法,其特征在于,所述人脸类别信息包括下述子信息中的至少其一:The video clip marking method according to claim 3, wherein the face category information includes at least one of the following sub-information:用于标识所述目标图像帧对应的表情类别的识别标记的表情子属性信息;Expression sub-attribute information of the recognition tag used to identify the expression category corresponding to the target image frame;用于标识所述目标图像帧对应的朝向类别的识别标记的朝向子属性信息;Orientation sub-attribute information of the identification mark used to identify the orientation category corresponding to the target image frame;用于标识所述目标图像帧对应的性别的识别标记的性别子属性信息。The gender sub-attribute information of the identification mark used to identify the gender corresponding to the target image frame.
- 根据权利要求2所述的视频片段标记方法,其特征在于,所述方法还包括:The video segment marking method according to claim 2, wherein the method further comprises:根据所述目标图像帧对应的所述识别标记信息,获得所述视频片段的标记记录信息,其中,所述标记记录信息用于记录所述目标图像帧中的至少一个第二目标类别对应的识别标记。According to the identification mark information corresponding to the target image frame, the mark record information of the video clip is obtained, wherein the mark record information is used to record the identification corresponding to at least one second target category in the target image frame mark.
- 根据权利要求2所述的视频片段标记方法,其特征在于,所述属性信息包括用于标识所述目标图像帧对应的时间戳的时间信息,所述方法还包括:The video segment marking method according to claim 2, wherein the attribute information includes time information used to identify a timestamp corresponding to the target image frame, and the method further comprises:根据所述目标图像帧对应的所述时间信息,获得所述视频片段的第一时间描述信息和/或第二时间描述信息;其中,所述第一时间描述信息用于记录包括至少一目标标记的所述目标图像帧对应的时间戳,所述第二时间描述信息用于记录所述视频片段的开始时间戳和结束时间戳。According to the time information corresponding to the target image frame, obtain the first time description information and/or the second time description information of the video clip; wherein the first time description information is used for recording including at least one target mark The time stamp corresponding to the target image frame, and the second time description information is used to record the start time stamp and the end time stamp of the video segment.
- 根据权利要求1所述的视频片段标记方法,其特征在于,所述N等于8。The video segment marking method according to claim 1, wherein the N is equal to 8.
- 一种视频片段标记设备,其特征在于,包括:存储器、处理器、视频采集器,所述视频采集器用于采集目标区域的待跟踪目标;所述存储器用于存储程序代码;所述处理器,调用所述程序代码,当程序代码被执行时,用于执行以下操作:A video segment marking device, which is characterized by comprising: a memory, a processor, and a video collector, the video collector is used to collect a target to be tracked in a target area; the memory is used to store program codes; the processor, The program code is called, and when the program code is executed, it is used to perform the following operations:对视频片段中的连续图像帧进行识别,获得所述连续图像帧中至少一个目标图像帧所对应的属性信息;Recognizing continuous image frames in the video segment to obtain attribute information corresponding to at least one target image frame in the continuous image frames;根据所述目标图像帧对应的属性信息,获得所述视频片段的标记描述信息,其中,所述标记描述信息包括基于比特位来记录的信息,所述比特位的长度为T*N,T表示所述目标图像帧中的对象类别数量,N为大于或者等于4的整数。According to the attribute information corresponding to the target image frame, the mark description information of the video segment is obtained, wherein the mark description information includes information recorded based on bits, and the length of the bits is T*N, where T means For the number of object categories in the target image frame, N is an integer greater than or equal to 4.
- 根据权利要求8所述视频片段标记设备,其特征在于,所述属性信息包括用于标识所述目标图像帧对应的至少一对象类别的识别标记的识别标记信息;对应的,根据所述目标图像帧对应的属性信息,获得所述视频片段的标记描述信息包括:The video clip marking device according to claim 8, wherein the attribute information includes identification mark information used to identify an identification mark of at least one object category corresponding to the target image frame; correspondingly, according to the target image The attribute information corresponding to the frame, and obtaining the mark description information of the video segment includes:根据所述目标图像帧对应的所述识别标记信息,确定所述目标图像帧中的至少一第一目标类别对应的所述识别标记的数量;Determine the number of the identification marks corresponding to at least one first target category in the target image frame according to the identification mark information corresponding to the target image frame;根据所述目标图像帧中的至少一第一目标类别对应的所述识别标记的数 量,获得所述视频片段的标记描述信息。According to the number of identification marks corresponding to at least one first target category in the target image frame, the mark description information of the video clip is obtained.
- 根据权利要求9所述视频片段标记设备,其特征在于,所述识别标记信息包括下述信息中的至少其一:The video clip marking device according to claim 9, wherein the identification marking information includes at least one of the following information:用于标识所述目标图像帧对应的物体对象类别的识别标记的物体类别信息;Object category information used to identify the identification mark of the object category corresponding to the target image frame;用于标识所述目标图像帧对应的场景对象类别的识别标记的场景类别信息;Scene category information used to identify the identification mark of the scene object category corresponding to the target image frame;用于标识所述目目标图像帧对应的人脸对象类别的识别标记的人脸类别信息。Face category information used to identify the recognition mark of the face object category corresponding to the target image frame.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010296290.4A CN112052357B (en) | 2020-04-15 | 2020-04-15 | Video clip marking method and device and handheld camera |
CN202010296290.4 | 2020-04-15 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021208255A1 true WO2021208255A1 (en) | 2021-10-21 |
Family
ID=73609655
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/099832 WO2021208255A1 (en) | 2020-04-15 | 2020-07-02 | Video clip marking method and device, and handheld camera |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN112052357B (en) |
WO (1) | WO2021208255A1 (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113163086B (en) * | 2021-04-07 | 2023-04-07 | 惠州Tcl云创科技有限公司 | Be applied to display device's intelligence and shoot accessory structure |
CN114598919B (en) * | 2022-03-01 | 2024-03-01 | 腾讯科技(深圳)有限公司 | Video processing method, device, computer equipment and storage medium |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106777114A (en) * | 2016-12-15 | 2017-05-31 | 北京奇艺世纪科技有限公司 | A kind of video classification methods and system |
CN108694217A (en) * | 2017-04-12 | 2018-10-23 | 合信息技术(北京)有限公司 | The label of video determines method and device |
CN109165573A (en) * | 2018-08-03 | 2019-01-08 | 百度在线网络技术(北京)有限公司 | Method and apparatus for extracting video feature vector |
US20190057258A1 (en) * | 2015-10-30 | 2019-02-21 | Hewlett-Packard Development Company, L.P. | Video Content Summarization and Class Selection |
CN110263217A (en) * | 2019-06-28 | 2019-09-20 | 北京奇艺世纪科技有限公司 | A kind of video clip label identification method and device |
CN110781960A (en) * | 2019-10-25 | 2020-02-11 | Oppo广东移动通信有限公司 | Training method, classification method, device and equipment of video classification model |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3005297B1 (en) * | 2013-06-04 | 2023-09-06 | HRL Laboratories, LLC | A system for detecting an object of interest in a scene |
CN108337532A (en) * | 2018-02-13 | 2018-07-27 | 腾讯科技(深圳)有限公司 | Perform mask method, video broadcasting method, the apparatus and system of segment |
CN109121022B (en) * | 2018-09-28 | 2020-05-05 | 百度在线网络技术(北京)有限公司 | Method and apparatus for marking video segments |
CN110166827B (en) * | 2018-11-27 | 2022-09-13 | 深圳市腾讯信息技术有限公司 | Video clip determination method and device, storage medium and electronic device |
CN110119711B (en) * | 2019-05-14 | 2021-06-11 | 北京奇艺世纪科技有限公司 | Method and device for acquiring character segments of video data and electronic equipment |
CN110458008A (en) * | 2019-07-04 | 2019-11-15 | 深圳壹账通智能科技有限公司 | Method for processing video frequency, device, computer equipment and storage medium |
-
2020
- 2020-04-15 CN CN202010296290.4A patent/CN112052357B/en active Active
- 2020-07-02 WO PCT/CN2020/099832 patent/WO2021208255A1/en active Application Filing
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190057258A1 (en) * | 2015-10-30 | 2019-02-21 | Hewlett-Packard Development Company, L.P. | Video Content Summarization and Class Selection |
CN106777114A (en) * | 2016-12-15 | 2017-05-31 | 北京奇艺世纪科技有限公司 | A kind of video classification methods and system |
CN108694217A (en) * | 2017-04-12 | 2018-10-23 | 合信息技术(北京)有限公司 | The label of video determines method and device |
CN109165573A (en) * | 2018-08-03 | 2019-01-08 | 百度在线网络技术(北京)有限公司 | Method and apparatus for extracting video feature vector |
CN110263217A (en) * | 2019-06-28 | 2019-09-20 | 北京奇艺世纪科技有限公司 | A kind of video clip label identification method and device |
CN110781960A (en) * | 2019-10-25 | 2020-02-11 | Oppo广东移动通信有限公司 | Training method, classification method, device and equipment of video classification model |
Also Published As
Publication number | Publication date |
---|---|
CN112052357B (en) | 2022-04-01 |
CN112052357A (en) | 2020-12-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11606482B2 (en) | Methods for camera movement compensation | |
US11368567B2 (en) | System and method for improving a photographic camera feature on a portable electronic device | |
WO2021208253A1 (en) | Tracking object determination method and device, and handheld camera | |
US9007431B1 (en) | Enabling the integration of a three hundred and sixty degree panoramic camera within a consumer device case | |
WO2021208256A1 (en) | Video processing method and apparatus, and handheld camera | |
WO2021208255A1 (en) | Video clip marking method and device, and handheld camera | |
US20060239648A1 (en) | System and method for marking and tagging wireless audio and video recordings | |
CN105704369B (en) | A kind of information processing method and device, electronic equipment | |
US8760551B2 (en) | Systems and methods for image capturing based on user interest | |
US20100287502A1 (en) | Image search device and image search method | |
EP2092461A1 (en) | User interface for face recognition | |
CN108886574A (en) | A kind of shooting bootstrap technique, equipment and system | |
CN102158649A (en) | Photographic device and photographic method thereof | |
CN105893997A (en) | Image and text scanning pen and scanning method for area-of-interest for users | |
WO2021208251A1 (en) | Face tracking method and face tracking device | |
CN109257649A (en) | A kind of multimedia file producting method and terminal device | |
WO2021208252A1 (en) | Tracking target determination method, device, and hand-held camera | |
WO2021208254A1 (en) | Tracking target recovery method and device, and handheld camera | |
WO2021208258A1 (en) | Method and apparatus for searching for tracked object, and hand-held camera thereof | |
WO2021208257A1 (en) | Tracking state determination method and device, and handheld camera | |
WO2022206605A1 (en) | Method for determining target object, and photographing method and device | |
WO2021208250A1 (en) | Face tracking method and face tracking device | |
WO2021208260A1 (en) | Method and device for displaying tracking frame of target object, and handheld camera | |
CN205621029U (en) | User regional image and characters wand interested | |
WO2021208261A1 (en) | Tracking target retrieving method and device, and handheld camera |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20930843 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20930843 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 03.07.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20930843 Country of ref document: EP Kind code of ref document: A1 |