US11323659B2 - Video communication device, video communication method, and video communication mediating method - Google Patents

Video communication device, video communication method, and video communication mediating method Download PDF

Info

Publication number
US11323659B2
US11323659B2 US17/183,890 US202117183890A US11323659B2 US 11323659 B2 US11323659 B2 US 11323659B2 US 202117183890 A US202117183890 A US 202117183890A US 11323659 B2 US11323659 B2 US 11323659B2
Authority
US
United States
Prior art keywords
terminal
user
video communication
abusive
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US17/183,890
Other versions
US20210203880A1 (en
Inventor
Sangil AHN
Kangsik Jung
Hyountaek YONG
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hyperconnect LLC
Original Assignee
Hyperconnect LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hyperconnect LLC filed Critical Hyperconnect LLC
Assigned to HYPERCONNECT, INC. reassignment HYPERCONNECT, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Ahn, Sangil, Jung, Kangsik, Yong, Hyountaek
Publication of US20210203880A1 publication Critical patent/US20210203880A1/en
Priority to US17/659,130 priority Critical patent/US11722638B2/en
Application granted granted Critical
Publication of US11323659B2 publication Critical patent/US11323659B2/en
Assigned to HYPERCONNECT INC. reassignment HYPERCONNECT INC. CONFIRMATION OF ASSIGNMENT Assignors: YONG, HYOUN TAEK
Assigned to HYPERCONNECT INC. reassignment HYPERCONNECT INC. CONFIRMATION OF ASSIGNMENT Assignors: JUNG, KANG SIK
Assigned to HYPERCONNECT INC. reassignment HYPERCONNECT INC. CONFIRMATION OF ASSIGNMENT Assignors: AHN, SANG IL
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/141Systems for two-way working between two video terminals, e.g. videophone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/141Systems for two-way working between two video terminals, e.g. videophone
    • H04N7/147Communication arrangements, e.g. identifying the communication as a video-communication, intermediate storage of the signals
    • G06K9/00711
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • G10L15/30Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/439Processing of audio elementary streams
    • H04N21/4394Processing of audio elementary streams involving operations for analysing the audio stream, e.g. detecting features or characteristics in audio streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44008Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/454Content or additional data filtering, e.g. blocking advertisements
    • H04N21/4542Blocking scenes or portions of the received content, e.g. censoring scenes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • H04N7/152Multipoint control units therefor

Definitions

  • One or more embodiments relate to a video communication device, a video communication method, and a video communication mediating method, which are capable of preventing a user who is making a video communication from being exposed to an inappropriate image or sound from the other party.
  • terminals With the development of communication technology and the miniaturization of electronic devices, personal terminals have become widespread to general consumers. In particular, portable personal terminals such as smart phones or smart tablets have been widely used in recent years. Most of terminals include a communication function. A user may use a terminal to perform a search on the Internet or exchange messages with the other party.
  • terminals such as smart phones include cameras, microphones, displays, and speakers.
  • a user can record a voice or capture a video containing voice using a terminal. The user can confirm the recorded voice through a speaker included in the terminal or confirm the captured video through a display.
  • a user can share the recorded voice or the captured video with the other party using a communication function of a terminal.
  • the user can transmit the previously recorded voice or the previously captured video to the other party.
  • the user can also transmit a voice currently being recorded by the terminal or a video currently being captured to the other party in real time.
  • the other party can transmit a voice currently being recorded or a movie currently being captured by a terminal to the user in real time.
  • the display included in the user's terminal can simultaneously display a video currently being captured by the user's terminal and a video currently being captured by the other party's terminal.
  • a speaker included in the terminal of the user can simultaneously reproduce the voice currently being recorded by the terminal of the other party.
  • the user and the other party can make a video communication with each other using their own terminals.
  • the user and the other party may already be acquainted with each other or may not be acquainted with each other.
  • a plurality of terminals including the terminal of the user and the terminal of the other party may be connected to the server.
  • the server can mediate between the terminal of the user and the terminal of the other party. Therefore, even if the user and the other party do not know each other, they can make a video communication with each other through mediation of the server.
  • a naked body image of the other party captured by the other party's terminal may be received at the user's terminal through the video communication session.
  • the user's terminal displays the received image so that the user can be exposed to the sexually obscene image.
  • the user may feel sexual shame or discomfort.
  • One or more embodiments may provide a video communication device, a video communication method, and a video communication mediating method, which are capable of preventing a user who is making a video communication from being exposed to an inappropriate video or sound from the other party.
  • One or more embodiments may provide a video communication device, a video communication method, and a video communication mediating method, which are capable of preventing a user from transmitting an inappropriate image or sound.
  • One or more embodiments may provide a video communication device, a video communication method, and a video communication mediating method, which are capable of preventing a user who is making a video communication from feeling sexual shame or discomfort caused by an image or sound from the other party.
  • One or more embodiments may provide a video communication device, a video communication method, and a video communication mediating method, which are capable of inducing a healthy video communication between users.
  • a video communication method may comprise establishing a first video communication session with a first terminal, consecutively receiving images or sounds from the first terminal through the first video communication session, examining at least a portion of images or sounds received in a period of recent specific time among the received images or sounds, and blocking an output of at least a portion of the received images or sounds or reporting a user of the first terminal as an abusive user to a first server according to a result of the examining.
  • the examining may comprise detecting a specific portion from the at least a portion of images or sounds received in a period of recent specific time.
  • the detecting the specific portion may comprise detecting a portion corresponding to a specific facial expression, body language, or body part from the at least a portion of images or sounds received in a period of recent specific time.
  • the detecting the portion corresponding to a specific facial expression, body language, or body part may comprise detecting a portion corresponding at least one of an obscene gesture, an insult gesture, a Vietnamese gesture, sexual content, nudity, genitalia, sexual activity, unsanitary content, excreta, excretion, antisocial content, antihuman content, illegal conduct, criminal conduct, hate crime, violent conduct, maltreatment conduct, self-harmful conduct, a weapon, a drug, an antisocial symbol, hateful content, threatening content, phobia-inducing content, or blood from the at least a portion of images or sounds received in a period of recent specific time.
  • the examining may comprise detecting a portion corresponding to a specific voice or language from the at least a portion of images or sounds received in a period of recent specific time.
  • the detecting a portion corresponding to a specific voice or language may comprise detecting a portion corresponding to a language related to at least one of a swear word, insult, racial discrimination, defamation, obscenity, sexual harassment, antisocial conduct, antihuman conduct, terrorism, hate speech, violence, maltreatment, a drug, or threatening.
  • the video communication method may comprise repeating additionally examining at least one time after the examining and before the blocking or the reporting, wherein the blocking or the reporting comprises blocking an output of the at least a portion of the received images or sounds or reporting a user of the first terminal when a cumulative number of times in which the specific portion is detected is a first number or more or the number of times in which the specific portion is consecutively detected a second number of times or more after establishing the first video communication session.
  • the video communication method may comprise repeating the examining to the blocking or the reporting with regular or irregular cycles.
  • the video communication method may comprise consecutively outputting the received images or sounds after the receiving.
  • the video communication method may comprise preparing information on correlation between at least one feature included in an image or sound of an arbitrary user and information on whether the arbitrary user is an abusive user before the establishing the first video communication session, wherein the examining comprises calculating whether the user of the first terminal is an abusive user using the prepared information on correlation and the at least a portion of images or sounds received in a period of recent specific time.
  • the preparing the information on correlation may comprise preparing a machine learning model
  • the examining may comprise predicting whether the user of the first terminal is an abusive user using the prepared machine learning model and the at least a portion of images or sounds received in a period of recent specific time.
  • the preparing the machine learning model may comprise receiving the machine learning model from a second server.
  • the preparing the machine learning model may comprise preparing at least one image or sound corresponding to a non-abusive user or abusive user and training the machine learning model using the at least one image or sound and information on whether each of the at least one image or sound corresponds to the abusive user.
  • the preparing the at least one image or sound may comprise establishing at least one second video communication session with at least one terminal of the non-abusive user or at least one terminal of the abusive user and collecting at least a portion of images or sounds transmitted or received through the at least one second video communication session.
  • a video communication method may comprise establishing a video communication session between a first terminal and a second terminal, consecutively capturing images of a surrounding area of the first terminal or consecutively recording sounds from the surrounding area, examining at least a portion of images captured or sounds recorded in a period of recent specific time among the captured images or the recorded sounds by the first terminal, and according to a result of the examining, blocking transmission of at least a portion through the video communication session among the captured images or the recorded sounds by the first terminal or reporting a user of the first terminal as an abusive user to a first server.
  • the video communication method may comprise repeating, by the first terminal, additionally examining at least one time after the examining and before the blocking or the reporting, wherein the blocking or the reporting may comprise blocking the transmission of the at least a portion of the captured images or recorded sounds or reporting the user of the first terminal when a cumulative number of times in which the specific portion is detected is a first number or more or the number of times in which the specific portion is consecutively detected a second number of times or more after establishing the video communication session.
  • the video communication method may comprise receiving, by the first terminal, a machine learning model from a second server, which corresponds to correlation between at least one feature included in an image or sound of an arbitrary user and information on whether the arbitrary user is an abusive user, before the establishing the video communication session, wherein the examining comprises predicting whether the user of the first terminal is an abusive user using the received machine learning model and the at least a portion of images captured or sounds recorded in a period of recent specific time.
  • the video communication method may comprise, before the establishing the video communication, preparing, by the first terminal, at least one image or sound corresponding to a non-abusive user or abusive user and training, by the first terminal, the machine learning model using the trained machine learning model and the at least a portion of images captured or sounds recorded in a period of recent specific time, wherein the examining may comprise predicting, by the first terminal, whether the user of the first terminal is an abusive user using the trained machine learning model and the at least a portion of images captured or sounds recorded in a period of recent specific time.
  • a video communication mediating method may comprise receiving, by a server, a mediation request from a plurality of terminals, mediating, by the server, a video communication connection between a first terminal and a second terminal among the plurality of terminals, receiving, by the server, a report from the first terminal, which indicates that a user of the first terminal or a user of the second terminal is an abusive user as the result of the first terminal's examination on an image of a surrounding area of the first terminal, captured by the first terminal, a sound from the surrounding area of the first terminal, recorded by the first terminal, an image received by the first terminal from the second terminal, or a sound received by the first terminal from the second terminal, mediating, by the server, a video communication connection between a terminal of the reported user and a terminal of a monitoring agent in response to the report, receiving, by the server, a result of verification on the reported user from the terminal of the monitoring agent, and rejecting, by the server, an additional mediation request from the reported user according to the received result of verification
  • a video communication device may comprise a communication interface configured to establish a video communication session with a terminal and consecutively receive images or sounds from the terminal through the video communication session, an output interface configured to consecutively output the received images or sounds, and a controller configured to examine at least a portion of images or sounds received in a period of recent specific time, wherein according to a result of examination, the controller may control the output interface to block an output of at least a portion of the received images or sounds or the communication interface to report a user of the terminal as an abusive user to a server.
  • a video communication device may comprise an input interface configured to consecutively capture images of a surrounding area thereof or consecutively record sounds from the surrounding area, a communication interface configured to establish a video communication session with a terminal and consecutively transmit the captured images or recorded sounds to the terminal through the video communication session, and a controller configured to examine at least a portion of images captured or sounds recorded in a period of recent specific time among the captured images or recorded sounds, wherein according to a result of the examination, the communication interface may block transmission of at least a portion of the captured images or recorded sounds or report a user of the video communication device as an abusive user to a server.
  • One or more embodiments can prevent a user who is making a video communication from being exposed to an inappropriate image or sound from the other party.
  • One or more embodiments can prevent a user from transmitting an inappropriate image or sound.
  • One or more embodiments can prevent a user who is making a video communication from feeling sexual shame or discomfort caused by an image or sound from the other party.
  • One or more embodiments can induce a healthy video communication between users.
  • FIG. 1 is a network configuration diagram illustrating an environment in which a video communication device according to an embodiment operates
  • FIG. 2 is a block diagram illustrating a configuration of a video communication device according to an embodiment
  • FIG. 3 is a flowchart illustrating a process of performing a video communication method according to an embodiment
  • FIG. 4 is a flowchart illustrating a process of performing operation of preparing information on correlation according to an embodiment
  • FIG. 5 is a flowchart illustrating a process of performing operation of preparing at least one image or sound according to an embodiment
  • FIG. 6 is a front view illustrating a second video communication device in which an image of a user of the second video communication device and an image of the other party are displayed during a video communication according to an embodiment
  • FIG. 7 is a front view illustrating a second video communication device in which an image of a user of the second video communication device and an image of the other party are displayed during a video communication according to another embodiment
  • FIG. 8 is a front view illustrating a first video communication device or a second video communication device or a in which an image of a user of the first video communication device and an image of a user of the second video communication device are displayed during a video communication according to still another embodiment;
  • FIG. 9 is a flowchart illustrating a process of performing a video communication method according to another embodiment.
  • FIG. 10 is a flowchart illustrating a process of performing a video communication mediating method according to an embodiment.
  • first may be used to describe various components, these components should not be limited by these terms. These terms may be only used to distinguish one component from another. Therefore, a first component used herein may be a second component within the technical scope of the present disclosure.
  • video communication devices 200 , 210 , and 300 a video communication method, video communication medicating method, according to embodiments, will be described in detail with reference to FIGS. 1 to 10 .
  • FIG. 1 is a network configuration diagram illustrating an environment in which the video communication devices 200 , 210 , and 300 according to an embodiment operate.
  • the environment in which the video communication devices 200 , 210 , and 300 operate may include a server 100 and at least two of the video communication devices 200 , 210 , and 300 connected to each other.
  • the video communication devices 200 , 210 , and 300 are only illustrated in FIG. 1 as a matter of convenience, an additional video communication device other than the video communication devices 200 , 210 , and 300 may be included.
  • the server 100 may be connected to a communication network.
  • the server 100 may be connected to an external device via the communication network.
  • the server 100 may transmit data to other devices connected to each other or receive data from other devices.
  • the communication network connected to the server 100 may include a wired communication network, a wireless communication network, or a composite communication network.
  • the communication network may include a mobile communication network such as 3G, Long Term Evolution (LTE), or LTE-A.
  • the communication network may include a wired or wireless communication network such as Wi-Fi, Universal Mobile Telephone System/General Packet Radio Service (UMTS/GPRS), or Ethernet.
  • the communication network may include a short-range communication network such as Magnetic Secure Transmission (MST), Radio Frequency Identification (RFID), Near Field Communication (NFC), ZigBee, Z-Wave, Bluetooth, Bluetooth Low Energy (BLE), or InfraRed (IR) communication.
  • the communication network may include Local Area Network (LAN), Metropolitan Area Network (MAN), or Wide Area Network (WAN).
  • the server 100 may be connected to each of the video communication devices 200 , 210 , and 300 through a communication network.
  • the server 100 may transceive data with the first video communication device 200 through a communication network.
  • the server 100 may transceive data with the second video communication device 210 through a communication network.
  • the server 100 may transceive data with the third video communication device 300 through a communication network.
  • the server 100 may receive data from the video communication devices 200 , 210 , and 300 .
  • the server 100 may operate calculations using data received from the video all devices 200 , 210 , and 300 .
  • the server 100 may transmit the calculation result to the video communication devices 200 , 210 , and 300 .
  • the server 100 may receive mediation requests of video communications from a plurality of video communication devices 200 , 210 , and 300 .
  • the server 100 may select at least two video communication devices among the plurality of video communication devices 200 , 210 , and 300 that transmit the mediation requests.
  • the server 100 may select the first video communication device 200 and the second video communication device 210 as the at least two video communication devices 200 , 210 , and 300 .
  • the server 100 may mediate a connection of a video communication between the selected at least two video communication devices 200 , 210 , and 300 .
  • the server 100 may transmit the connection information on the first video communication device 200 to the second video communication device 210 .
  • the server 100 may transmit the connection information on the second video communication device 210 to the first video communication device 200 .
  • the connection information on the first video communication device 200 may include, for example, an IP address and a port number of the first video communication device 200 .
  • the second video communication device 210 receiving the connection information on the first video communication device 200 may attempt to connect to the first video communication device 200 using the received connection information.
  • the connection information on the second video communication device 210 may include, for example, an IP address and a port number of the second video communication device 210 .
  • the first video communication device 200 receiving the connection information on the second video communication device 210 may attempt to connect to the second video communication device 210 using the received connection information.
  • Succeeding in an attempt to connect the first video communication device 200 to the second video communication device 210 or an attempt to connect the second video communication device 210 to the first video communication device 200 may enable establishment of a video communication session between the first video communication device 200 and the second video communication device 210 .
  • the first video communication device 200 may transmit an image or sound to the second video communication device 210 through the video communication session.
  • the first video communication device 200 may encode an image or sound into a digital signal and transmit the encoded result to the second video communication device 210 .
  • the first video communication device 200 may receive an image or sound from the second video communication device 210 through the video communication session.
  • the first video communication device 200 may receive an image or sound encoded in a digital signal and may decode the received encoded image or sound.
  • the second video communication device 210 may transmit an image or sound to the first video communication device 200 through the video communication session. Further, the second video communication device 210 may receive an image or sound from the first video communication device 200 through the video communication session. Accordingly, a user of the first video communication device 200 and another user of the second video communication device 210 may make a video communication with each other.
  • the video communication devices 200 , 210 and 300 may be terminals.
  • the video communication devices 200 , 210 and 300 may be, for example, desktop computers, laptop computers, smart phones, smart tablets, smart watches, mobile terminals, digital cameras, wearable devices, or portable electronic devices.
  • the video communication devices 200 , 210 , and 300 may execute programs or applications.
  • Each of the video communication devices 200 , 210 and 300 may be the same kind of device each other or may not be the same kind of device each other.
  • the first video communication device 200 and a first terminal 200 may be interchangeably used as a term to refer to the first video communication device 200 for convenience of description.
  • the second video communication device 210 and a second terminal 210 will be interchangeably used as a term to refer to the second video communication device 210 and a third video communication device 300
  • the third terminal 300 will be interchangeably used as a term to refer to the third video communication device 300 .
  • FIG. 2 is a block diagram showing a configuration of the first terminal 200 according to an embodiment.
  • the configuration of the second terminal 210 or the configuration of the third terminal 300 may be the same as or different from the configuration of the first terminal 200 .
  • the first terminal 200 according to the embodiment may include an input interface 201 , a storage 202 , a communication interface 203 , an output interface 204 , and a controller 205 .
  • the input interface 201 may receive input from the outside.
  • the input interface 201 may receive input from a user of the first terminal 200 . Further, the input interface 201 may receive a signal from an external device as the input.
  • the input interface 201 may include, for example, a microphone, a camera, a keyboard, a mouse, a trackball, a touch screen, a button, a switch, a sensor, a network interface, or other input devices.
  • the input interface 201 may receive light from outside through a camera included in the input interface 201 .
  • the input interface 201 may receive light emitted from an external object or light reflected from an external object.
  • the input interface 201 may consecutively capture images of the surrounding area of the first terminal 200 through the camera.
  • the input interface 201 may capture an image of a user of the first terminal 200 .
  • the input interface 201 may receive a sound from outside through a microphone included in the input interface 201 .
  • the input interface 201 may receive a sound from the user of the first terminal 200 . Further, the input interface 201 may receive a sound from an external sound reproducing device.
  • the input interface 201 may consecutively record sounds from the surrounding area of the first terminal 200 through the microphone.
  • the input interface 201 may record a voice of the user of the first terminal 200 .
  • the storage 202 may store data.
  • the storage 202 may store the input received by the input interface 201 .
  • the storage 202 may store the image captured or the sound recorded by the input interface 201 .
  • the storage 202 may store data to be used by the controller 205 to perform an arithmetic operation.
  • the storage 202 may store the result of the arithmetic operation performed by the controller 205 .
  • the storage 202 may store the image or voice encoded by the controller 205 .
  • the storage 202 may store data to be transmitted to the outside through the communication interface 203 or may store data received from the outside through the communication interface 203 .
  • the storage 202 may include a volatile memory or a non-volatile memory.
  • the storage 202 may include, for example, flash memory, read-only memory (ROM), random access memory (RAM), electrically erasable read-only memory (EEROM), erasable programmable read-only memory (EPROM), electrically erasable and programmable read-only memory (EEPROM), hard disk drive (HDD), register, or the like.
  • the storage 202 may include, for example, a file system, a database, or an embedded database.
  • the communication interface 203 may transmit data to the exterior or receive data from the exterior.
  • the communication interface 203 may transmit data to the other video communication devices 210 and 300 or the server 100 .
  • the communication interface 203 may receive data from the other video communication devices 210 and 300 or the server 100 .
  • the communication interface 203 may transmit the result of the arithmetic operation performed by the controller 205 to the exterior. Further, the communication interface 203 may transmit data stored in the storage 202 to the exterior.
  • the communication interface 203 may transmit the image captured or the sound recorded by the input interface 201 to the exterior.
  • Data to be transmitted by the communication interface 203 or data received by the communication interface 203 may be stored in the storage 202 .
  • the communication interface 203 may transmit an encoded image or voice stored in the storage 202 to the server 100 or the other video communication devices 210 and 300 .
  • the communication interface 203 may receive an image or a voice encoded by the other video communication devices 210 and 300 from the other video communication devices 210 and 300 .
  • the communication interface 203 may include, for example, a long-range network interface such as a 3G module, an LTE module, an LTE-A module, a Wi-Fi module, a WiGig module, an Ultra Wide Band (UWB) module, or a LAN card. Further, the communication interface 203 may include, for example, a short-range network interface such as an MST module, a Bluetooth module, an NFC module, an RFID module, a ZigBee module, a Z-Wave module, or an IR module. Further, the communication interface 203 may include other network interfaces.
  • a long-range network interface such as a 3G module, an LTE module, an LTE-A module, a Wi-Fi module, a WiGig module, an Ultra Wide Band (UWB) module, or a LAN card.
  • the communication interface 203 may include, for example, a short-range network interface such as an MST module, a Bluetooth module, an NFC module, an RFID module, a ZigBee
  • the output interface 204 may output a signal to the exterior.
  • the output interface 204 may, for example, display a screen, reproduce a sound, or output a vibration.
  • the output interface 204 may include a display, a speaker, a vibrator, an oscillator, or other output devices.
  • the output interface 204 may display a screen.
  • the controller 205 may control the output interface 204 to display a screen.
  • the output interface 204 may display a user interface.
  • the output interface 204 may display another screen in response to an input of a user.
  • the output interface 204 may display data.
  • the output interface 204 may display the result of the arithmetic operation performed by the controller 205 .
  • the output interface 204 may display data stored in the storage 202 .
  • the output interface 204 may display data received by the communication interface 204 .
  • the output interface 204 may display an image of a user of the first terminal 200 captured by the input interface 201 or images of the other users captured by the other video communication devices 210 and 300 . Further, the output interface 204 may reproduce the voice of the user of the first terminal 200 recorded by the input interface 201 or the voice of the other users recorded by the other video communication devices 210 and 300 .
  • the output interface 204 may include, for example, a flat-panel display device such as a liquid crystal display (LCD), an organic light-emitting diode (OLED), or a plasma display panel (PDP).
  • the output interface 204 may include a curved display or a flexible display.
  • the output interface 204 may include a touch screen.
  • the controller 205 may control an operation of the first terminal 200 .
  • the controller 205 may be connected to the components included in the first terminal 200 .
  • the controller 205 may control each operation of the components included in the first terminal 200
  • the controller 205 may process a signal.
  • the controller 205 may process a signal received by the input interface 201 .
  • the controller 205 may process the images of the users, captured by the camera included in the input interface 201 .
  • the controller 205 may process a sound and a voice of a user, recorded by the microphone included in the input interface 201 .
  • the controller 205 may process the signal received by the communication interface 203 .
  • the controller 205 may process a request of the user, received through the communication interface 203 .
  • the controller 205 may process images or voices of other users received through the communication interface 203 .
  • the controller 205 may control an operation of the first terminal 200 in response to a signal received by the input interface 201 .
  • the controller 205 may control operation of the first terminal 200 in response to a request of a user received by the input interface 201 .
  • the controller 205 may perform an arithmetic operation.
  • the controller 205 may perform an arithmetic operation according to a signal received by the input interface 201 .
  • the controller 205 may perform an arithmetic operation using a signal received by the input interface 201 or data stored in the storage 202 .
  • the controller 205 may perform image-processing using an image captured by the camera included in the input interface 201 .
  • the controller 205 may encode an image captured by the camera included in the input interface 201 .
  • the controller 205 may encode a sound or a voice recorded by the microphone included in the input interface 201 .
  • the controller 205 may encode a sound or a voice recorded by the camera included in the input interface 201 .
  • the controller 205 may decode images or voices of other users, received by the communication interface 203 .
  • the controller 205 may perform post-processing on the decoded image or voice.
  • the controller 205 may perform image-processing on images of other users, received by the communication interface 203 .
  • the controller 205 may control such that the result of the arithmetic operation is stored in the storage 202 .
  • the controller 205 may control such that the result of the arithmetic operation is output by the output interface 105 .
  • the controller 204 may control that the result of the arithmetic operation is transmitted to the other video communication devices 210 and 300 by the communication interface 203 .
  • the controller 205 may include, for example, a central processing unit (CPU), a graphic processing unit (GPU), a micro controller unit (MCU), or a microprocessor.
  • CPU central processing unit
  • GPU graphic processing unit
  • MCU micro controller unit
  • a type of the second terminal 210 may be same as or different from a type of the first terminal 200 .
  • the second terminal 210 may execute a program or an application.
  • the second terminal 210 may be connected to a communication network.
  • the second terminal 210 may be connected to an external device with each other via the communication network.
  • the second terminal 210 may transmit data to the connected external device or receive data from the connected external device.
  • the second terminal 210 may be connected to the first terminal 200 with each other by mediation of the server 100 .
  • the second terminal 210 may transmit data to the first terminal 200 or receive data from the first terminal 200 .
  • the first terminal 200 and the second terminal 210 which are connected to each other, may transceive a message, a file, data, an image, a video, a sound, or a voice, and so on with each other.
  • the second terminal 210 may establish a voice communication session or a video communication session with the first terminal 200 .
  • the video communication session may be established using, for example, transmission control protocol (TCP), user datagram protocol (UDP), or web real-time communication (WebRTC).
  • TCP transmission control protocol
  • UDP user datagram protocol
  • WebRTC web real-time communication
  • the user of the first terminal 200 and the user of the second terminal 210 may make a voice communication or a video communication with each other.
  • the video communication session may directly be established between the first terminal 200 and the second terminal 210 .
  • the video communication session may directly be established between the first terminal 200 and the second terminal 210 via at least one other device.
  • the video communication session may include a session established between the first terminal 200 and the server 100 and a session established between the server 100 and the second terminal 210 .
  • the first terminal 200 and the second terminal 210 may directly exchange data with each other. Further, the first terminal 200 and the second terminal 210 may exchange data with each other via the server 100 or other mediating devices.
  • a type of the third terminal 300 may be identical to or different from a type of the first terminal 200 .
  • the third terminal 300 may execute a program or an application.
  • the third terminal 300 may be connected to a communication network.
  • the third terminal 300 may be connected to an external device with each other via the communication network.
  • the third terminal 300 may transmit data to the connected external device or receive data from the connected external device.
  • FIG. 3 is a flowchart illustrating a process of performing a video communication method by the second terminal 210 , according to an embodiment.
  • operation S 100 may be performed to prepare information on correlation between at least one feature included in an image or sound of an arbitrary user and information on whether the arbitrary user is an abusive user.
  • the storage 202 of the second terminal 210 may prepare information on the correlation.
  • An arbitrary image or sound of the arbitrary user may include at least one feature.
  • the feature included in the arbitrary image may be a specific shape, pattern, color, or a combination thereof included in the arbitrary image.
  • the feature may be, for example, a specific facial expression, a specific body language, a specific shape, pattern, or color corresponding to a body part, or combination thereof.
  • Each feature included in the arbitrary image may correspond to a non-abusive element or an abusive element.
  • the feature may correspond to at least one abusive element of an obscene gesture, an insult gesture, a Vietnamese gesture, sexual content, nudity, genitalia, sexual activity, unsanitary content, excreta, excretion, antisocial content, antihuman content, illegal conduct, criminal conduct, hate crime, violent conduct, maltreatment conduct, self-harmful conduct, a weapon, a drug, an antisocial symbol, hateful content, threatening content, phobia-inducing content, or blood.
  • the feature included in the arbitrary sound may be a specific waveform, a pattern, or a combination thereof included in the arbitrary sound.
  • the feature may be, for example, a specific waveform, pattern, or a combination thereof corresponding to a specific voice or language.
  • Each feature included in the arbitrary sound may correspond to a non-abusive element an abusive element.
  • the feature may correspond to at least one abusive element of groaning, gasping sound, screaming, or shrieking.
  • the feature may correspond to at least one abusive element among a crash, a gun-shooting sound, a vomit sound, a hit sound, a whip sound, a knife sound, a swing sound, a stab sound, a sound having a pitch equal to or greater than an arbitrary threshold value, or a sound having a loudness equal to and greater than a specific threshold value.
  • the feature may correspond to a language related to at least one of a swear word, insult, racial discrimination, defamation, obscenity, sexual harassment, antisocial conduct, antihuman conduct, terrorism, hate speech, violence, maltreatment, a drug, or threatening as the abusive element.
  • the storage 202 of the second terminal 210 may prepare at least one of a list, a table, a database, or a machine learning model including information on the correlation.
  • the storage 202 of the second terminal 210 may prepare a list including at least one item.
  • the storage 202 of the second terminal 210 may prepare a database corresponding to the list.
  • Each item included in the list may include information on one feature.
  • Each feature corresponding to each item may correspond to an abusive element. Therefore, when any feature among features corresponding to the item included in the list is identical to the feature included in the arbitrary image or sound, it may be determined that the arbitrary image or sound includes an abusive element. When it is determined that the arbitrary image or sound includes an abusive element, the arbitrary user may be determined as an abusive user.
  • the communication interface 203 of the second terminal 210 may receive the list or a database corresponding to the list from the server 100 or a separate server.
  • the storage 202 of the second terminal 210 may store the received list or database.
  • the controller 205 of the second terminal 210 may generate the list or database by itself, and the storage 202 of the second terminal 210 may store the generated list or database.
  • the storage 202 of the second terminal 210 may prepare a table including at least one record.
  • the storage 202 of the second terminal 210 may prepare a database corresponding to the table.
  • Each record included in the table may include information on a feature and information on whether the feature corresponds to a non-abusive element or an abusive element. Therefore, when the information indicating that a record corresponding to a feature included in an arbitrary image or sound is present among the records included in the table and information indicating that the feature corresponds to an abusive element is included in the record, the arbitrary image or sound may be determined to contain an abusive element. When the arbitrary image or sound is determined to include an abusive element, the arbitrary user may be determined as an abusive user.
  • the communication interface 203 of the second terminal 210 may receive the table or a database corresponding to the table from the server 100 or a separate server.
  • the storage 202 of the second terminal 210 may store the received table or database.
  • the controller 205 of the second terminal 210 may generate the table or database by itself, and the storage 202 of the second terminal 210 may store the generated table or database.
  • the storage 202 of the second terminal 210 may prepare a machine learning model corresponding to the correlation.
  • the machine learning model may be a neural network including at least two layers.
  • the machine learning model may include an input layer and an output layer.
  • the machine learning model may further include at least one hidden layer.
  • the machine learning model may be a classification model trained using at least one image or sound and information on whether the at least one image or sound corresponds to a non-abusive element or an abusive element.
  • the machine learning model may use at least one feature included in an arbitrary image or sound as an input value.
  • the machine learning model may predict whether the at least one feature corresponds to a non-abusive element or an abusive element. When it is predicted that the at least one feature corresponds to an abusive element, the arbitrary user may be determined as an abusive user.
  • the communication interface 203 of the second terminal 210 may receive the machine learning model from the server 100 or a separate server.
  • the storage 202 of the second terminal 210 may store the received machine learning model.
  • the controller 205 of the second terminal 210 may train the machine learning model by itself and the storage 202 of the second terminal 210 may store the trained machine learning model.
  • FIG. 4 is a flowchart of operation S 100 of preparing information on correlation according to an embodiment.
  • operation S 101 may be performed to prepare at least one image or sound corresponding to a non-abusive user or an abusive user.
  • the storage 202 of the second terminal 210 may prepare at least one image or sound corresponding to a non-abusive user or an abusive user.
  • FIG. 5 is a flowchart of operation S 101 of preparing at least one image or sound according to an embodiment.
  • operation S 102 may be performed to establish at least one second video communication session between the terminals of at least one non-abusive user terminal or at least one abusive user terminal.
  • the communication interface 203 of the second terminal 210 may establish the at least one second video communication session.
  • the communication interface 203 of the second terminal 210 may transmit an image or sound recorded by the input interface 201 to the other video communication devices 200 and 300 through the second video communication session.
  • the output interface 204 of the second terminal 210 may output the captured image or the recorded sound.
  • the communication interface 203 of the second terminal 210 may receive an image or sound recorded by the other video communication devices 200 and 300 from the other video communication devices 200 and 300 through the second video communication session.
  • the output interface 204 of the second terminal 210 may output the received image or sound.
  • operation S 103 may be performed to collect at least one portion of the image or sound transmitted or received through the at least one second video communication session.
  • the input interface 201 of the second terminal 210 receives the feedback indicating that the output image or sound does not contain an abusive element from the user of the second terminal 210 .
  • the storage 202 of the second terminal 210 may collect the image or sound output when the feedback is received, as an image or sound corresponding to s non-abusive user.
  • the input interface 201 of the second terminal 210 may receive the feedback indicating that the output image or sound contains an abusive element from the user of the second terminal 210 .
  • the user of the second terminal 210 may touch a report button displayed on a touch screen.
  • the storage 202 of the second terminal 210 may collect the image or sound output when the report button was touched, as an image or sound corresponding to an abusive user.
  • operation S 104 may be performed to train the machine learning model using the at least one image or sound and information on whether each of the at least one image or sound corresponds to an abusive user.
  • the controller 205 of the second terminal 210 may train the machine learning model using information on the at least one image or sound and information on whether each of the at least one image or sound corresponds to an abusive user, collected by the storage 202
  • the controller 205 may train the machine learning model using a deep learning algorithm.
  • the machine learning model may include at least one of deep neural network (DNN), convolutional neural network (CNN), recurrent neural network (RNN), region-based convolutional neural networks (R-CNN), restricted Boltzmann machine (RBM), deep belief network (DBN), or deep Q-networks.
  • DNN deep neural network
  • CNN convolutional neural network
  • RNN recurrent neural network
  • R-CNN region-based convolutional neural networks
  • R-CNN restricted Boltzmann machine
  • DBN deep belief network
  • the machine learning model may be convolutional neural network (CNN).
  • CNN convolutional neural network
  • the machine learning model may include AlexNet, ZFNet, GoogLeNet, VGGNet, ResNet, Inception-ResNet, Inception-v2, Inception-v3, or Inception-v4.
  • the controller 205 may convert each of the at least one image into an RGB format image having a width of 60 pixels and a height of 80 pixels using a convolutional layer.
  • the controller 205 may train the machine learning model including Inception-v4 using each of the transformed images.
  • the trained model may include total 4,330,627 variables.
  • the size of each variable may be 4 bytes.
  • the size of the trained model may be at least 17,322,508 bytes.
  • operation S 110 may be performed to establish the first video communication session with the first terminal 200 .
  • the communication interface 203 of the second terminal 210 may establish the first video communication session.
  • operation 120 may be performed to consecutively receive images or sounds from the first terminal 200 through the first video communication session.
  • the communication interface 203 of the second terminal 210 may consecutively receive images or sounds through the first video communication session.
  • the communication interface 203 may receive a data stream including an encoded image or sound from the first terminal 200 .
  • the controller 205 of the second terminal 210 may decode the received encoded image or sound.
  • operation S 130 may be performed to consecutively output the received images or sounds.
  • the display included in the output interface 204 of the second terminal 210 may consecutively show the received image.
  • the output interface 204 may display images of the surrounding area of the second terminal 210 , being consecutively captured by the camera included in the input interface 201 of the second terminal 210 together with the received image.
  • FIGS. 6 to 8 is a front view of the second terminal 210 displaying an image of a user of the second terminal 210 and an image of the other party during a video communication according to an embodiment. 6 to 8 .
  • the output interface 204 of the second terminal 210 may display images captured by the input interface 201 on a rectangular area positioned at the lower left of the display. Further, the output interface 204 of the second terminal 210 may display the images received by the communication interface 203 on the remaining area of the entire area of the display except for the rectangular area.
  • the speaker included in the output interface 204 of the second terminal 210 may consecutively reproduce the received sounds.
  • the output interface 204 may reproduce sounds in the surrounding area of the second terminal 210 , being consecutively recorded by the microphone included in the input interface 201 of the second terminal 210 together with the received sounds.
  • consecutively outputting the received images or sounds may be excluded.
  • the control interface 205 of the second terminal 210 may control the output interface 204 according to a result of the examination so that the received images or sounds is consecutively output.
  • operation S 140 may be performed to examine the at least a portion of the images or sounds received in a period of recent specific time.
  • the controller 205 of the second terminal 210 may examine at least a portion of the received images or sounds in a period of recent specific time.
  • the specific time may be, for example, one second.
  • the controller 205 may examine all of the images or sounds received in a period of the recent specific time. When the images are being received at 30 frames per second (FPS), the controller 205 may examine all 30 images received for recent one second.
  • FPS frames per second
  • the controller 205 may select a portion of the images or sounds received in a period of the recent specific time.
  • the controller 205 may check the selected portion of the images or sounds. For example, the controller 205 may arbitrarily select one image among the 30 images received for recent one second.
  • the controller 205 may examine the selected single image.
  • the controller 205 may examine an image of the user of the first terminal 200 displayed in the remaining area except for the rectangular area positioned at the lower left of the display in FIGS. 6 to 8 .
  • the controller 205 may detect a specific portion from at least a portion of the image received in a period of the recent specific time.
  • the controller 205 may detect a portion corresponding to a specific facial expression, body language, or body part as the specific portion.
  • the controller 205 may detect a portion corresponding to at least one of an obscene gesture, an insult gesture, a Vietnamese gesture, sexual content, nudity, genitalia, sexual activity, unsanitary content, excreta, excretion, antisocial content, antihuman content, illegal conduct, criminal conduct, hate crime, violent conduct, maltreatment conduct, self-harmful conduct, a weapon, a drug, an antisocial symbol, hateful content, threatening content, phobia-inducing content, or blood.
  • the controller 205 may detect a specific portion from at least one portion of the images received in a period of the recent specific time.
  • the controller 205 may detect a portion corresponding to a specific voice or language.
  • the controller 205 may detect a portion corresponding to at least one of an obscene gesture, an insult gesture, a Vietnamese gesture, defamation, obscenity, sexual harassment, antisocial conduct, antihuman conduct, terrorism, hate speech, violence, maltreatment, a drug, or threatening.
  • the controller 205 of the second terminal 210 may calculate whether the user of the first terminal 200 is an abusive user by detecting the specific portion. When the specific portion is detected from the image or sound which is the examination object, the controller 205 may determine that the user of the first terminal 200 is an abusive user. When the specific portion is not detected from the image or sound which is the examination object, the controller 205 may determine that the user of the first terminal 200 is a non-abusive user.
  • the controller 205 may calculate whether the user of the first terminal 200 is an abusive user using information on the prepared correlation and the at least a portion of the images or sounds received in a period of the recent specific time.
  • the controller 205 may extract a feature included in the at least a portion of the images or sounds received in a period of the recent specific time.
  • the controller 205 may calculate whether the user of the first terminal 200 is an abusive user using the extracted feature and the information on the prepared correlation.
  • the information on the prepared correlation may include the list.
  • the controller 205 may compare a feature corresponding to at least one item included in the list and the extracted feature with each other. When one among features corresponding to the at least one item is the same as the extracted feature, the controller 205 may determine that the image or sound to be examined includes an abusive element. Further, the controller 205 may determine that the user of the first terminal 200 is an abusive user.
  • the information on the prepared correlation may include the table.
  • the controller 205 may compare the extracted feature with the feature corresponding to at least one record included in the table. If information indicating that a record corresponding to the extracted feature is included in the records included in the table and that the extracted feature corresponds to an abusive element is included in the record, it may be determined that the image or sound includes an abusive element. Further, the controller 205 may determine that the user of the first terminal 200 is an abusive user.
  • the information on the prepared correlation may include the machine learning model.
  • the controller 205 may predict whether the user of the first terminal 200 is an abusive user using the prepared machine learning model and the at least a portion of the images or sounds received in a period of the recent specific time.
  • the controller 205 may convert an image subjected to the examination into an RGB format image having a width of 60 pixels and a height of 80 pixels using a convolutional layer.
  • the controller 205 may use the converted image as an input value of the machine learning model.
  • the controller 205 may run the machine learning model.
  • the result value may be calculated from the output layer of the machine learning model.
  • the result value may be, for example, a vector including three numerical values.
  • the three numerical values respectively, may correspond to the probability that the image to be examined does not include an abusive element, the probability that the image to be examined includes an abusive element, and the probability that the image to be examined includes the appearance of a user of the first terminal 200 , who has taken off his or her clothes that the user wear on the upper part of the body.
  • the sum of the three numerical values may be one.
  • the controller 205 may determine that the user of the first terminal 200 is a non-abusive user. For example, in FIG. 6 , when an image of a user of the first terminal 200 displayed in an area other than the rectangular area positioned at the lower left of the display is examined, the greatest value may be the first numeric value among the three numerical values. The controller 205 may determine that the user of the first terminal 200 is a non-abusive user.
  • the controller 205 may determine that the user of the first terminal 200 is an abusive user. For example, in FIG. 7 , when an image of a user of the first terminal 200 displayed in an area other than the rectangular area positioned at the lower left of the display is examined, the greatest value may be the second numeric value among the three numerical values. The controller 205 may determine that the user of the first terminal 200 is an abusive user.
  • operation S 150 may be performed to repeat the operation S 140 of examining at least one time.
  • the controller 205 of the second terminal 210 may perform an examination on an image or sound received later than the image or sound that is the object of the latest examination after performing the latest examination.
  • the controller 205 may further perform four examinations after performing the first examination.
  • the controller 205 may examine each of the five images or sounds received at an interval of one second from each other.
  • Each result of the examinations may be accumulated. For example, when a portion corresponding to an abusive element is detected in each of five images, the cumulative number of times may be five.
  • the operation S 150 of repeating may be excluded.
  • operation S 160 may be performed to block the output of at least a portion of the received image or sound or to report the user of the first terminal 200 as an abusive user to the server 100 according to the result of the examination.
  • the controller may control the output interface 204 to consecutively output the received image or sound.
  • the controller 205 may control the output interface 204 to block the at least a portion of the received image or sound.
  • the controller 205 may control the communication interface 203 to report the user of the first terminal 200 as an abusive user to the server 100 .
  • the at least a portion of the received image or sound may be blocked, and the user of the first terminal 200 may be reported as an abusive user to the server 100 .
  • the controller 205 may control the output interface 204 to block the output of the at least a portion or the communication interface 205 to report the user of the first terminal 200 as an abusive user.
  • the first number of times may be, for example, one or five times.
  • the controller 205 may control the output interface 204 to block the output of the at least a portion or the communication interface 203 to report the user of the first terminal 200 as an abusive user.
  • the controller 205 may control the output interface 204 to block the output of the at least a portion or the communication interface 205 to report the user of the first terminal 200 as an abusive user.
  • the second number of times may be, for example, one or five times.
  • the controller 205 may control the output interface 204 to block the output of the at least a portion or the communication interface 203 to report the user of the first terminal 200 as an abusive user.
  • the output is blocked or the user of the first terminal 200 is reported, thereby tolerating intermittent errors of the determination that may occur in the result of the examination. Therefore, more stable operation of the second terminal 210 may further be ensured.
  • operation S 170 may be performed to repeat the operation S 140 of examining or the operation S 160 of blocking or reporting with regular or irregular cycles. For example, after the controller 205 performed the latest operation S 140 of examining or the latest operation S 160 of blocking or reporting, the operation S 140 of examining or operation S 160 of blocking or reporting may be repeated after at least 10 seconds have elapsed. Therefore, it may be consistently examined whether the images or sounds which the first terminal 200 consecutively transmits may include an abusive element.
  • a user in a video communication may be prevented from being exposed to an inappropriate image or sound from the other party. Further, a user in a video communication may be prevented from feeling sexual shame or discomfort. Further, a healthy video communication between users can be induced.
  • FIG. 9 is a flowchart illustrating a process of a video communication method by the first terminal 200 , according to another embodiment.
  • operation S 200 may be performed in which the first terminal 200 receives a machine learning model from the server 100 , which corresponds to correlation between at least one feature included in an image or sound of an arbitrary user and information on whether the arbitrary user is an abusive user.
  • the storage 202 of the first terminal 200 may prepare information on the correlation.
  • the storage 202 of the first terminal 200 may prepare at least one of a list, a table, a database, or a machine learning model including information on the correlation.
  • the storage 202 of the first terminal 200 may prepare a machine learning model corresponding to the correlation.
  • the communication interface 203 of the first terminal 200 may receive the machine learning model from the server 100 or a separate server.
  • the storage 202 of the first terminal 200 may store the received machine learning model.
  • the controller 205 of the first terminal 200 may train the machine learning model by itself, and the storage 202 of the first terminal 200 may store the trained machine learning model.
  • the storage 202 of the first terminal 200 may prepare at least one image or sound corresponding to a non-abusive user or an abusive user.
  • the controller 205 of the first terminal 200 may train the machine learning model using the at least one image or sound and information on whether each of the at least one image or sound corresponds to an abusive user.
  • operation S 210 may be performed to establish a video communication session between the first terminal 200 and the second terminal 210 .
  • the communication interface 203 of the first terminal 200 may establish a video communication session with the second terminal 210 .
  • operation S 220 may be performed in which the first terminal 200 consecutively captures an image of the surrounding area of the first terminal 200 or consecutively records a sound from the surrounding area.
  • the camera included in the input interface 201 of the first terminal 200 may consecutively capture an image of the surrounding area of the first terminal 200 .
  • the microphone included in the input interface 201 of the first terminal 200 may consecutively record the sound from the surrounding area.
  • the storage 202 of the first terminal 200 may store the captured image or the recorded sound.
  • operation S 230 may be performed in which the first terminal 200 consecutively transmits the captured images or recorded sounds to the second terminal 210 through the video communication session.
  • the communication interface 203 of the first terminal 200 may consecutively transmit the captured images or recorded sounds. Further, at the same time, the communication interface 203 may consecutively receive images or sounds from the second terminal 210 .
  • the output interface 204 may display the captured images together with the received images. The output interface 204 may reproduce the recorded sounds together with the received sounds.
  • the operation S 230 of consecutively transmitting the captured images or the recorded sounds may be excluded.
  • the controller 205 of the first terminal 200 may control the communication interface 203 according to the result of the examination to consecutively transmit the captured images or the recorded sounds.
  • operation S 240 may be performed in which the first terminal 200 examines at least a portion of the images captured or sounds recorded in a period of the recent specific time among the captured images or the recorded sounds.
  • the controller 205 of the first terminal 200 may detect a specific portion from the at least a portion of the images captured or sounds recorded in a period of the recent specific time.
  • the specific time may be, for example, one second.
  • the controller 205 may predict whether the user of the first terminal 200 is an abusive user using the prepared machine learning model and the at least a portion of the images captured or sounds recorded in a period of the recent specific time. For example, the controller 205 may convert an image which is the examination object into an RGB format image having a width of 60 pixels and a height of 80 pixels using a convolutional layer. The controller 205 may use the converted image as an input value of the machine learning model. The controller 205 may run the machine learning model.
  • the result value may be calculated from the output layer of the machine learning model.
  • the result value may be, for example, a vector including three numerical values.
  • the three numerical values respectively, may correspond to the probability that the image to be examined does not include an abusive element, the probability that the image to be examined includes an abusive element, and the probability that the image to be examined includes the appearance of a user, who has taken off his or her clothes that the user wear on the upper part of the body, of the first terminal 200 .
  • the sum of the three numerical values may be one.
  • the controller 205 may determine that the user of the first terminal 200 is a non-abusive user. For example, in FIG. 6 , when an image of a user of the first terminal 200 displayed in the rectangular area positioned at the lower left of the display is examined, the greatest value may be the first numeric value among the three numerical values. The controller 205 may determine that the user of the first terminal 200 is a non-abusive user.
  • FIG. 8 is a front view of a first terminal 200 displaying an image of a user of the first terminal 200 and an image of a user of the second terminal 210 during a video communication according to the embodiment.
  • the greatest value may be the third numeric value among the three numerical values.
  • the controller 205 may determine that the user of the first terminal 200 is an abusive user.
  • operation S 250 may be performed in which the first terminal 200 repeats the operation S 240 of examining at least one time.
  • the controller 205 of the first terminal 200 may perform an examination on an image or sound received later than the image or sound that is the object of the latest examination after performing the latest examination.
  • the controller 205 may further perform four examinations after performing the first examination.
  • the controller 205 may examine each of the five images or sounds received at an interval of one second from each other.
  • Each result of the examinations may be accumulated. For example, when a portion corresponding to an abusive element is detected in each of five images, the cumulative number of times may be five.
  • the operation S 250 of repeating may be excluded.
  • operation S 260 may be performed in which the first terminal 200 blocks the transmission of the at least a portion of the captured image or the recorded sound through a video communication session or reports the user of the first terminal 200 as an abusive user to the server 100 according to the result of the examination.
  • the controller may control the communication interface 203 to consecutively transmit the captured images or the recorded sounds.
  • the controller 205 may control the communication interface 203 to block the transmission of the at least a portion of the captured images or the recorded sounds.
  • the controller 205 may control the communication interface 203 to report the user of the first terminal 200 as an abusive user to the server 100 .
  • the transmission of the at least a portion of the captured images or the recorded sounds may be blocked, and the user of the first terminal 200 may be reported as an abusive user to the server 100
  • the controller 205 may control the communication interface 203 to block the transmission of the at least a portion or to report the user of the first terminal 200 as an abusive user.
  • the first number of times may be, for example, one or five times.
  • the controller 205 may control the communication interface 203 to block the output of the at least a portion or to report the user of the first terminal 200 as an abusive user.
  • the controller 205 may control the communication interface 203 to block the transmission of the at least a portion or to report the user of the first terminal 200 as an abusive user.
  • the second number of times may be, for example, one or five times.
  • the controller 205 may control the communication interface 203 to block the transmission of the at least a portion or to report the user of the first terminal 200 as an abusive user.
  • the transmission is blocked or the user of the first terminal 200 is reported, thereby tolerating intermittent errors of the determination that may occur in the result of the examination. Therefore, more stable operation of the first terminal 200 may further be ensured.
  • operation S 270 may be performed that the first terminal 200 repeats the operation S 240 of examining or the operation S 260 of blocking or reporting with regular or irregular cycles.
  • the operation S 140 of examining or the operation S 160 of blocking or reporting may be repeated after at least 10 seconds have elapsed. Therefore, it may be consistently examined whether the images or sounds which the first terminal 200 consecutively transmits may include an abusive element.
  • a user in a video communication may be prevented from being exposed to an inappropriate image or sound from the other party. Further, a healthy video communication between users can be induced.
  • FIG. 10 is a flowchart of illustrating a video communication mediating method according to an embodiment. Referring to FIG. 10 , operation S 300 may be performed that the server 100 receives mediation requests from a plurality of terminals.
  • operation S 310 may be performed that the server 100 may mediate a video communication connection between the first terminal 200 and the second terminal 210 among the plurality of terminals.
  • the server 100 may select the first terminal 200 and the second terminal 210 among the plurality of terminals.
  • the server 100 may mediate a video communication connection between the selected first terminal 200 and second terminal 210 .
  • the server 100 may transmit information on the connection for the first terminal 200 to the second terminal 210 .
  • the server 100 may transmit information on the connection for the second terminal 210 to the first terminal 200 .
  • operation S 320 may be performed that the server 100 receives the report, from the first terminal 200 , indicating that a user of the first terminal 200 or a user of the second terminal 210 is an abusive user as the result of the examination on, by the first terminal 200 , an image of the surrounding area of the first terminal 200 , captured by the first terminal 200 , a sound from the surrounding area of the first terminal 200 , recorded by the first terminal 200 , an image received from the second terminal 210 , or a sound received from the second terminal 210
  • the first terminal 200 may examine the image captured or the sound recorded by the first terminal 200 . When it is determined that the user of the first terminal 200 is an abusive user, the first terminal 200 may transmit the report indicating that the user of the first terminal 200 is an abusive user to the server 100 .
  • the terminal 200 may examine the image captured or the sound received from the second terminal 210 .
  • the first terminal may transmit the report indicating that the user of the second terminal 210 is an abusive user to the server 100 .
  • operation S 330 may be performed that the server 100 mediates a video communication connection between the terminal of the reported user and the terminal of a monitoring agent in response to the report.
  • the report indicating that the user of the first terminal 200 is an abusive user is received from the first terminal 200 .
  • the server 100 may mediates a video communication connection between the first terminal 100 and the third terminal 300 in response to the report.
  • the third terminal 300 may be a terminal of a monitoring agent. In other words, the user of the third terminal 300 may be a monitoring agent.
  • operation S 340 may be performed that the server 100 receives the result of verification on the reported user from the terminal of the monitoring agent.
  • the user of the third terminal 300 may input information on whether the user of the first terminal 200 is an abusive user after making a video communication with the user of the first terminal 200 .
  • the third terminal 300 may transmit the input information as the result of the verification on the user of the first terminal 200 .
  • operation S 350 may be performed that the server 100 rejects the additional mediation request from the terminal of the reported user according to the received result of the verification.
  • the server 100 may mediate a video call connection between the first terminal 200 and other terminal in response to the additional mediation request from the first terminal 200 .
  • the server 100 may reject the additional mediation request from the first terminal 200 .
  • a video call mediation request from a terminal of a user who is verified as an abusive user may be rejected, so that a healthy video communication between users is induced.
  • a user in a video communication may be prevented from being exposed to an inappropriate image or sound from the other party. Further, a user in a video communication may be prevented from transmitting an inappropriate image or sound to the other party. Further, a user in a video communication may be prevented from feeling sexual shame or discomfort due to the image or sound of the other party. Further, a healthy video communication between users can be induced.
  • the embodiment as described above may be embodied in a non-transitory computer-readable recording medium having recorded thereon computer-executable instruction codes, such as a program module executable by the computer.
  • the non-transitory computer-readable recording medium may be any available medium which can be accessed by the computer and may include any volatile and non-volatile media and any removable and non-removable media.
  • non-transitory computer-readable recording medium may include any computer storage medium and communication medium.
  • the computer storage medium may include any volatile and non-volatile media or any removable and non-removable media embodied by a certain method or technology for storing information such as computer-readable instruction codes, a data structure, a program module, or other data.
  • the communication medium may include computer-readable instruction code, a data structure, a program module, other data of modulated data signals, or other transmission mechanisms, and may include any information transmission medium.

Abstract

One or more embodiment relates to a video communication device, video communication method, and a video communication mediating method, which are capable of preventing a user in a video communication from being exposed to an inappropriate image or sound from the other party. One or more embodiment may provide a video communication method comprising establishing a first video communication session with a first terminal, consecutively receiving images or sounds from the first terminal through the first video communication session, examining at least a portion of images or sounds received in a period of recent specific time among the received images or sounds, and blocking an output of at least a portion of the received images or sounds or reporting a user of the first terminal as an abusive user to a first server according to the result of the examining.

Description

CROSS-REFERENCE TO RELATED APPLICATION
This application is a continuation of U.S. application Ser. No. 16/605,657, filed Oct. 16, 2019, which is a National Stage of International Application No. PCT/KR2018/000787 filed Jan. 17, 2018, claiming priority based on Korean Patent Application No. 10-2017-0048893 filed Apr. 17, 2017.
BACKGROUND 1. Field
One or more embodiments relate to a video communication device, a video communication method, and a video communication mediating method, which are capable of preventing a user who is making a video communication from being exposed to an inappropriate image or sound from the other party.
2. Description of the Related Art
With the development of communication technology and the miniaturization of electronic devices, personal terminals have become widespread to general consumers. In particular, portable personal terminals such as smart phones or smart tablets have been widely used in recent years. Most of terminals include a communication function. A user may use a terminal to perform a search on the Internet or exchange messages with the other party.
In addition, with the development of small camera, small microphone, small display, and small speaker technology, most terminals such as smart phones include cameras, microphones, displays, and speakers. A user can record a voice or capture a video containing voice using a terminal. The user can confirm the recorded voice through a speaker included in the terminal or confirm the captured video through a display.
In addition, a user can share the recorded voice or the captured video with the other party using a communication function of a terminal. The user can transmit the previously recorded voice or the previously captured video to the other party. The user can also transmit a voice currently being recorded by the terminal or a video currently being captured to the other party in real time.
At the same time, the other party can transmit a voice currently being recorded or a movie currently being captured by a terminal to the user in real time. The display included in the user's terminal can simultaneously display a video currently being captured by the user's terminal and a video currently being captured by the other party's terminal. In addition, a speaker included in the terminal of the user can simultaneously reproduce the voice currently being recorded by the terminal of the other party. In other words, the user and the other party can make a video communication with each other using their own terminals.
The user and the other party may already be acquainted with each other or may not be acquainted with each other. A plurality of terminals including the terminal of the user and the terminal of the other party may be connected to the server. The server can mediate between the terminal of the user and the terminal of the other party. Therefore, even if the user and the other party do not know each other, they can make a video communication with each other through mediation of the server.
While a user is in a video communication with the other party, the user may be exposed to an inappropriate image or sound from the other party. Particularly, while the user is in a video communication with the other party having an unknown relationship with each other, there is a high possibility that the user is exposed to obscene images from the other party.
For example, a naked body image of the other party captured by the other party's terminal may be received at the user's terminal through the video communication session. The user's terminal displays the received image so that the user can be exposed to the sexually obscene image. When the user is exposed to an obscene image that she or he does not want, the user may feel sexual shame or discomfort.
SUMMARY
One or more embodiments may provide a video communication device, a video communication method, and a video communication mediating method, which are capable of preventing a user who is making a video communication from being exposed to an inappropriate video or sound from the other party.
One or more embodiments may provide a video communication device, a video communication method, and a video communication mediating method, which are capable of preventing a user from transmitting an inappropriate image or sound.
One or more embodiments may provide a video communication device, a video communication method, and a video communication mediating method, which are capable of preventing a user who is making a video communication from feeling sexual shame or discomfort caused by an image or sound from the other party.
One or more embodiments may provide a video communication device, a video communication method, and a video communication mediating method, which are capable of inducing a healthy video communication between users.
According to an embodiment, a video communication method may comprise establishing a first video communication session with a first terminal, consecutively receiving images or sounds from the first terminal through the first video communication session, examining at least a portion of images or sounds received in a period of recent specific time among the received images or sounds, and blocking an output of at least a portion of the received images or sounds or reporting a user of the first terminal as an abusive user to a first server according to a result of the examining.
Further, the examining may comprise detecting a specific portion from the at least a portion of images or sounds received in a period of recent specific time.
Further, the detecting the specific portion may comprise detecting a portion corresponding to a specific facial expression, body language, or body part from the at least a portion of images or sounds received in a period of recent specific time.
Further, the detecting the portion corresponding to a specific facial expression, body language, or body part may comprise detecting a portion corresponding at least one of an obscene gesture, an insult gesture, a racism gesture, sexual content, nudity, genitalia, sexual activity, unsanitary content, excreta, excretion, antisocial content, antihuman content, illegal conduct, criminal conduct, hate crime, violent conduct, maltreatment conduct, self-harmful conduct, a weapon, a drug, an antisocial symbol, hateful content, threatening content, phobia-inducing content, or blood from the at least a portion of images or sounds received in a period of recent specific time.
Further, the examining may comprise detecting a portion corresponding to a specific voice or language from the at least a portion of images or sounds received in a period of recent specific time.
Further, the detecting a portion corresponding to a specific voice or language may comprise detecting a portion corresponding to a language related to at least one of a swear word, insult, racial discrimination, defamation, obscenity, sexual harassment, antisocial conduct, antihuman conduct, terrorism, hate speech, violence, maltreatment, a drug, or threatening.
Further, the video communication method may comprise repeating additionally examining at least one time after the examining and before the blocking or the reporting, wherein the blocking or the reporting comprises blocking an output of the at least a portion of the received images or sounds or reporting a user of the first terminal when a cumulative number of times in which the specific portion is detected is a first number or more or the number of times in which the specific portion is consecutively detected a second number of times or more after establishing the first video communication session.
Further, the video communication method may comprise repeating the examining to the blocking or the reporting with regular or irregular cycles.
Further, the video communication method may comprise consecutively outputting the received images or sounds after the receiving.
Further, the video communication method may comprise preparing information on correlation between at least one feature included in an image or sound of an arbitrary user and information on whether the arbitrary user is an abusive user before the establishing the first video communication session, wherein the examining comprises calculating whether the user of the first terminal is an abusive user using the prepared information on correlation and the at least a portion of images or sounds received in a period of recent specific time.
Further, the preparing the information on correlation may comprise preparing a machine learning model, and the examining may comprise predicting whether the user of the first terminal is an abusive user using the prepared machine learning model and the at least a portion of images or sounds received in a period of recent specific time.
Further, the preparing the machine learning model may comprise receiving the machine learning model from a second server.
Further, the preparing the machine learning model may comprise preparing at least one image or sound corresponding to a non-abusive user or abusive user and training the machine learning model using the at least one image or sound and information on whether each of the at least one image or sound corresponds to the abusive user.
Further, the preparing the at least one image or sound may comprise establishing at least one second video communication session with at least one terminal of the non-abusive user or at least one terminal of the abusive user and collecting at least a portion of images or sounds transmitted or received through the at least one second video communication session.
According to another embodiment, a video communication method may comprise establishing a video communication session between a first terminal and a second terminal, consecutively capturing images of a surrounding area of the first terminal or consecutively recording sounds from the surrounding area, examining at least a portion of images captured or sounds recorded in a period of recent specific time among the captured images or the recorded sounds by the first terminal, and according to a result of the examining, blocking transmission of at least a portion through the video communication session among the captured images or the recorded sounds by the first terminal or reporting a user of the first terminal as an abusive user to a first server.
Further, the video communication method may comprise repeating, by the first terminal, additionally examining at least one time after the examining and before the blocking or the reporting, wherein the blocking or the reporting may comprise blocking the transmission of the at least a portion of the captured images or recorded sounds or reporting the user of the first terminal when a cumulative number of times in which the specific portion is detected is a first number or more or the number of times in which the specific portion is consecutively detected a second number of times or more after establishing the video communication session.
Further, the video communication method may comprise receiving, by the first terminal, a machine learning model from a second server, which corresponds to correlation between at least one feature included in an image or sound of an arbitrary user and information on whether the arbitrary user is an abusive user, before the establishing the video communication session, wherein the examining comprises predicting whether the user of the first terminal is an abusive user using the received machine learning model and the at least a portion of images captured or sounds recorded in a period of recent specific time.
Further, the video communication method may comprise, before the establishing the video communication, preparing, by the first terminal, at least one image or sound corresponding to a non-abusive user or abusive user and training, by the first terminal, the machine learning model using the trained machine learning model and the at least a portion of images captured or sounds recorded in a period of recent specific time, wherein the examining may comprise predicting, by the first terminal, whether the user of the first terminal is an abusive user using the trained machine learning model and the at least a portion of images captured or sounds recorded in a period of recent specific time.
According to still another embodiment, a video communication mediating method may comprise receiving, by a server, a mediation request from a plurality of terminals, mediating, by the server, a video communication connection between a first terminal and a second terminal among the plurality of terminals, receiving, by the server, a report from the first terminal, which indicates that a user of the first terminal or a user of the second terminal is an abusive user as the result of the first terminal's examination on an image of a surrounding area of the first terminal, captured by the first terminal, a sound from the surrounding area of the first terminal, recorded by the first terminal, an image received by the first terminal from the second terminal, or a sound received by the first terminal from the second terminal, mediating, by the server, a video communication connection between a terminal of the reported user and a terminal of a monitoring agent in response to the report, receiving, by the server, a result of verification on the reported user from the terminal of the monitoring agent, and rejecting, by the server, an additional mediation request from the reported user according to the received result of verification.
According to yet another embodiment, a video communication device may comprise a communication interface configured to establish a video communication session with a terminal and consecutively receive images or sounds from the terminal through the video communication session, an output interface configured to consecutively output the received images or sounds, and a controller configured to examine at least a portion of images or sounds received in a period of recent specific time, wherein according to a result of examination, the controller may control the output interface to block an output of at least a portion of the received images or sounds or the communication interface to report a user of the terminal as an abusive user to a server.
According to yet another embodiment, a video communication device may comprise an input interface configured to consecutively capture images of a surrounding area thereof or consecutively record sounds from the surrounding area, a communication interface configured to establish a video communication session with a terminal and consecutively transmit the captured images or recorded sounds to the terminal through the video communication session, and a controller configured to examine at least a portion of images captured or sounds recorded in a period of recent specific time among the captured images or recorded sounds, wherein according to a result of the examination, the communication interface may block transmission of at least a portion of the captured images or recorded sounds or report a user of the video communication device as an abusive user to a server.
One or more embodiments can prevent a user who is making a video communication from being exposed to an inappropriate image or sound from the other party.
One or more embodiments can prevent a user from transmitting an inappropriate image or sound.
One or more embodiments can prevent a user who is making a video communication from feeling sexual shame or discomfort caused by an image or sound from the other party.
One or more embodiments can induce a healthy video communication between users.
BRIEF DESCRIPTION OF THE DRAWINGS
These and/or other aspects will become apparent and more readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings in which:
FIG. 1 is a network configuration diagram illustrating an environment in which a video communication device according to an embodiment operates;
FIG. 2 is a block diagram illustrating a configuration of a video communication device according to an embodiment;
FIG. 3 is a flowchart illustrating a process of performing a video communication method according to an embodiment;
FIG. 4 is a flowchart illustrating a process of performing operation of preparing information on correlation according to an embodiment;
FIG. 5 is a flowchart illustrating a process of performing operation of preparing at least one image or sound according to an embodiment;
FIG. 6 is a front view illustrating a second video communication device in which an image of a user of the second video communication device and an image of the other party are displayed during a video communication according to an embodiment;
FIG. 7 is a front view illustrating a second video communication device in which an image of a user of the second video communication device and an image of the other party are displayed during a video communication according to another embodiment;
FIG. 8 is a front view illustrating a first video communication device or a second video communication device or a in which an image of a user of the first video communication device and an image of a user of the second video communication device are displayed during a video communication according to still another embodiment;
FIG. 9 is a flowchart illustrating a process of performing a video communication method according to another embodiment; and
FIG. 10 is a flowchart illustrating a process of performing a video communication mediating method according to an embodiment.
DETAILED DESCRIPTION
Advantages and features of the present disclosure and a method of achieving the same will be more clearly understood from embodiments described below in detail with reference to the accompanying drawings. However, the present disclosure is not limited to the following embodiments and may be implemented in various different forms. The embodiments are provided merely to complete the present disclosure and to provide those of ordinary skill in the art with the category of the present disclosure. The present disclosure is defined only by the claims. Wherever possible, the same reference numbers will be used throughout the specification to refer to the same or like components.
It will be understood that although the terms “first,” “second,” etc. may be used to describe various components, these components should not be limited by these terms. These terms may be only used to distinguish one component from another. Therefore, a first component used herein may be a second component within the technical scope of the present disclosure.
The terminology used herein is for the purpose of describing embodiments and is not intended to limit the present disclosure. As used herein, the singular forms are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” or “comprising” used in this specification may denote that the used components or steps do not preclude the presence or addition of one or more other components and/or steps.
Unless otherwise defined, all terms used herein have the same meaning as commonly understood by those of ordinary skill in the art. It will be further understood that terms, such as those defined in commonly used dictionaries, will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
Hereinafter, video communication devices 200, 210, and 300, a video communication method, video communication medicating method, according to embodiments, will be described in detail with reference to FIGS. 1 to 10.
FIG. 1 is a network configuration diagram illustrating an environment in which the video communication devices 200, 210, and 300 according to an embodiment operate. Referring to FIG. 1, the environment in which the video communication devices 200, 210, and 300 operate may include a server 100 and at least two of the video communication devices 200, 210, and 300 connected to each other. Although the video communication devices 200, 210, and 300 are only illustrated in FIG. 1 as a matter of convenience, an additional video communication device other than the video communication devices 200, 210, and 300 may be included.
The server 100 may be connected to a communication network. The server 100 may be connected to an external device via the communication network. The server 100 may transmit data to other devices connected to each other or receive data from other devices.
The communication network connected to the server 100 may include a wired communication network, a wireless communication network, or a composite communication network. The communication network may include a mobile communication network such as 3G, Long Term Evolution (LTE), or LTE-A. The communication network may include a wired or wireless communication network such as Wi-Fi, Universal Mobile Telephone System/General Packet Radio Service (UMTS/GPRS), or Ethernet. The communication network may include a short-range communication network such as Magnetic Secure Transmission (MST), Radio Frequency Identification (RFID), Near Field Communication (NFC), ZigBee, Z-Wave, Bluetooth, Bluetooth Low Energy (BLE), or InfraRed (IR) communication. The communication network may include Local Area Network (LAN), Metropolitan Area Network (MAN), or Wide Area Network (WAN).
The server 100 may be connected to each of the video communication devices 200, 210, and 300 through a communication network. When the server 100 is connected to a first video communication device 200, the server 100 may transceive data with the first video communication device 200 through a communication network. When the server 100 is connected to a second video communication device 210, the server 100 may transceive data with the second video communication device 210 through a communication network. When the server 100 is connected to a third video communication device 300, the server 100 may transceive data with the third video communication device 300 through a communication network.
The server 100 may receive data from the video communication devices 200, 210, and 300. The server 100 may operate calculations using data received from the video all devices 200, 210, and 300. The server 100 may transmit the calculation result to the video communication devices 200, 210, and 300.
The server 100 may receive mediation requests of video communications from a plurality of video communication devices 200, 210, and 300. The server 100 may select at least two video communication devices among the plurality of video communication devices 200, 210, and 300 that transmit the mediation requests. For example, the server 100 may select the first video communication device 200 and the second video communication device 210 as the at least two video communication devices 200, 210, and 300.
The server 100 may mediate a connection of a video communication between the selected at least two video communication devices 200, 210, and 300. For example, the server 100 may transmit the connection information on the first video communication device 200 to the second video communication device 210. As another example, the server 100 may transmit the connection information on the second video communication device 210 to the first video communication device 200.
The connection information on the first video communication device 200 may include, for example, an IP address and a port number of the first video communication device 200. The second video communication device 210 receiving the connection information on the first video communication device 200 may attempt to connect to the first video communication device 200 using the received connection information.
The connection information on the second video communication device 210 may include, for example, an IP address and a port number of the second video communication device 210. The first video communication device 200 receiving the connection information on the second video communication device 210 may attempt to connect to the second video communication device 210 using the received connection information.
Succeeding in an attempt to connect the first video communication device 200 to the second video communication device 210 or an attempt to connect the second video communication device 210 to the first video communication device 200 may enable establishment of a video communication session between the first video communication device 200 and the second video communication device 210. The first video communication device 200 may transmit an image or sound to the second video communication device 210 through the video communication session. The first video communication device 200 may encode an image or sound into a digital signal and transmit the encoded result to the second video communication device 210.
Further, the first video communication device 200 may receive an image or sound from the second video communication device 210 through the video communication session. The first video communication device 200 may receive an image or sound encoded in a digital signal and may decode the received encoded image or sound.
The second video communication device 210 may transmit an image or sound to the first video communication device 200 through the video communication session. Further, the second video communication device 210 may receive an image or sound from the first video communication device 200 through the video communication session. Accordingly, a user of the first video communication device 200 and another user of the second video communication device 210 may make a video communication with each other.
The video communication devices 200, 210 and 300 may be terminals. The video communication devices 200, 210 and 300 may be, for example, desktop computers, laptop computers, smart phones, smart tablets, smart watches, mobile terminals, digital cameras, wearable devices, or portable electronic devices. The video communication devices 200, 210, and 300 may execute programs or applications. Each of the video communication devices 200, 210 and 300 may be the same kind of device each other or may not be the same kind of device each other.
Hereinafter, the first video communication device 200 and a first terminal 200 may be interchangeably used as a term to refer to the first video communication device 200 for convenience of description. Hereinafter, the second video communication device 210 and a second terminal 210 will be interchangeably used as a term to refer to the second video communication device 210 and a third video communication device 300, and the third terminal 300 will be interchangeably used as a term to refer to the third video communication device 300.
FIG. 2 is a block diagram showing a configuration of the first terminal 200 according to an embodiment. The configuration of the second terminal 210 or the configuration of the third terminal 300 may be the same as or different from the configuration of the first terminal 200. Referring to FIG. 2, the first terminal 200 according to the embodiment may include an input interface 201, a storage 202, a communication interface 203, an output interface 204, and a controller 205.
The input interface 201 may receive input from the outside. The input interface 201 may receive input from a user of the first terminal 200. Further, the input interface 201 may receive a signal from an external device as the input. The input interface 201 may include, for example, a microphone, a camera, a keyboard, a mouse, a trackball, a touch screen, a button, a switch, a sensor, a network interface, or other input devices.
The input interface 201 may receive light from outside through a camera included in the input interface 201. The input interface 201 may receive light emitted from an external object or light reflected from an external object. The input interface 201 may consecutively capture images of the surrounding area of the first terminal 200 through the camera. The input interface 201 may capture an image of a user of the first terminal 200.
The input interface 201 may receive a sound from outside through a microphone included in the input interface 201. The input interface 201 may receive a sound from the user of the first terminal 200. Further, the input interface 201 may receive a sound from an external sound reproducing device. The input interface 201 may consecutively record sounds from the surrounding area of the first terminal 200 through the microphone. The input interface 201 may record a voice of the user of the first terminal 200.
The storage 202 may store data. The storage 202 may store the input received by the input interface 201. The storage 202 may store the image captured or the sound recorded by the input interface 201. The storage 202 may store data to be used by the controller 205 to perform an arithmetic operation. The storage 202 may store the result of the arithmetic operation performed by the controller 205. For example, the storage 202 may store the image or voice encoded by the controller 205. The storage 202 may store data to be transmitted to the outside through the communication interface 203 or may store data received from the outside through the communication interface 203.
The storage 202 may include a volatile memory or a non-volatile memory. The storage 202 may include, for example, flash memory, read-only memory (ROM), random access memory (RAM), electrically erasable read-only memory (EEROM), erasable programmable read-only memory (EPROM), electrically erasable and programmable read-only memory (EEPROM), hard disk drive (HDD), register, or the like. The storage 202 may include, for example, a file system, a database, or an embedded database.
The communication interface 203 may transmit data to the exterior or receive data from the exterior. The communication interface 203 may transmit data to the other video communication devices 210 and 300 or the server 100. The communication interface 203 may receive data from the other video communication devices 210 and 300 or the server 100. The communication interface 203 may transmit the result of the arithmetic operation performed by the controller 205 to the exterior. Further, the communication interface 203 may transmit data stored in the storage 202 to the exterior. The communication interface 203 may transmit the image captured or the sound recorded by the input interface 201 to the exterior.
Data to be transmitted by the communication interface 203 or data received by the communication interface 203 may be stored in the storage 202. For example, the communication interface 203 may transmit an encoded image or voice stored in the storage 202 to the server 100 or the other video communication devices 210 and 300. Further, the communication interface 203 may receive an image or a voice encoded by the other video communication devices 210 and 300 from the other video communication devices 210 and 300.
The communication interface 203 may include, for example, a long-range network interface such as a 3G module, an LTE module, an LTE-A module, a Wi-Fi module, a WiGig module, an Ultra Wide Band (UWB) module, or a LAN card. Further, the communication interface 203 may include, for example, a short-range network interface such as an MST module, a Bluetooth module, an NFC module, an RFID module, a ZigBee module, a Z-Wave module, or an IR module. Further, the communication interface 203 may include other network interfaces.
The output interface 204 may output a signal to the exterior. The output interface 204 may, for example, display a screen, reproduce a sound, or output a vibration. The output interface 204 may include a display, a speaker, a vibrator, an oscillator, or other output devices.
The output interface 204 may display a screen. The controller 205 may control the output interface 204 to display a screen. The output interface 204 may display a user interface. The output interface 204 may display another screen in response to an input of a user.
The output interface 204 may display data. The output interface 204 may display the result of the arithmetic operation performed by the controller 205. The output interface 204 may display data stored in the storage 202. The output interface 204 may display data received by the communication interface 204.
The output interface 204 may display an image of a user of the first terminal 200 captured by the input interface 201 or images of the other users captured by the other video communication devices 210 and 300. Further, the output interface 204 may reproduce the voice of the user of the first terminal 200 recorded by the input interface 201 or the voice of the other users recorded by the other video communication devices 210 and 300.
The output interface 204 may include, for example, a flat-panel display device such as a liquid crystal display (LCD), an organic light-emitting diode (OLED), or a plasma display panel (PDP). The output interface 204 may include a curved display or a flexible display. The output interface 204 may include a touch screen.
The controller 205 may control an operation of the first terminal 200. The controller 205 may be connected to the components included in the first terminal 200. The controller 205 may control each operation of the components included in the first terminal 200
The controller 205 may process a signal. The controller 205 may process a signal received by the input interface 201. For example, the controller 205 may process the images of the users, captured by the camera included in the input interface 201. The controller 205 may process a sound and a voice of a user, recorded by the microphone included in the input interface 201.
The controller 205 may process the signal received by the communication interface 203. For example, the controller 205 may process a request of the user, received through the communication interface 203. The controller 205 may process images or voices of other users received through the communication interface 203.
The controller 205 may control an operation of the first terminal 200 in response to a signal received by the input interface 201. The controller 205 may control operation of the first terminal 200 in response to a request of a user received by the input interface 201.
The controller 205 may perform an arithmetic operation. The controller 205 may perform an arithmetic operation according to a signal received by the input interface 201. The controller 205 may perform an arithmetic operation using a signal received by the input interface 201 or data stored in the storage 202. For example, the controller 205 may perform image-processing using an image captured by the camera included in the input interface 201. The controller 205 may encode an image captured by the camera included in the input interface 201. The controller 205 may encode a sound or a voice recorded by the microphone included in the input interface 201. The controller 205 may encode a sound or a voice recorded by the camera included in the input interface 201.
The controller 205 may decode images or voices of other users, received by the communication interface 203. The controller 205 may perform post-processing on the decoded image or voice. For example, the controller 205 may perform image-processing on images of other users, received by the communication interface 203.
The controller 205 may control such that the result of the arithmetic operation is stored in the storage 202. The controller 205 may control such that the result of the arithmetic operation is output by the output interface 105. The controller 204 may control that the result of the arithmetic operation is transmitted to the other video communication devices 210 and 300 by the communication interface 203.
The controller 205 may include, for example, a central processing unit (CPU), a graphic processing unit (GPU), a micro controller unit (MCU), or a microprocessor.
A type of the second terminal 210 may be same as or different from a type of the first terminal 200. The second terminal 210 may execute a program or an application.
The second terminal 210 may be connected to a communication network. The second terminal 210 may be connected to an external device with each other via the communication network. The second terminal 210 may transmit data to the connected external device or receive data from the connected external device.
The second terminal 210 may be connected to the first terminal 200 with each other by mediation of the server 100. The second terminal 210 may transmit data to the first terminal 200 or receive data from the first terminal 200. The first terminal 200 and the second terminal 210, which are connected to each other, may transceive a message, a file, data, an image, a video, a sound, or a voice, and so on with each other.
The second terminal 210 may establish a voice communication session or a video communication session with the first terminal 200. The video communication session may be established using, for example, transmission control protocol (TCP), user datagram protocol (UDP), or web real-time communication (WebRTC). The user of the first terminal 200 and the user of the second terminal 210 may make a voice communication or a video communication with each other.
The video communication session may directly be established between the first terminal 200 and the second terminal 210. According to another example, the video communication session may directly be established between the first terminal 200 and the second terminal 210 via at least one other device. For example, the video communication session may include a session established between the first terminal 200 and the server 100 and a session established between the server 100 and the second terminal 210.
The first terminal 200 and the second terminal 210 may directly exchange data with each other. Further, the first terminal 200 and the second terminal 210 may exchange data with each other via the server 100 or other mediating devices.
A type of the third terminal 300 may be identical to or different from a type of the first terminal 200. The third terminal 300 may execute a program or an application.
The third terminal 300 may be connected to a communication network. The third terminal 300 may be connected to an external device with each other via the communication network. The third terminal 300 may transmit data to the connected external device or receive data from the connected external device.
Hereinafter, the specific operations of the video communication devices 200, 210, and 300 in the video communication method, according to an embodiment, will be described in detail with reference to FIGS. 3 to 9.
FIG. 3 is a flowchart illustrating a process of performing a video communication method by the second terminal 210, according to an embodiment. Referring to FIG. 3, in the video communication method according to an embodiment, first, operation S100 may be performed to prepare information on correlation between at least one feature included in an image or sound of an arbitrary user and information on whether the arbitrary user is an abusive user. The storage 202 of the second terminal 210 may prepare information on the correlation.
An arbitrary image or sound of the arbitrary user may include at least one feature. The feature included in the arbitrary image may be a specific shape, pattern, color, or a combination thereof included in the arbitrary image. The feature may be, for example, a specific facial expression, a specific body language, a specific shape, pattern, or color corresponding to a body part, or combination thereof.
Each feature included in the arbitrary image may correspond to a non-abusive element or an abusive element. For example, the feature may correspond to at least one abusive element of an obscene gesture, an insult gesture, a racism gesture, sexual content, nudity, genitalia, sexual activity, unsanitary content, excreta, excretion, antisocial content, antihuman content, illegal conduct, criminal conduct, hate crime, violent conduct, maltreatment conduct, self-harmful conduct, a weapon, a drug, an antisocial symbol, hateful content, threatening content, phobia-inducing content, or blood.
The feature included in the arbitrary sound may be a specific waveform, a pattern, or a combination thereof included in the arbitrary sound. The feature may be, for example, a specific waveform, pattern, or a combination thereof corresponding to a specific voice or language.
Each feature included in the arbitrary sound may correspond to a non-abusive element an abusive element. For example, the feature may correspond to at least one abusive element of groaning, gasping sound, screaming, or shrieking. As another example, the feature may correspond to at least one abusive element among a crash, a gun-shooting sound, a vomit sound, a hit sound, a whip sound, a knife sound, a swing sound, a stab sound, a sound having a pitch equal to or greater than an arbitrary threshold value, or a sound having a loudness equal to and greater than a specific threshold value. As still another example, the feature may correspond to a language related to at least one of a swear word, insult, racial discrimination, defamation, obscenity, sexual harassment, antisocial conduct, antihuman conduct, terrorism, hate speech, violence, maltreatment, a drug, or threatening as the abusive element.
The storage 202 of the second terminal 210 may prepare at least one of a list, a table, a database, or a machine learning model including information on the correlation. The storage 202 of the second terminal 210 may prepare a list including at least one item. The storage 202 of the second terminal 210 may prepare a database corresponding to the list.
Each item included in the list may include information on one feature. Each feature corresponding to each item may correspond to an abusive element. Therefore, when any feature among features corresponding to the item included in the list is identical to the feature included in the arbitrary image or sound, it may be determined that the arbitrary image or sound includes an abusive element. When it is determined that the arbitrary image or sound includes an abusive element, the arbitrary user may be determined as an abusive user.
The communication interface 203 of the second terminal 210 may receive the list or a database corresponding to the list from the server 100 or a separate server. The storage 202 of the second terminal 210 may store the received list or database. According to another embodiment, the controller 205 of the second terminal 210 may generate the list or database by itself, and the storage 202 of the second terminal 210 may store the generated list or database.
According to another embodiment, the storage 202 of the second terminal 210 may prepare a table including at least one record. The storage 202 of the second terminal 210 may prepare a database corresponding to the table.
Each record included in the table may include information on a feature and information on whether the feature corresponds to a non-abusive element or an abusive element. Therefore, when the information indicating that a record corresponding to a feature included in an arbitrary image or sound is present among the records included in the table and information indicating that the feature corresponds to an abusive element is included in the record, the arbitrary image or sound may be determined to contain an abusive element. When the arbitrary image or sound is determined to include an abusive element, the arbitrary user may be determined as an abusive user.
The communication interface 203 of the second terminal 210 may receive the table or a database corresponding to the table from the server 100 or a separate server. The storage 202 of the second terminal 210 may store the received table or database. According to another embodiment, the controller 205 of the second terminal 210 may generate the table or database by itself, and the storage 202 of the second terminal 210 may store the generated table or database.
According to another embodiment, the storage 202 of the second terminal 210 may prepare a machine learning model corresponding to the correlation. The machine learning model may be a neural network including at least two layers. The machine learning model may include an input layer and an output layer. The machine learning model may further include at least one hidden layer.
The machine learning model may be a classification model trained using at least one image or sound and information on whether the at least one image or sound corresponds to a non-abusive element or an abusive element. The machine learning model may use at least one feature included in an arbitrary image or sound as an input value. The machine learning model may predict whether the at least one feature corresponds to a non-abusive element or an abusive element. When it is predicted that the at least one feature corresponds to an abusive element, the arbitrary user may be determined as an abusive user.
The communication interface 203 of the second terminal 210 may receive the machine learning model from the server 100 or a separate server. The storage 202 of the second terminal 210 may store the received machine learning model. According to another embodiment, the controller 205 of the second terminal 210 may train the machine learning model by itself and the storage 202 of the second terminal 210 may store the trained machine learning model.
FIG. 4 is a flowchart of operation S100 of preparing information on correlation according to an embodiment. Referring to FIG. 4, first, operation S101 may be performed to prepare at least one image or sound corresponding to a non-abusive user or an abusive user. The storage 202 of the second terminal 210 may prepare at least one image or sound corresponding to a non-abusive user or an abusive user.
FIG. 5 is a flowchart of operation S101 of preparing at least one image or sound according to an embodiment. Referring to FIG. 5, first, operation S102 may be performed to establish at least one second video communication session between the terminals of at least one non-abusive user terminal or at least one abusive user terminal. The communication interface 203 of the second terminal 210 may establish the at least one second video communication session.
The communication interface 203 of the second terminal 210 may transmit an image or sound recorded by the input interface 201 to the other video communication devices 200 and 300 through the second video communication session. The output interface 204 of the second terminal 210 may output the captured image or the recorded sound. The communication interface 203 of the second terminal 210 may receive an image or sound recorded by the other video communication devices 200 and 300 from the other video communication devices 200 and 300 through the second video communication session. The output interface 204 of the second terminal 210 may output the received image or sound.
Next, operation S103 may be performed to collect at least one portion of the image or sound transmitted or received through the at least one second video communication session. In response to the output of the received image or sound, the input interface 201 of the second terminal 210 receives the feedback indicating that the output image or sound does not contain an abusive element from the user of the second terminal 210. The storage 202 of the second terminal 210 may collect the image or sound output when the feedback is received, as an image or sound corresponding to s non-abusive user.
Further, in response to the output of the received image or sound, the input interface 201 of the second terminal 210 may receive the feedback indicating that the output image or sound contains an abusive element from the user of the second terminal 210. For example, when an image or sound including an abusive element is output, the user of the second terminal 210 may touch a report button displayed on a touch screen. The storage 202 of the second terminal 210 may collect the image or sound output when the report button was touched, as an image or sound corresponding to an abusive user.
Referring again to FIG. 4, next, operation S104 may be performed to train the machine learning model using the at least one image or sound and information on whether each of the at least one image or sound corresponds to an abusive user. The controller 205 of the second terminal 210 may train the machine learning model using information on the at least one image or sound and information on whether each of the at least one image or sound corresponds to an abusive user, collected by the storage 202
For example, the controller 205 may train the machine learning model using a deep learning algorithm. The machine learning model may include at least one of deep neural network (DNN), convolutional neural network (CNN), recurrent neural network (RNN), region-based convolutional neural networks (R-CNN), restricted Boltzmann machine (RBM), deep belief network (DBN), or deep Q-networks.
The machine learning model may be convolutional neural network (CNN). The machine learning model may include AlexNet, ZFNet, GoogLeNet, VGGNet, ResNet, Inception-ResNet, Inception-v2, Inception-v3, or Inception-v4.
The controller 205 may convert each of the at least one image into an RGB format image having a width of 60 pixels and a height of 80 pixels using a convolutional layer. The controller 205 may train the machine learning model including Inception-v4 using each of the transformed images. The trained model may include total 4,330,627 variables. The size of each variable may be 4 bytes. Thus, the size of the trained model may be at least 17,322,508 bytes.
Referring again to FIG. 3, operation S110 may be performed to establish the first video communication session with the first terminal 200. The communication interface 203 of the second terminal 210 may establish the first video communication session.
Next, operation 120 may be performed to consecutively receive images or sounds from the first terminal 200 through the first video communication session. The communication interface 203 of the second terminal 210 may consecutively receive images or sounds through the first video communication session. The communication interface 203 may receive a data stream including an encoded image or sound from the first terminal 200. The controller 205 of the second terminal 210 may decode the received encoded image or sound.
Next, operation S130 may be performed to consecutively output the received images or sounds. The display included in the output interface 204 of the second terminal 210 may consecutively show the received image. The output interface 204 may display images of the surrounding area of the second terminal 210, being consecutively captured by the camera included in the input interface 201 of the second terminal 210 together with the received image.
Each of FIGS. 6 to 8 is a front view of the second terminal 210 displaying an image of a user of the second terminal 210 and an image of the other party during a video communication according to an embodiment. 6 to 8. As illustrated in FIGS. 6 to 8, the output interface 204 of the second terminal 210 may display images captured by the input interface 201 on a rectangular area positioned at the lower left of the display. Further, the output interface 204 of the second terminal 210 may display the images received by the communication interface 203 on the remaining area of the entire area of the display except for the rectangular area.
The speaker included in the output interface 204 of the second terminal 210 may consecutively reproduce the received sounds. The output interface 204 may reproduce sounds in the surrounding area of the second terminal 210, being consecutively recorded by the microphone included in the input interface 201 of the second terminal 210 together with the received sounds.
According to another embodiment, in operation S130, consecutively outputting the received images or sounds may be excluded. Alternatively, after operation S140 of examining at least a portion of images or sounds received in a period of recent specific time, the control interface 205 of the second terminal 210 may control the output interface 204 according to a result of the examination so that the received images or sounds is consecutively output.
Referring again to FIG. 5, next, operation S140 may be performed to examine the at least a portion of the images or sounds received in a period of recent specific time. The controller 205 of the second terminal 210 may examine at least a portion of the received images or sounds in a period of recent specific time. The specific time may be, for example, one second.
The controller 205 may examine all of the images or sounds received in a period of the recent specific time. When the images are being received at 30 frames per second (FPS), the controller 205 may examine all 30 images received for recent one second.
The controller 205 may select a portion of the images or sounds received in a period of the recent specific time. The controller 205 may check the selected portion of the images or sounds. For example, the controller 205 may arbitrarily select one image among the 30 images received for recent one second. The controller 205 may examine the selected single image. As another example, the controller 205 may examine an image of the user of the first terminal 200 displayed in the remaining area except for the rectangular area positioned at the lower left of the display in FIGS. 6 to 8.
The controller 205 may detect a specific portion from at least a portion of the image received in a period of the recent specific time. The controller 205 may detect a portion corresponding to a specific facial expression, body language, or body part as the specific portion. For example, the controller 205 may detect a portion corresponding to at least one of an obscene gesture, an insult gesture, a racism gesture, sexual content, nudity, genitalia, sexual activity, unsanitary content, excreta, excretion, antisocial content, antihuman content, illegal conduct, criminal conduct, hate crime, violent conduct, maltreatment conduct, self-harmful conduct, a weapon, a drug, an antisocial symbol, hateful content, threatening content, phobia-inducing content, or blood.
The controller 205 may detect a specific portion from at least one portion of the images received in a period of the recent specific time. The controller 205 may detect a portion corresponding to a specific voice or language. For example, the controller 205 may detect a portion corresponding to at least one of an obscene gesture, an insult gesture, a racism gesture, defamation, obscenity, sexual harassment, antisocial conduct, antihuman conduct, terrorism, hate speech, violence, maltreatment, a drug, or threatening.
The controller 205 of the second terminal 210 may calculate whether the user of the first terminal 200 is an abusive user by detecting the specific portion. When the specific portion is detected from the image or sound which is the examination object, the controller 205 may determine that the user of the first terminal 200 is an abusive user. When the specific portion is not detected from the image or sound which is the examination object, the controller 205 may determine that the user of the first terminal 200 is a non-abusive user.
According to another embodiment, the controller 205 may calculate whether the user of the first terminal 200 is an abusive user using information on the prepared correlation and the at least a portion of the images or sounds received in a period of the recent specific time. The controller 205 may extract a feature included in the at least a portion of the images or sounds received in a period of the recent specific time. The controller 205 may calculate whether the user of the first terminal 200 is an abusive user using the extracted feature and the information on the prepared correlation.
For example, the information on the prepared correlation may include the list. The controller 205 may compare a feature corresponding to at least one item included in the list and the extracted feature with each other. When one among features corresponding to the at least one item is the same as the extracted feature, the controller 205 may determine that the image or sound to be examined includes an abusive element. Further, the controller 205 may determine that the user of the first terminal 200 is an abusive user.
As another example, the information on the prepared correlation may include the table. The controller 205 may compare the extracted feature with the feature corresponding to at least one record included in the table. If information indicating that a record corresponding to the extracted feature is included in the records included in the table and that the extracted feature corresponds to an abusive element is included in the record, it may be determined that the image or sound includes an abusive element. Further, the controller 205 may determine that the user of the first terminal 200 is an abusive user.
As another example, the information on the prepared correlation may include the machine learning model. The controller 205 may predict whether the user of the first terminal 200 is an abusive user using the prepared machine learning model and the at least a portion of the images or sounds received in a period of the recent specific time.
For example, the controller 205 may convert an image subjected to the examination into an RGB format image having a width of 60 pixels and a height of 80 pixels using a convolutional layer. The controller 205 may use the converted image as an input value of the machine learning model. The controller 205 may run the machine learning model.
As the result of the running as described above, the result value may be calculated from the output layer of the machine learning model. The result value may be, for example, a vector including three numerical values. The three numerical values, respectively, may correspond to the probability that the image to be examined does not include an abusive element, the probability that the image to be examined includes an abusive element, and the probability that the image to be examined includes the appearance of a user of the first terminal 200, who has taken off his or her clothes that the user wear on the upper part of the body. The sum of the three numerical values may be one.
When the greatest value is the first numeric value among the three numeric values, the controller 205 may determine that the user of the first terminal 200 is a non-abusive user. For example, in FIG. 6, when an image of a user of the first terminal 200 displayed in an area other than the rectangular area positioned at the lower left of the display is examined, the greatest value may be the first numeric value among the three numerical values. The controller 205 may determine that the user of the first terminal 200 is a non-abusive user.
When the greatest value is the second or the third numeric value among the three numeric values, the controller 205 may determine that the user of the first terminal 200 is an abusive user. For example, in FIG. 7, when an image of a user of the first terminal 200 displayed in an area other than the rectangular area positioned at the lower left of the display is examined, the greatest value may be the second numeric value among the three numerical values. The controller 205 may determine that the user of the first terminal 200 is an abusive user.
Referring again to FIG. 3, operation S150 may be performed to repeat the operation S140 of examining at least one time. The controller 205 of the second terminal 210 may perform an examination on an image or sound received later than the image or sound that is the object of the latest examination after performing the latest examination. For example, the controller 205 may further perform four examinations after performing the first examination. In other words, the controller 205 may examine each of the five images or sounds received at an interval of one second from each other. Each result of the examinations may be accumulated. For example, when a portion corresponding to an abusive element is detected in each of five images, the cumulative number of times may be five.
According to another embodiment, the operation S150 of repeating may be excluded.
Next, operation S160 may be performed to block the output of at least a portion of the received image or sound or to report the user of the first terminal 200 as an abusive user to the server 100 according to the result of the examination. When the user of the first terminal 200 is determined to be a non-abusive user, the controller may control the output interface 204 to consecutively output the received image or sound.
When the user of the first terminal 200 is determined to be an abusive user, the controller 205 may control the output interface 204 to block the at least a portion of the received image or sound. According to another embodiment, when the user of the first terminal 200 is determined to be an abusive user, the controller 205 may control the communication interface 203 to report the user of the first terminal 200 as an abusive user to the server 100. According to still another embodiment, when the user of the first terminal 200 is determined to be an abusive user, the at least a portion of the received image or sound may be blocked, and the user of the first terminal 200 may be reported as an abusive user to the server 100.
According to another embodiment, when the cumulative number of times that the specific portion is detected after establishing the first video communication session is a first number of times or more, the controller 205 may control the output interface 204 to block the output of the at least a portion or the communication interface 205 to report the user of the first terminal 200 as an abusive user. The first number of times may be, for example, one or five times. According to still another embodiment, when the cumulative number of times that the user of the first terminal 200 is determined to be an abusive user after establishing the first video communication session is the first number of times or more, the controller 205 may control the output interface 204 to block the output of the at least a portion or the communication interface 203 to report the user of the first terminal 200 as an abusive user.
According to still another embodiment, the number of times that the specific portion is consecutively detected is a second number of times or more, the controller 205 may control the output interface 204 to block the output of the at least a portion or the communication interface 205 to report the user of the first terminal 200 as an abusive user. The second number of times may be, for example, one or five times. According to yet another embodiment, when the cumulative number of times that the user of the first terminal 200 is consecutively determined to be an abusive user is the second number of times or more, the controller 205 may control the output interface 204 to block the output of the at least a portion or the communication interface 203 to report the user of the first terminal 200 as an abusive user.
As described above, only when the cumulative number of times or the consecutively determined number of times is a specific number of times or more, the output is blocked or the user of the first terminal 200 is reported, thereby tolerating intermittent errors of the determination that may occur in the result of the examination. Therefore, more stable operation of the second terminal 210 may further be ensured.
Next, operation S170 may be performed to repeat the operation S140 of examining or the operation S160 of blocking or reporting with regular or irregular cycles. For example, after the controller 205 performed the latest operation S140 of examining or the latest operation S160 of blocking or reporting, the operation S140 of examining or operation S160 of blocking or reporting may be repeated after at least 10 seconds have elapsed. Therefore, it may be consistently examined whether the images or sounds which the first terminal 200 consecutively transmits may include an abusive element.
According to the embodiment as described above, a user in a video communication may be prevented from being exposed to an inappropriate image or sound from the other party. Further, a user in a video communication may be prevented from feeling sexual shame or discomfort. Further, a healthy video communication between users can be induced.
FIG. 9 is a flowchart illustrating a process of a video communication method by the first terminal 200, according to another embodiment. Referring to FIG. 9, first, operation S200 may be performed in which the first terminal 200 receives a machine learning model from the server 100, which corresponds to correlation between at least one feature included in an image or sound of an arbitrary user and information on whether the arbitrary user is an abusive user.
The storage 202 of the first terminal 200 may prepare information on the correlation. The storage 202 of the first terminal 200 may prepare at least one of a list, a table, a database, or a machine learning model including information on the correlation.
The storage 202 of the first terminal 200 may prepare a machine learning model corresponding to the correlation. The communication interface 203 of the first terminal 200 may receive the machine learning model from the server 100 or a separate server. The storage 202 of the first terminal 200 may store the received machine learning model. According to another embodiment, the controller 205 of the first terminal 200 may train the machine learning model by itself, and the storage 202 of the first terminal 200 may store the trained machine learning model.
The storage 202 of the first terminal 200 may prepare at least one image or sound corresponding to a non-abusive user or an abusive user. The controller 205 of the first terminal 200 may train the machine learning model using the at least one image or sound and information on whether each of the at least one image or sound corresponds to an abusive user.
Next, operation S210 may be performed to establish a video communication session between the first terminal 200 and the second terminal 210. The communication interface 203 of the first terminal 200 may establish a video communication session with the second terminal 210.
Next, operation S220 may be performed in which the first terminal 200 consecutively captures an image of the surrounding area of the first terminal 200 or consecutively records a sound from the surrounding area. The camera included in the input interface 201 of the first terminal 200 may consecutively capture an image of the surrounding area of the first terminal 200. The microphone included in the input interface 201 of the first terminal 200 may consecutively record the sound from the surrounding area. The storage 202 of the first terminal 200 may store the captured image or the recorded sound.
Next, operation S230 may be performed in which the first terminal 200 consecutively transmits the captured images or recorded sounds to the second terminal 210 through the video communication session. The communication interface 203 of the first terminal 200 may consecutively transmit the captured images or recorded sounds. Further, at the same time, the communication interface 203 may consecutively receive images or sounds from the second terminal 210. The output interface 204 may display the captured images together with the received images. The output interface 204 may reproduce the recorded sounds together with the received sounds.
According to another embodiment, the operation S230 of consecutively transmitting the captured images or the recorded sounds may be excluded.
Alternatively, after the operation S240 of examining at least a portion of the captured images or the recorded sounds in a period of the recent specific time is performed, the controller 205 of the first terminal 200 may control the communication interface 203 according to the result of the examination to consecutively transmit the captured images or the recorded sounds.
Next, operation S240 may be performed in which the first terminal 200 examines at least a portion of the images captured or sounds recorded in a period of the recent specific time among the captured images or the recorded sounds. The controller 205 of the first terminal 200 may detect a specific portion from the at least a portion of the images captured or sounds recorded in a period of the recent specific time. The specific time may be, for example, one second.
The controller 205 may predict whether the user of the first terminal 200 is an abusive user using the prepared machine learning model and the at least a portion of the images captured or sounds recorded in a period of the recent specific time. For example, the controller 205 may convert an image which is the examination object into an RGB format image having a width of 60 pixels and a height of 80 pixels using a convolutional layer. The controller 205 may use the converted image as an input value of the machine learning model. The controller 205 may run the machine learning model.
As the result of the operation of running as described above, the result value may be calculated from the output layer of the machine learning model. The result value may be, for example, a vector including three numerical values. The three numerical values, respectively, may correspond to the probability that the image to be examined does not include an abusive element, the probability that the image to be examined includes an abusive element, and the probability that the image to be examined includes the appearance of a user, who has taken off his or her clothes that the user wear on the upper part of the body, of the first terminal 200. The sum of the three numerical values may be one.
When the greatest value is the first numeric value among the three numeric values, the controller 205 may determine that the user of the first terminal 200 is a non-abusive user. For example, in FIG. 6, when an image of a user of the first terminal 200 displayed in the rectangular area positioned at the lower left of the display is examined, the greatest value may be the first numeric value among the three numerical values. The controller 205 may determine that the user of the first terminal 200 is a non-abusive user.
When the greatest value is the second or the third numeric value among the three numeric values, the controller 205 may determine that the user of the first terminal 200 is an abusive user. FIG. 8 is a front view of a first terminal 200 displaying an image of a user of the first terminal 200 and an image of a user of the second terminal 210 during a video communication according to the embodiment. For example, in FIG. 8, when an image of a user of the first terminal 200 displayed in the rectangular area positioned at the lower left of the display is examined, the greatest value may be the third numeric value among the three numerical values. The controller 205 may determine that the user of the first terminal 200 is an abusive user.
Next, operation S250 may be performed in which the first terminal 200 repeats the operation S240 of examining at least one time. The controller 205 of the first terminal 200 may perform an examination on an image or sound received later than the image or sound that is the object of the latest examination after performing the latest examination. For example, the controller 205 may further perform four examinations after performing the first examination. In other words, the controller 205 may examine each of the five images or sounds received at an interval of one second from each other. Each result of the examinations may be accumulated. For example, when a portion corresponding to an abusive element is detected in each of five images, the cumulative number of times may be five.
According to another embodiment, the operation S250 of repeating may be excluded.
Next, operation S260 may be performed in which the first terminal 200 blocks the transmission of the at least a portion of the captured image or the recorded sound through a video communication session or reports the user of the first terminal 200 as an abusive user to the server 100 according to the result of the examination.
When the user of the first terminal 200 is determined to be a non-abusive user, the controller may control the communication interface 203 to consecutively transmit the captured images or the recorded sounds.
When the user of the first terminal 200 is determined to be an abusive user, the controller 205 may control the communication interface 203 to block the transmission of the at least a portion of the captured images or the recorded sounds. According to another embodiment, when the user of the first terminal 200 is determined to be an abusive user, the controller 205 may control the communication interface 203 to report the user of the first terminal 200 as an abusive user to the server 100. According to still another embodiment, when the user of the first terminal 200 is determined to be an abusive user, the transmission of the at least a portion of the captured images or the recorded sounds may be blocked, and the user of the first terminal 200 may be reported as an abusive user to the server 100
According to another embodiment, when the cumulative number of times that the specific portion is detected after establishing the video communication session is a first number of times or more, the controller 205 may control the communication interface 203 to block the transmission of the at least a portion or to report the user of the first terminal 200 as an abusive user. The first number of times may be, for example, one or five times. According to still another embodiment, when the cumulative number of times that the user of the first terminal 200 is determined to be an abusive user after establishing the video communication session is the first number of times or more, the controller 205 may control the communication interface 203 to block the output of the at least a portion or to report the user of the first terminal 200 as an abusive user.
According to still another embodiment, the number of times that the specific portion is consecutively detected is a second number of times or more, the controller 205 may control the communication interface 203 to block the transmission of the at least a portion or to report the user of the first terminal 200 as an abusive user. The second number of times may be, for example, one or five times. According to yet another embodiment, when the cumulative the number of times that the user of the first terminal 200 is consecutively determined to be an abusive user is the second number of times or more, the controller 205 may control the communication interface 203 to block the transmission of the at least a portion or to report the user of the first terminal 200 as an abusive user.
As described above, only when the cumulative number of times or the consecutively determined number of times is a specific number of times or more, the transmission is blocked or the user of the first terminal 200 is reported, thereby tolerating intermittent errors of the determination that may occur in the result of the examination. Therefore, more stable operation of the first terminal 200 may further be ensured.
Next, operation S270 may be performed that the first terminal 200 repeats the operation S240 of examining or the operation S260 of blocking or reporting with regular or irregular cycles. For example, after the controller 205 performed the latest operation S140 of examining or the latest operation S160 of blocking or reporting, the operation S140 of examining or the operation S160 of blocking or reporting may be repeated after at least 10 seconds have elapsed. Therefore, it may be consistently examined whether the images or sounds which the first terminal 200 consecutively transmits may include an abusive element.
According to the embodiment as described above, a user in a video communication may be prevented from being exposed to an inappropriate image or sound from the other party. Further, a healthy video communication between users can be induced.
FIG. 10 is a flowchart of illustrating a video communication mediating method according to an embodiment. Referring to FIG. 10, operation S300 may be performed that the server 100 receives mediation requests from a plurality of terminals.
Next, operation S310 may be performed that the server 100 may mediate a video communication connection between the first terminal 200 and the second terminal 210 among the plurality of terminals. The server 100 may select the first terminal 200 and the second terminal 210 among the plurality of terminals. The server 100 may mediate a video communication connection between the selected first terminal 200 and second terminal 210. For example, the server 100 may transmit information on the connection for the first terminal 200 to the second terminal 210. As another example, the server 100 may transmit information on the connection for the second terminal 210 to the first terminal 200.
Next, operation S320 may be performed that the server 100 receives the report, from the first terminal 200, indicating that a user of the first terminal 200 or a user of the second terminal 210 is an abusive user as the result of the examination on, by the first terminal 200, an image of the surrounding area of the first terminal 200, captured by the first terminal 200, a sound from the surrounding area of the first terminal 200, recorded by the first terminal 200, an image received from the second terminal 210, or a sound received from the second terminal 210
The first terminal 200 may examine the image captured or the sound recorded by the first terminal 200. When it is determined that the user of the first terminal 200 is an abusive user, the first terminal 200 may transmit the report indicating that the user of the first terminal 200 is an abusive user to the server 100.
Further, the terminal 200 may examine the image captured or the sound received from the second terminal 210. When the user of the second terminal 210 is determined to be an abusive user as the result of the examination, the first terminal may transmit the report indicating that the user of the second terminal 210 is an abusive user to the server 100.
Next, operation S330 may be performed that the server 100 mediates a video communication connection between the terminal of the reported user and the terminal of a monitoring agent in response to the report. For example, it is assumed that the report indicating that the user of the first terminal 200 is an abusive user is received from the first terminal 200. The server 100 may mediates a video communication connection between the first terminal 100 and the third terminal 300 in response to the report. The third terminal 300 may be a terminal of a monitoring agent. In other words, the user of the third terminal 300 may be a monitoring agent.
Next, operation S340 may be performed that the server 100 receives the result of verification on the reported user from the terminal of the monitoring agent. The user of the third terminal 300 may input information on whether the user of the first terminal 200 is an abusive user after making a video communication with the user of the first terminal 200. The third terminal 300 may transmit the input information as the result of the verification on the user of the first terminal 200.
Next, operation S350 may be performed that the server 100 rejects the additional mediation request from the terminal of the reported user according to the received result of the verification. When it is verified that the user of the first terminal 200 is not an abusive user, the server 100 may mediate a video call connection between the first terminal 200 and other terminal in response to the additional mediation request from the first terminal 200. When it is verified that the user of the first terminal 200 is an abusive user, the server 100 may reject the additional mediation request from the first terminal 200.
According to the embodiment as described above, a video call mediation request from a terminal of a user who is verified as an abusive user may be rejected, so that a healthy video communication between users is induced.
According to the embodiment as described above, a user in a video communication may be prevented from being exposed to an inappropriate image or sound from the other party. Further, a user in a video communication may be prevented from transmitting an inappropriate image or sound to the other party. Further, a user in a video communication may be prevented from feeling sexual shame or discomfort due to the image or sound of the other party. Further, a healthy video communication between users can be induced.
The embodiment as described above may be embodied in a non-transitory computer-readable recording medium having recorded thereon computer-executable instruction codes, such as a program module executable by the computer. The non-transitory computer-readable recording medium may be any available medium which can be accessed by the computer and may include any volatile and non-volatile media and any removable and non-removable media.
Furthermore, the non-transitory computer-readable recording medium may include any computer storage medium and communication medium. The computer storage medium may include any volatile and non-volatile media or any removable and non-removable media embodied by a certain method or technology for storing information such as computer-readable instruction codes, a data structure, a program module, or other data. The communication medium may include computer-readable instruction code, a data structure, a program module, other data of modulated data signals, or other transmission mechanisms, and may include any information transmission medium.
It should be understood that embodiments described herein should be considered in a descriptive sense only and not for purposes of limitation. Descriptions of features or aspects within each embodiment should typically be considered as available for other similar features or aspects in other embodiments.

Claims (11)

What is claimed is:
1. A video communication mediating method comprising:
receiving, by a server, a mediation request from a plurality of terminals;
mediating, by the server, a first video communication connection between a first terminal and a second terminal among the plurality of terminals;
receiving, by the server, a report from the first terminal, which indicates that a user of the first terminal or a user of the second terminal is an abusive user;
mediating, by the server, a second video communication connection between a terminal of the reported user and a terminal of a monitoring agent after ending the first video communication connection in response to the report; and
receiving, by the server, a result of verification on the reported user from the terminal of the monitoring agent.
2. The video communication mediating method of claim 1, wherein the receiving comprises receive the report that the user of at least one of the first terminal and the second terminal is the abusive user according to an examining result of examining images or sounds received by the at least one of the terminal by running a machine learning model in the at least one of the terminal.
3. The video communication mediating method of claim 1, wherein the method further comprises determining, by the server, whether to reject an additional mediation request from the reported user based on the received result of verification.
4. The video communication mediating method of claim 2, the machine learning model, as a result of running, calculates result value of probability that the image to be examined does not include an abusive element, the probability that the image to be examined includes an abusive element, and the probability that the image to be examined includes a specific abusive element.
5. The video communication mediating method of claim 4, wherein the receiving comprises receive the report that the user of the at least one terminal is the abusive user, if the number of times that the examining result includes the abusive element or the specific abusive element is more than the preset number.
6. A non-transitory computer-readable storage medium storing instructions that, when executed by a processor, cause the processor to:
receive a mediation request from a plurality of terminals;
mediate a first video communication connection between a first terminal and a second terminal among the plurality of terminals;
receive a report from the first terminal, which indicates that a user of the first terminal or a user of the second terminal is an abusive user;
mediate a second video communication connection between a terminal of the reported user and a terminal of a monitoring agent after ending the first video communication connection in response to the report; and
receive a result of verification on the reported user from the terminal of the monitoring agent.
7. A server for a video call mediation comprising at least one processor which is configured to:
receive a mediation request from a plurality of terminals;
mediate a first video communication connection between a first terminal and a second terminal among the plurality of terminals;
receive a report from the first terminal, which indicates that a user of the first terminal or a user of the second terminal is an abusive user;
mediate a second video communication connection between a terminal of the reported user and a terminal of a monitoring agent after ending the first video communication connection in response to the report; and
receive a result of verification on the reported user from the terminal of the monitoring agent.
8. The server of claim 7, wherein processor is further configured to receive the report that the user of at least one of the first terminal and the second terminal is the abusive user according to an examining result of examining images or sounds received by the at least one of the terminal by running a machine learning model in the at least one of the terminal.
9. The server of claim 7, wherein the processor is further configured to determine whether to reject an additional mediation request from the reported user based on the received result of verification.
10. The server of claim 8, wherein the machine learning model, as a result of running, calculates result value of probability that the image to be examined does not include an abusive element, the probability that the image to be examined includes an abusive element, and the probability that the image to be examined includes a specific abusive element.
11. The server of claim 10, wherein the processor is further configured to receive the report that the user of the at least one terminal is the abusive user, if the number of times that the examining result includes the abusive element or the specific abusive element is more than the preset number.
US17/183,890 2017-04-17 2021-02-24 Video communication device, video communication method, and video communication mediating method Active US11323659B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/659,130 US11722638B2 (en) 2017-04-17 2022-04-13 Video communication device, video communication method, and video communication mediating method

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
KR1020170048893A KR101932844B1 (en) 2017-04-17 2017-04-17 Device and method of making video calls and method of mediating video calls
KR10-2017-0048893 2017-04-17
PCT/KR2018/000787 WO2018194243A1 (en) 2017-04-17 2018-01-17 Video communication device, video communication method, and video communication mediating method

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
US16/605,657 Continuation US10965907B2 (en) 2017-04-17 2018-01-17 Video communication device, video communication method, and video communication mediating method
PCT/KR2018/000787 Continuation WO2018194243A1 (en) 2017-04-17 2018-01-17 Video communication device, video communication method, and video communication mediating method

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/659,130 Continuation US11722638B2 (en) 2017-04-17 2022-04-13 Video communication device, video communication method, and video communication mediating method

Publications (2)

Publication Number Publication Date
US20210203880A1 US20210203880A1 (en) 2021-07-01
US11323659B2 true US11323659B2 (en) 2022-05-03

Family

ID=63856376

Family Applications (3)

Application Number Title Priority Date Filing Date
US16/605,657 Active US10965907B2 (en) 2017-04-17 2018-01-17 Video communication device, video communication method, and video communication mediating method
US17/183,890 Active US11323659B2 (en) 2017-04-17 2021-02-24 Video communication device, video communication method, and video communication mediating method
US17/659,130 Active US11722638B2 (en) 2017-04-17 2022-04-13 Video communication device, video communication method, and video communication mediating method

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US16/605,657 Active US10965907B2 (en) 2017-04-17 2018-01-17 Video communication device, video communication method, and video communication mediating method

Family Applications After (1)

Application Number Title Priority Date Filing Date
US17/659,130 Active US11722638B2 (en) 2017-04-17 2022-04-13 Video communication device, video communication method, and video communication mediating method

Country Status (3)

Country Link
US (3) US10965907B2 (en)
KR (1) KR101932844B1 (en)
WO (1) WO2018194243A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220141396A1 (en) * 2020-10-29 2022-05-05 Acer Incorporated Video conferencing system and method of removing interruption thereof
US20220239848A1 (en) * 2021-01-26 2022-07-28 Dell Products, Lp System and method for operating an intelligent videoframe privacy monitoring management system for videoconferencing applications
US11496709B2 (en) 2020-01-31 2022-11-08 Hyperconnect Inc. Terminal, operating method thereof, and computer-readable recording medium
US11716424B2 (en) 2019-05-10 2023-08-01 Hyperconnect Inc. Video call mediation method
US11722638B2 (en) 2017-04-17 2023-08-08 Hyperconnect Inc. Video communication device, video communication method, and video communication mediating method
US11825236B2 (en) 2020-01-31 2023-11-21 Hyperconnect Inc. Terminal and operating method thereof
US11854263B2 (en) * 2018-07-23 2023-12-26 Tencent Technology (Shenzhen) Company Limited Video processing method and apparatus, terminal device, server, and storage medium

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11553157B2 (en) 2016-10-10 2023-01-10 Hyperconnect Inc. Device and method of displaying images
CN109982124A (en) * 2019-03-26 2019-07-05 深圳创维-Rgb电子有限公司 User's scene intelligent analysis method, device and storage medium
WO2021025574A1 (en) * 2019-08-05 2021-02-11 Марк Александрович НЕЧАЕВ Moderation of audio and video calls
KR102311603B1 (en) * 2019-10-01 2021-10-13 주식회사 하이퍼커넥트 Mobile and operating method thereof
WO2021078565A1 (en) * 2019-10-24 2021-04-29 Signify Holding B.V. System and method for space monitoring
EP3872779A1 (en) * 2020-02-27 2021-09-01 Axis AB Method, software, and device for training an alarm system to classify audio of an event
KR20210115442A (en) 2020-03-13 2021-09-27 주식회사 하이퍼커넥트 Report evaluation device and operating method thereof
US20240121358A1 (en) * 2021-03-11 2024-04-11 Hewlett-Packard Development Company, L.P. Meeting video substitutes
KR102576636B1 (en) * 2021-03-22 2023-09-11 하이퍼커넥트 유한책임회사 Method and apparatus for providing video stream based on machine learning
US11622147B2 (en) * 2021-07-22 2023-04-04 Popio Mobile Video Cloud, Llc Blurring digital video streams upon initiating digital video communications
US11895101B2 (en) * 2021-11-22 2024-02-06 Dell Products, L.P. Machine learning development hub

Citations (106)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH09200714A (en) 1996-01-17 1997-07-31 Casio Comput Co Ltd Image processor with video telephone
KR20010000426A (en) 2000-09-28 2001-01-05 김용환 Method of Intelligent Image Interface
JP2001309325A (en) 2000-04-19 2001-11-02 Mitsubishi Electric Corp Video telephone system
US20030091239A1 (en) 2001-11-13 2003-05-15 Kazuyuki Imagawa Communications method using images and device for the same
US20030126267A1 (en) 2001-12-27 2003-07-03 Koninklijke Philips Electronics N.V. Method and apparatus for preventing access to inappropriate content over a network based on audio or visual content
JP2003219383A (en) 2001-11-13 2003-07-31 Matsushita Electric Ind Co Ltd Communications method using image and device for the same
US20040049780A1 (en) 2002-09-10 2004-03-11 Jeanette Gee System, method, and computer program product for selective replacement of objectionable program content with less-objectionable content
KR20040064972A (en) 2003-01-13 2004-07-21 엘지전자 주식회사 televiewing limit method for display device
JP2006270380A (en) 2005-03-23 2006-10-05 Victor Co Of Japan Ltd Image information transmitter
US20070195158A1 (en) 2006-02-23 2007-08-23 Kies Jonathan K Apparatus and methods for managing video calls on a wireless device
US20080012935A1 (en) 2005-11-22 2008-01-17 Gateway Inc. Inappropriate content detection and distribution prevention for wireless cameras/camcorders with e-mail capabilities and camera phones
US7321384B1 (en) 2003-06-03 2008-01-22 Cisco Technology, Inc. Method and apparatus for using far end camera control (FECC) messages to implement participant and layout selection in a multipoint videoconference
KR100844334B1 (en) 2007-06-07 2008-07-08 엘지전자 주식회사 Apparatus and method for controlling display in mobile communication terminal
US20080259154A1 (en) 2007-04-20 2008-10-23 General Instrument Corporation Simulating Short Depth of Field to Maximize Privacy in Videotelephony
US20080298571A1 (en) 2007-05-31 2008-12-04 Kurtz Andrew F Residential video communication system
KR20080110064A (en) 2007-06-14 2008-12-18 유성준 Method for distinguishing obscene image and apparatus therefor
US20090049467A1 (en) 2002-07-02 2009-02-19 Caption Tv, Inc. System, method and computer program product for selective filtering of objectionable content from a program
US7564476B1 (en) 2005-05-13 2009-07-21 Avaya Inc. Prevent video calls based on appearance
US20100017818A1 (en) 2008-07-17 2010-01-21 Cisco Technology, Inc. Non-bandwidth intensive method for providing multiple levels of censoring in an a/v stream
US20100054592A1 (en) 2004-10-28 2010-03-04 Fotonation Ireland Limited Analyzing partial face regions for red-eye detection in acquired digital images
US20100124941A1 (en) 2008-11-19 2010-05-20 Samsung Electronics Co., Ltd. Method and device for synthesizing image
US20100226261A1 (en) * 2005-09-16 2010-09-09 Eyeball Networks Inc. Method and system to prevent spam over internet telephony
KR20110019499A (en) 2009-08-20 2011-02-28 엘지전자 주식회사 Mobile terminal
KR20110025720A (en) 2009-09-05 2011-03-11 에스케이텔레콤 주식회사 Method for video call based on an avatar and system, apparatus thereof
JP2011082694A (en) 2009-10-05 2011-04-21 Sharp Corp System and device for communication conference
KR20110073238A (en) 2009-12-21 2011-06-29 한국전자통신연구원 Apparatus and method for blocking the objectionable multimedia based on multimodal and multiscale features
JP2011205243A (en) 2010-03-24 2011-10-13 Sharp Corp Information processing apparatus, conference system, information processing method, and computer program
JP2012018571A (en) 2010-07-08 2012-01-26 Ntt Docomo Inc Content recommendation apparatus and method
JP2012054897A (en) 2010-09-03 2012-03-15 Sharp Corp Conference system, information processing apparatus, and information processing method
US8165416B2 (en) 2007-06-29 2012-04-24 Microsoft Corporation Automatic gain and exposure control using region of interest detection
US20120155759A1 (en) 2010-12-21 2012-06-21 Microsoft Corporation Establishing clusters of user preferences for image enhancement
US20120182379A1 (en) 2009-09-24 2012-07-19 Zte Corporation Method, Application Server and System for Privacy Protection in Video Call
JP2012146209A (en) 2011-01-13 2012-08-02 Nikon Corp Electronic device and control program for electronic device
KR20120090870A (en) 2012-05-09 2012-08-17 김명훈 Image security apparatus and image security system using face detect algorithm and image security method thereof
WO2012131932A1 (en) 2011-03-30 2012-10-04 富士通株式会社 Information processing device including image transmission control function
EP2515526A2 (en) 2011-04-08 2012-10-24 DigitalOptics Corporation Europe Limited Display device with image capture and analysis module
KR20120126677A (en) 2011-05-12 2012-11-21 재단법인 서울호서직업전문학교 Method and apparatus for automatically detecting pornographic image, and computer readable medium thereof
US20120320141A1 (en) 2011-06-16 2012-12-20 Vtel Products Corporation, Inc. Video conference control system and method
US20130083155A1 (en) 2011-09-30 2013-04-04 Cisco Technology Inc. Method, endpoint, and system for establishing a video conference
US8441356B1 (en) 2009-02-16 2013-05-14 Handhold Adaptive, LLC Methods for remote assistance of disabled persons
US20130147897A1 (en) 2010-09-10 2013-06-13 Shigehiro Ichimura Mobile terminal, remote operation system, data transmission control method by mobile terminal, and non-transitory computer readable medium
JP2013531923A (en) 2010-05-17 2013-08-08 アマゾン テクノロジーズ インコーポレイテッド Selective content presentation engine
US20130234826A1 (en) 2011-01-13 2013-09-12 Nikon Corporation Electronic device and electronic device control program
US8626879B2 (en) 2009-12-22 2014-01-07 Sling Media, Inc. Systems and methods for establishing network connections using local mediation services
KR20140012474A (en) 2012-07-20 2014-02-03 한국전자통신연구원 Apparatus and method for protecting privacy based on face recognition
US20140176732A1 (en) 2012-12-21 2014-06-26 Google Inc. Recommending transformations for photography
US20140267583A1 (en) 2013-03-13 2014-09-18 Futurewei Technologies, Inc. Augmented Video Calls on Mobile Devices
EP2782326A2 (en) 2013-03-20 2014-09-24 Samsung Electronics Co., Ltd. Method and apparatus for processing an image based on an image property and ambient environment information
US20150030314A1 (en) 2012-12-11 2015-01-29 Unify Gmbh & Co. Kg Method of processing video data, device, computer program product, and data construct
US8977063B2 (en) 2005-03-09 2015-03-10 Qualcomm Incorporated Region-of-interest extraction for video telephony
US20150070516A1 (en) 2012-12-14 2015-03-12 Biscotti Inc. Automatic Content Filtering
WO2015091487A1 (en) 2013-12-19 2015-06-25 Alcatel Lucent Method for real-time reporting and prevention of call abuse
US20150220777A1 (en) 2014-01-31 2015-08-06 Google Inc. Self-initiated change of appearance for subjects in video and images
US20150229882A1 (en) 2014-02-10 2015-08-13 Alibaba Group Holding Limited Video communication method and system in instant communication
US20150286858A1 (en) 2015-03-18 2015-10-08 Looksery, Inc. Emotion recognition in video conferencing
WO2015162647A1 (en) 2014-04-25 2015-10-29 Sony Corporation Processing digital photographs in response to external applications
US20150325023A1 (en) 2014-05-09 2015-11-12 Google Inc. Providing pre-edits for photos
US20150370474A1 (en) 2014-06-19 2015-12-24 BrightSky Labs, Inc. Multiple view interface for video editing system
US9225897B1 (en) 2014-07-07 2015-12-29 Snapchat, Inc. Apparatus and method for supplying content aware photo filters
US9230328B1 (en) 2013-06-28 2016-01-05 Google Inc. Providing image parameters
US20160019412A1 (en) 2014-07-18 2016-01-21 Htc Corporation Method for performing a face tracking function and an electric device having the same
US20160034786A1 (en) 2014-07-29 2016-02-04 Microsoft Corporation Computerized machine learning of interesting video sections
US9282287B1 (en) 2014-09-09 2016-03-08 Google Inc. Real-time video transformations in video conferences
US9319632B2 (en) 2013-01-02 2016-04-19 Samsung Electronics Co., Ltd. Display apparatus and method for video calling thereof
US20160127653A1 (en) 2014-11-03 2016-05-05 Samsung Electronics Co., Ltd. Electronic Device and Method for Providing Filter in Electronic Device
EP3035283A1 (en) 2013-08-15 2016-06-22 Xiaomi Inc. Image processing method and apparatus, and terminal device
US9380264B1 (en) * 2015-02-16 2016-06-28 Siva Prasad Vakalapudi System and method for video communication
JP2016122975A (en) 2014-12-25 2016-07-07 セイコーエプソン株式会社 Display device and display device control method
WO2016112299A1 (en) 2015-01-09 2016-07-14 Snapchat, Inc. Object recognition based photo filters
US20160316265A1 (en) * 2014-04-29 2016-10-27 At&T Intellectual Property I, Lp Method and apparatus for augmenting media content
US9503456B1 (en) * 2015-01-22 2016-11-22 Google Inc. Video chat abuse detection based on external context
US20160350675A1 (en) * 2015-06-01 2016-12-01 Facebook, Inc. Systems and methods to identify objectionable content
US20160373571A1 (en) 2014-06-19 2016-12-22 Microsoft Technology Licensing, Llc Use of a digital assistant in communications
US9569697B1 (en) 2015-02-19 2017-02-14 Google Inc. Object oriented image editing
US9591261B2 (en) 2013-09-24 2017-03-07 Hitachi Maxell, Ltd. Video telephone device and video telephone processing method
US20170251336A1 (en) 2012-09-11 2017-08-31 Amazon Technologies, Inc. Content display controls based on environmental factors
US9848167B1 (en) 2016-06-21 2017-12-19 Amazon Technologies, Inc. Low bandwidth video
JP2017228224A (en) 2016-06-24 2017-12-28 キヤノン株式会社 Information processing device, information processing method, and program
KR20180000255A (en) 2016-06-22 2018-01-02 엘지전자 주식회사 Mobile terminal and method for controlling the same
US20180027042A1 (en) 2015-12-15 2018-01-25 Line Corporation Method and system for video call using two-way communication of visual or auditory effect
US20180103234A1 (en) 2016-10-10 2018-04-12 Hyperconnect, Inc. Device and method of displaying images
US20180176641A1 (en) 2016-12-19 2018-06-21 Samsung Electronics Co., Ltd. Method and apparatus for filtering video
US20180260577A1 (en) 2017-03-08 2018-09-13 Blackberry Limited Data messaging with privacy tokens
US20180309801A1 (en) 2015-05-23 2018-10-25 Yogesh Chunilal Rathod Initiate call to present one or more types of applications and media up-to end of call
WO2018194243A1 (en) 2017-04-17 2018-10-25 Hyperconnect, Inc. Video communication device, video communication method, and video communication mediating method
US20180316892A1 (en) 2016-01-06 2018-11-01 Samsung Electronics Co., Ltd. Display apparatus and control methods thereof
WO2018221824A1 (en) 2017-05-27 2018-12-06 Hyperconnect, Inc. Video communication device and video communication method
KR20190007934A (en) 2017-07-14 2019-01-23 주식회사 하이퍼커넥트 Communication device, server and method of operation thereof
KR20190016671A (en) 2017-08-09 2019-02-19 주식회사 하이퍼커넥트 Communication device, server and communication method thereof
JP2019047357A (en) 2017-09-04 2019-03-22 キヤノン株式会社 Display unit with imaging apparatus
KR101989842B1 (en) 2017-12-07 2019-06-17 주식회사 하이퍼커넥트 Terminal and image processing method thereof
US20190188453A1 (en) 2017-12-15 2019-06-20 Hyperconnect, Inc. Terminal and server for providing video call service
US20190199963A1 (en) 2017-12-27 2019-06-27 Hyperconnect, Inc. Terminal and server for providing video call service
US20190238759A1 (en) 2018-01-31 2019-08-01 Hyperconnect, Inc. Terminal and image processing method thereof
US20190251118A1 (en) 2018-02-14 2019-08-15 Hyperconnect, Inc. Server and operating method thereof
US20190266444A1 (en) 2015-09-02 2019-08-29 Pocketguardian, Llc System and Method of Detecting Offensive Content Sent or Received on a Portable Electronic Device
US20190342246A1 (en) 2018-05-05 2019-11-07 Thomas P. Theriault Novel decency filtration tools and their use in interfacing systems
JP2020010158A (en) 2018-07-06 2020-01-16 本田技研工業株式会社 Provision method of peripheral image, and program for communication terminal
US20200213530A1 (en) 2018-12-31 2020-07-02 Hyperconnect, Inc. Terminal and server providing a video call service
US20200242750A1 (en) 2019-01-30 2020-07-30 Walmart Apollo, Llc Systems and methods for detecting offensive content in images using neural network architectures
US10791261B2 (en) 2014-10-02 2020-09-29 Apple Inc. Interactive video conferencing
US20200358904A1 (en) 2019-05-10 2020-11-12 Hyperconnect, Inc. Mobile, server and operating method thereof
US20210099672A1 (en) 2019-10-01 2021-04-01 Hyperconnect, Inc. Terminal and operating method thereof
US20210203882A1 (en) 2016-10-10 2021-07-01 Hyperconnect, Inc. Device and method of displaying images
US20210243407A1 (en) 2020-01-31 2021-08-05 Hyperconnect, Inc. Terminal and operating method thereof
US20210243408A1 (en) 2020-01-31 2021-08-05 Hyperconnect, Inc. Terminal, operating method thereof, and computer-readable recording medium

Family Cites Families (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0690279A (en) 1992-09-08 1994-03-29 Fujitsu General Ltd Mischievous call repulsing device
US20030028875A1 (en) 2001-05-02 2003-02-06 Koninklijke Philips Electronics N.V. Television access control system
US7444403B1 (en) 2003-11-25 2008-10-28 Microsoft Corporation Detecting sexually predatory content in an electronic communication
US20080059198A1 (en) 2006-09-01 2008-03-06 Pudding Ltd. Apparatus and method for detecting and reporting online predators
GB0709574D0 (en) 2007-05-18 2007-06-27 Aurix Ltd Speech Screening
US20090041311A1 (en) 2007-08-09 2009-02-12 Jon Hundley Facial recognition based content blocking system
JP2008159034A (en) 2007-11-07 2008-07-10 Fujitsu Ltd Server and voice signal collection/distribution method
US8230011B2 (en) 2008-03-03 2012-07-24 John Vandenburgh Networked computer alert system
US20090288131A1 (en) 2008-05-13 2009-11-19 Porto Technology, Llc Providing advance content alerts to a mobile device during playback of a media item
US8904547B2 (en) 2009-01-05 2014-12-02 International Business Machines Corporation Notification upon exposure to offensive behavioral patterns in collaboration
US8301653B2 (en) 2010-01-25 2012-10-30 Glenn Adamousky System and method for capturing and reporting online sessions
US8856165B1 (en) * 2010-03-26 2014-10-07 Google Inc. Ranking of users who report abuse
US9716599B1 (en) 2013-03-14 2017-07-25 Ca, Inc. Automated assessment of organization mood
US20160023116A1 (en) 2014-07-03 2016-01-28 Spitfire Technologies, Llc Electronically mediated reaction game
WO2017004241A1 (en) 2015-07-02 2017-01-05 Krush Technologies, Llc Facial gesture recognition and video analysis tool
KR20170059201A (en) 2015-11-20 2017-05-30 삼성전자주식회사 Electronic device and content ouputting method thereof
BR102016007265B1 (en) * 2016-04-01 2022-11-16 Samsung Eletrônica da Amazônia Ltda. MULTIMODAL AND REAL-TIME METHOD FOR FILTERING SENSITIVE CONTENT
US10104025B2 (en) 2016-05-23 2018-10-16 Oath Inc. Virtual chat rooms
US10349224B2 (en) 2017-01-24 2019-07-09 Essential Products, Inc. Media and communications in a connected environment
US10440063B1 (en) 2018-07-10 2019-10-08 Eturi Corp. Media device content review and management
US10440324B1 (en) 2018-09-06 2019-10-08 Amazon Technologies, Inc. Altering undesirable communication data for communication sessions
US20200335089A1 (en) 2019-04-16 2020-10-22 International Business Machines Corporation Protecting chat with artificial intelligence
US20220229488A1 (en) 2019-06-14 2022-07-21 Semiconductor Energy Laboratory Co., Ltd. Data Processing Device Executing Operation Based on User's Emotion
US11330313B2 (en) 2019-08-02 2022-05-10 Dell Products L.P. Crowd rating media content based on micro-expressions of viewers

Patent Citations (126)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH09200714A (en) 1996-01-17 1997-07-31 Casio Comput Co Ltd Image processor with video telephone
JP2001309325A (en) 2000-04-19 2001-11-02 Mitsubishi Electric Corp Video telephone system
US6590601B2 (en) 2000-04-19 2003-07-08 Mitsubishi Denki Kabushiki Kaisha Videophone apparatus with privacy protection
KR20010000426A (en) 2000-09-28 2001-01-05 김용환 Method of Intelligent Image Interface
US20030091239A1 (en) 2001-11-13 2003-05-15 Kazuyuki Imagawa Communications method using images and device for the same
JP2003219383A (en) 2001-11-13 2003-07-31 Matsushita Electric Ind Co Ltd Communications method using image and device for the same
US20030126267A1 (en) 2001-12-27 2003-07-03 Koninklijke Philips Electronics N.V. Method and apparatus for preventing access to inappropriate content over a network based on audio or visual content
US20090049467A1 (en) 2002-07-02 2009-02-19 Caption Tv, Inc. System, method and computer program product for selective filtering of objectionable content from a program
US20040049780A1 (en) 2002-09-10 2004-03-11 Jeanette Gee System, method, and computer program product for selective replacement of objectionable program content with less-objectionable content
KR20040064972A (en) 2003-01-13 2004-07-21 엘지전자 주식회사 televiewing limit method for display device
US7321384B1 (en) 2003-06-03 2008-01-22 Cisco Technology, Inc. Method and apparatus for using far end camera control (FECC) messages to implement participant and layout selection in a multipoint videoconference
US20100054592A1 (en) 2004-10-28 2010-03-04 Fotonation Ireland Limited Analyzing partial face regions for red-eye detection in acquired digital images
US8977063B2 (en) 2005-03-09 2015-03-10 Qualcomm Incorporated Region-of-interest extraction for video telephony
JP2006270380A (en) 2005-03-23 2006-10-05 Victor Co Of Japan Ltd Image information transmitter
US7564476B1 (en) 2005-05-13 2009-07-21 Avaya Inc. Prevent video calls based on appearance
US20100226261A1 (en) * 2005-09-16 2010-09-09 Eyeball Networks Inc. Method and system to prevent spam over internet telephony
US20080012935A1 (en) 2005-11-22 2008-01-17 Gateway Inc. Inappropriate content detection and distribution prevention for wireless cameras/camcorders with e-mail capabilities and camera phones
US20070195158A1 (en) 2006-02-23 2007-08-23 Kies Jonathan K Apparatus and methods for managing video calls on a wireless device
US20080259154A1 (en) 2007-04-20 2008-10-23 General Instrument Corporation Simulating Short Depth of Field to Maximize Privacy in Videotelephony
US20080298571A1 (en) 2007-05-31 2008-12-04 Kurtz Andrew F Residential video communication system
KR100844334B1 (en) 2007-06-07 2008-07-08 엘지전자 주식회사 Apparatus and method for controlling display in mobile communication terminal
KR20080110064A (en) 2007-06-14 2008-12-18 유성준 Method for distinguishing obscene image and apparatus therefor
US8165416B2 (en) 2007-06-29 2012-04-24 Microsoft Corporation Automatic gain and exposure control using region of interest detection
US20100017818A1 (en) 2008-07-17 2010-01-21 Cisco Technology, Inc. Non-bandwidth intensive method for providing multiple levels of censoring in an a/v stream
US20100124941A1 (en) 2008-11-19 2010-05-20 Samsung Electronics Co., Ltd. Method and device for synthesizing image
US8441356B1 (en) 2009-02-16 2013-05-14 Handhold Adaptive, LLC Methods for remote assistance of disabled persons
KR20110019499A (en) 2009-08-20 2011-02-28 엘지전자 주식회사 Mobile terminal
KR20110025720A (en) 2009-09-05 2011-03-11 에스케이텔레콤 주식회사 Method for video call based on an avatar and system, apparatus thereof
US20120182379A1 (en) 2009-09-24 2012-07-19 Zte Corporation Method, Application Server and System for Privacy Protection in Video Call
JP2011082694A (en) 2009-10-05 2011-04-21 Sharp Corp System and device for communication conference
KR20110073238A (en) 2009-12-21 2011-06-29 한국전자통신연구원 Apparatus and method for blocking the objectionable multimedia based on multimodal and multiscale features
US8626879B2 (en) 2009-12-22 2014-01-07 Sling Media, Inc. Systems and methods for establishing network connections using local mediation services
JP2011205243A (en) 2010-03-24 2011-10-13 Sharp Corp Information processing apparatus, conference system, information processing method, and computer program
US10127195B2 (en) 2010-05-17 2018-11-13 Amazon Technologies, Inc. Selective content presentation engine
JP2013531923A (en) 2010-05-17 2013-08-08 アマゾン テクノロジーズ インコーポレイテッド Selective content presentation engine
JP2012018571A (en) 2010-07-08 2012-01-26 Ntt Docomo Inc Content recommendation apparatus and method
JP2012054897A (en) 2010-09-03 2012-03-15 Sharp Corp Conference system, information processing apparatus, and information processing method
US20130147897A1 (en) 2010-09-10 2013-06-13 Shigehiro Ichimura Mobile terminal, remote operation system, data transmission control method by mobile terminal, and non-transitory computer readable medium
US20120155759A1 (en) 2010-12-21 2012-06-21 Microsoft Corporation Establishing clusters of user preferences for image enhancement
JP2012146209A (en) 2011-01-13 2012-08-02 Nikon Corp Electronic device and control program for electronic device
US20130234826A1 (en) 2011-01-13 2013-09-12 Nikon Corporation Electronic device and electronic device control program
WO2012131932A1 (en) 2011-03-30 2012-10-04 富士通株式会社 Information processing device including image transmission control function
US20130342633A1 (en) 2011-03-30 2013-12-26 Fujitsu Limited Information processing device including an image transmission control function
EP2515526A2 (en) 2011-04-08 2012-10-24 DigitalOptics Corporation Europe Limited Display device with image capture and analysis module
KR20120126677A (en) 2011-05-12 2012-11-21 재단법인 서울호서직업전문학교 Method and apparatus for automatically detecting pornographic image, and computer readable medium thereof
US20120320141A1 (en) 2011-06-16 2012-12-20 Vtel Products Corporation, Inc. Video conference control system and method
US20130083155A1 (en) 2011-09-30 2013-04-04 Cisco Technology Inc. Method, endpoint, and system for establishing a video conference
KR20120090870A (en) 2012-05-09 2012-08-17 김명훈 Image security apparatus and image security system using face detect algorithm and image security method thereof
KR20140012474A (en) 2012-07-20 2014-02-03 한국전자통신연구원 Apparatus and method for protecting privacy based on face recognition
US20170251336A1 (en) 2012-09-11 2017-08-31 Amazon Technologies, Inc. Content display controls based on environmental factors
US20150030314A1 (en) 2012-12-11 2015-01-29 Unify Gmbh & Co. Kg Method of processing video data, device, computer program product, and data construct
US20150070516A1 (en) 2012-12-14 2015-03-12 Biscotti Inc. Automatic Content Filtering
WO2014100774A1 (en) 2012-12-21 2014-06-26 Google Inc. Recommending transformations for photography
US20140176732A1 (en) 2012-12-21 2014-06-26 Google Inc. Recommending transformations for photography
KR20150087362A (en) 2012-12-21 2015-07-29 구글 인코포레이티드 Recommending transformations for photography
US9319632B2 (en) 2013-01-02 2016-04-19 Samsung Electronics Co., Ltd. Display apparatus and method for video calling thereof
US20140267583A1 (en) 2013-03-13 2014-09-18 Futurewei Technologies, Inc. Augmented Video Calls on Mobile Devices
EP2782326A2 (en) 2013-03-20 2014-09-24 Samsung Electronics Co., Ltd. Method and apparatus for processing an image based on an image property and ambient environment information
US9230328B1 (en) 2013-06-28 2016-01-05 Google Inc. Providing image parameters
EP3035283A1 (en) 2013-08-15 2016-06-22 Xiaomi Inc. Image processing method and apparatus, and terminal device
US9591261B2 (en) 2013-09-24 2017-03-07 Hitachi Maxell, Ltd. Video telephone device and video telephone processing method
WO2015091487A1 (en) 2013-12-19 2015-06-25 Alcatel Lucent Method for real-time reporting and prevention of call abuse
US9460340B2 (en) 2014-01-31 2016-10-04 Google Inc. Self-initiated change of appearance for subjects in video and images
US20150220777A1 (en) 2014-01-31 2015-08-06 Google Inc. Self-initiated change of appearance for subjects in video and images
US20150229882A1 (en) 2014-02-10 2015-08-13 Alibaba Group Holding Limited Video communication method and system in instant communication
WO2015162647A1 (en) 2014-04-25 2015-10-29 Sony Corporation Processing digital photographs in response to external applications
US20160316265A1 (en) * 2014-04-29 2016-10-27 At&T Intellectual Property I, Lp Method and apparatus for augmenting media content
US20150325023A1 (en) 2014-05-09 2015-11-12 Google Inc. Providing pre-edits for photos
US20150370474A1 (en) 2014-06-19 2015-12-24 BrightSky Labs, Inc. Multiple view interface for video editing system
US20160373571A1 (en) 2014-06-19 2016-12-22 Microsoft Technology Licensing, Llc Use of a digital assistant in communications
US9225897B1 (en) 2014-07-07 2015-12-29 Snapchat, Inc. Apparatus and method for supplying content aware photo filters
US20160019412A1 (en) 2014-07-18 2016-01-21 Htc Corporation Method for performing a face tracking function and an electric device having the same
US20160034786A1 (en) 2014-07-29 2016-02-04 Microsoft Corporation Computerized machine learning of interesting video sections
US9282287B1 (en) 2014-09-09 2016-03-08 Google Inc. Real-time video transformations in video conferences
US10791261B2 (en) 2014-10-02 2020-09-29 Apple Inc. Interactive video conferencing
US20160127653A1 (en) 2014-11-03 2016-05-05 Samsung Electronics Co., Ltd. Electronic Device and Method for Providing Filter in Electronic Device
US9792710B2 (en) 2014-12-25 2017-10-17 Seiko Epson Corporation Display device, and method of controlling display device
JP2016122975A (en) 2014-12-25 2016-07-07 セイコーエプソン株式会社 Display device and display device control method
WO2016112299A1 (en) 2015-01-09 2016-07-14 Snapchat, Inc. Object recognition based photo filters
US9503456B1 (en) * 2015-01-22 2016-11-22 Google Inc. Video chat abuse detection based on external context
US9380264B1 (en) * 2015-02-16 2016-06-28 Siva Prasad Vakalapudi System and method for video communication
US9569697B1 (en) 2015-02-19 2017-02-14 Google Inc. Object oriented image editing
US20150286858A1 (en) 2015-03-18 2015-10-08 Looksery, Inc. Emotion recognition in video conferencing
US20180309801A1 (en) 2015-05-23 2018-10-25 Yogesh Chunilal Rathod Initiate call to present one or more types of applications and media up-to end of call
US20160350675A1 (en) * 2015-06-01 2016-12-01 Facebook, Inc. Systems and methods to identify objectionable content
US20190266444A1 (en) 2015-09-02 2019-08-29 Pocketguardian, Llc System and Method of Detecting Offensive Content Sent or Received on a Portable Electronic Device
US20180027042A1 (en) 2015-12-15 2018-01-25 Line Corporation Method and system for video call using two-way communication of visual or auditory effect
US20180316892A1 (en) 2016-01-06 2018-11-01 Samsung Electronics Co., Ltd. Display apparatus and control methods thereof
US9848167B1 (en) 2016-06-21 2017-12-19 Amazon Technologies, Inc. Low bandwidth video
KR20180000255A (en) 2016-06-22 2018-01-02 엘지전자 주식회사 Mobile terminal and method for controlling the same
JP2017228224A (en) 2016-06-24 2017-12-28 キヤノン株式会社 Information processing device, information processing method, and program
US11102450B2 (en) 2016-10-10 2021-08-24 Hyperconnect, Inc. Device and method of displaying images
JP2020500488A (en) 2016-10-10 2020-01-09 ハイパーコネクト インコーポレイテッド Image display device and image display method (DEVICE AND METHOD OF DISPLAYING IMAGES)
US20210203882A1 (en) 2016-10-10 2021-07-01 Hyperconnect, Inc. Device and method of displaying images
US20180103234A1 (en) 2016-10-10 2018-04-12 Hyperconnect, Inc. Device and method of displaying images
US20180176641A1 (en) 2016-12-19 2018-06-21 Samsung Electronics Co., Ltd. Method and apparatus for filtering video
US20180260577A1 (en) 2017-03-08 2018-09-13 Blackberry Limited Data messaging with privacy tokens
KR20180116520A (en) 2017-04-17 2018-10-25 주식회사 하이퍼커넥트 Device and method of making video calls and method of mediating video calls
WO2018194243A1 (en) 2017-04-17 2018-10-25 Hyperconnect, Inc. Video communication device, video communication method, and video communication mediating method
US10965907B2 (en) 2017-04-17 2021-03-30 Hyperconnect, Inc. Video communication device, video communication method, and video communication mediating method
US20200137352A1 (en) 2017-04-17 2020-04-30 Hyperconnect, Inc. Video communication device, video communication method, and video communication mediating method
WO2018221824A1 (en) 2017-05-27 2018-12-06 Hyperconnect, Inc. Video communication device and video communication method
KR20190007934A (en) 2017-07-14 2019-01-23 주식회사 하이퍼커넥트 Communication device, server and method of operation thereof
KR20190016671A (en) 2017-08-09 2019-02-19 주식회사 하이퍼커넥트 Communication device, server and communication method thereof
JP2019047357A (en) 2017-09-04 2019-03-22 キヤノン株式会社 Display unit with imaging apparatus
KR101989842B1 (en) 2017-12-07 2019-06-17 주식회사 하이퍼커넥트 Terminal and image processing method thereof
US20190188453A1 (en) 2017-12-15 2019-06-20 Hyperconnect, Inc. Terminal and server for providing video call service
US10574941B2 (en) 2017-12-27 2020-02-25 Hyperconnect, Inc. Terminal and server for providing video call service
US20200145613A1 (en) 2017-12-27 2020-05-07 Hyperconnect, Inc. Terminal and server for providing video call service
US11140356B2 (en) 2017-12-27 2021-10-05 Hyperconnect, Inc. Terminal and server for providing video call service
US20190199963A1 (en) 2017-12-27 2019-06-27 Hyperconnect, Inc. Terminal and server for providing video call service
US20190238759A1 (en) 2018-01-31 2019-08-01 Hyperconnect, Inc. Terminal and image processing method thereof
US10855933B2 (en) 2018-01-31 2020-12-01 Hyperconnect, Inc. Terminal and image processing method thereof
US20190251118A1 (en) 2018-02-14 2019-08-15 Hyperconnect, Inc. Server and operating method thereof
US11080325B2 (en) 2018-02-14 2021-08-03 Hyperconnect, Inc. Server and operating method thereof
US20190342246A1 (en) 2018-05-05 2019-11-07 Thomas P. Theriault Novel decency filtration tools and their use in interfacing systems
JP2020010158A (en) 2018-07-06 2020-01-16 本田技研工業株式会社 Provision method of peripheral image, and program for communication terminal
US20200213530A1 (en) 2018-12-31 2020-07-02 Hyperconnect, Inc. Terminal and server providing a video call service
US11206362B2 (en) 2018-12-31 2021-12-21 Hyperconnect, Inc. Terminal and server providing a video call service
US20200242750A1 (en) 2019-01-30 2020-07-30 Walmart Apollo, Llc Systems and methods for detecting offensive content in images using neural network architectures
US20200358904A1 (en) 2019-05-10 2020-11-12 Hyperconnect, Inc. Mobile, server and operating method thereof
EP3800879A1 (en) 2019-10-01 2021-04-07 Hyperconnect, Inc. Terminal and operating method thereof
US20210099672A1 (en) 2019-10-01 2021-04-01 Hyperconnect, Inc. Terminal and operating method thereof
US11184582B2 (en) 2019-10-01 2021-11-23 Hyperconnect, Inc. Terminal and operating method thereof
US20210243407A1 (en) 2020-01-31 2021-08-05 Hyperconnect, Inc. Terminal and operating method thereof
US20210243408A1 (en) 2020-01-31 2021-08-05 Hyperconnect, Inc. Terminal, operating method thereof, and computer-readable recording medium

Non-Patent Citations (29)

* Cited by examiner, † Cited by third party
Title
"Talking To Complete Strangers On Holla App! Awkward" YouTube, Life With Nae, Published Jul. 21, 2018, Available online at https://www.youtube.com/watch?v=F9ddBJ4yJZA, 8 pgs.
Communication dated Apr. 16, 2018, issued by the Korean Intellectual Property Office in application No. 10-2017-0048893.
Communication in European Patent Office in Application No. 20173922.4, dated Oct. 9, 2020, 7 Pgs.
Communication in Korean Patent Office in Application No. 10-2018-0174192, dated Nov. 25, 2019, 9 Pgs.
Communication in Korean Patent Office in Application No. 10-2019-0054670, dated Apr. 21, 2020,11 Pgs.
Communication issued in Korean Patent Office in Application No. 10-2018-0012389, dated Jul. 31, 2019, 11 Pgs.
Dou et al., "Coarse-to-Fine Trained Multi-Scale Convolutional Neural Networks for Image Classification", IEEE, 2015, 7 pgs.
Extended European Search Report for Application No. 18213173.0, dated May 7, 2019, 8 Pgs.
Extended European Search Report for European Application No. 19156400.4, Search completed Apr. 11, 2019, dated Apr. 24, 2019, 11 Pgs.
Extended European Search Report for European Application No. 20199235.1, Search completed Feb. 10, 2021, dated Feb. 19, 2021, 7 Pgs.
Extended European Search Report for European Application No. 21154225.3, Search completed Apr. 20, 2021, dated Apr. 29, 2021, 7 Pgs.
Extended Search Report for European Application No. 17859740.7, Search completed Sep. 16, 2019, dated Sep. 24, 2019, 10 Pgs.
Extended Search Report for European Application No. 21154208.9, Search completed May 21, 2021, dated Jun. 4, 2021, 9 Pgs.
Freitas et al., "A Multimodal CNN-Based Tool to Censure Inappropriate Video Scenes", arXiv:1911.03974v1 [cs.MM], Nov. 10, 2019, 7 pgs.
International Search Report dated Jan. 18, 2018 in International Application No. PCT/KR2017/011139, 3 pages.
International Search Report for PCT/KR2018/000787 dated Apr. 26, 2018 [PCT/ISA/210].
Korean Office Action for Application No. 10-2017-0181512, dated Mar. 8, 2019, 8 Pgs.
Korean Office Action for Application No. 10-2019-0121524, dated Sep. 22, 2020, 9 Pgs.
Korean Office Action of KR 10-2020-0012270 dated Nov. 30, 2020, 6 pgs.
Office Action dated Aug. 29, 2017 of Korean Patent Application No. 10-2016-0130656 which is the parent Application—6 pages.
Office Action dated Feb. 14, 2017 of Korean Patent Application No. 10-2016-0130656 which is the parent Application—5 pages.
Office Action dated Nov. 30, 2018 in Korean Application No. 10-2018-0134869, in 5 pages.
Office Action for Japanese Patent Application No. 2020-165934, dated Sep. 24, 2021, 12 pgs.
Office Action for Japanese Patent Application No. 2021013320 dated Jan. 19, 2022, 6 pgs.
Office Action issued from Korean Patent Application No. 10-2020-0012233, dated Nov. 20, 2020, 5 Pgs.
Office Action issued in Japanese Patent Application No. 2018-238541, dated Dec. 10, 2019, 12 Pgs.
Sood et al., "Automatic Identification of Personal Insults on Social News Sites", Journal of The American Society for Information Science and Technology, Feb. 1, 2012, vol. 63, No. 2, pp. 270-285. Published Online Oct. 28, 2011.
Translation of Office Action dated Jun. 29, 2020 in Japanese Application No. 2019-540499, in 3 pages.
Xing et al., "SafeVchat: Detecting Obscene Content and Misbehaving Users in Online Video Chat Services", World Wide Web, ACM, Mar. 28, 2011, pp. 685-694.

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11722638B2 (en) 2017-04-17 2023-08-08 Hyperconnect Inc. Video communication device, video communication method, and video communication mediating method
US11854263B2 (en) * 2018-07-23 2023-12-26 Tencent Technology (Shenzhen) Company Limited Video processing method and apparatus, terminal device, server, and storage medium
US11716424B2 (en) 2019-05-10 2023-08-01 Hyperconnect Inc. Video call mediation method
US11496709B2 (en) 2020-01-31 2022-11-08 Hyperconnect Inc. Terminal, operating method thereof, and computer-readable recording medium
US11825236B2 (en) 2020-01-31 2023-11-21 Hyperconnect Inc. Terminal and operating method thereof
US20220141396A1 (en) * 2020-10-29 2022-05-05 Acer Incorporated Video conferencing system and method of removing interruption thereof
US11812185B2 (en) * 2020-10-29 2023-11-07 Acer Incorporated Video conferencing system and method of removing interruption thereof
US20220239848A1 (en) * 2021-01-26 2022-07-28 Dell Products, Lp System and method for operating an intelligent videoframe privacy monitoring management system for videoconferencing applications
US11838684B2 (en) * 2021-01-26 2023-12-05 Dell Products, Lp System and method for operating an intelligent videoframe privacy monitoring management system for videoconferencing applications

Also Published As

Publication number Publication date
KR101932844B1 (en) 2018-12-27
KR20180116520A (en) 2018-10-25
US20210203880A1 (en) 2021-07-01
US11722638B2 (en) 2023-08-08
US10965907B2 (en) 2021-03-30
WO2018194243A1 (en) 2018-10-25
US20220239862A1 (en) 2022-07-28
US20200137352A1 (en) 2020-04-30

Similar Documents

Publication Publication Date Title
US11323659B2 (en) Video communication device, video communication method, and video communication mediating method
US10701315B2 (en) Video communication device and video communication method
KR102311603B1 (en) Mobile and operating method thereof
KR102283972B1 (en) Communication device, server and method of operation thereof
KR102282963B1 (en) Mobile, server and operating method thereof
US11553157B2 (en) Device and method of displaying images
KR101979650B1 (en) Server and operating method thereof
CN107845062A (en) image generating method and device
CN108182626A (en) Service push method, information acquisition terminal and computer readable storage medium
WO2021047069A1 (en) Face recognition method and electronic terminal device
KR20190016671A (en) Communication device, server and communication method thereof
CN109842612A (en) Log security analysis method, device and storage medium based on picture library model
CN115497159A (en) Human body abnormal state detection method and device, computer equipment and storage medium
WO2018107031A1 (en) Personal emergency data capture and alerting
JP7275181B2 (en) Declaration evaluation device and method of operation thereof
KR102293416B1 (en) Communication device, server and communication method thereof
US10447968B1 (en) Controlled-environment facility video communications monitoring system
KR102051828B1 (en) Method of making video communication and device of mediating video communication
KR102315211B1 (en) Mobile and operating method thereof
CN110738812A (en) Bell prompting method and related product
US11984138B1 (en) Method for monitoring emotion and behavior during conversation for user in need of protection
CN113535926B (en) Active dialogue method and device and voice terminal
Liau IoT Based Personal Safety And Protection System
CN113918822A (en) Content searching method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

AS Assignment

Owner name: HYPERCONNECT, INC., KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:AHN, SANGIL;JUNG, KANGSIK;YONG, HYOUNTAEK;REEL/FRAME:056435/0052

Effective date: 20191016

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: AWAITING TC RESP., ISSUE FEE NOT PAID

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STPP Information on status: patent application and granting procedure in general

Free format text: AWAITING TC RESP, ISSUE FEE PAYMENT VERIFIED

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: HYPERCONNECT INC., KOREA, REPUBLIC OF

Free format text: CONFIRMATION OF ASSIGNMENT;ASSIGNOR:YONG, HYOUN TAEK;REEL/FRAME:061975/0837

Effective date: 20220820

AS Assignment

Owner name: HYPERCONNECT INC., KOREA, REPUBLIC OF

Free format text: CONFIRMATION OF ASSIGNMENT;ASSIGNOR:JUNG, KANG SIK;REEL/FRAME:062005/0483

Effective date: 20220919

AS Assignment

Owner name: HYPERCONNECT INC., KOREA, REPUBLIC OF

Free format text: CONFIRMATION OF ASSIGNMENT;ASSIGNOR:AHN, SANG IL;REEL/FRAME:062026/0401

Effective date: 20221104