US20240168701A1 - Head-mounted display control apparatus, method for controlling head-mounted display control apparatus, and non-transitory computer readable storage medium - Google Patents

Head-mounted display control apparatus, method for controlling head-mounted display control apparatus, and non-transitory computer readable storage medium Download PDF

Info

Publication number
US20240168701A1
US20240168701A1 US18/511,788 US202318511788A US2024168701A1 US 20240168701 A1 US20240168701 A1 US 20240168701A1 US 202318511788 A US202318511788 A US 202318511788A US 2024168701 A1 US2024168701 A1 US 2024168701A1
Authority
US
United States
Prior art keywords
display
image
head
phone call
control apparatus
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US18/511,788
Inventor
Masaaki Kobayashi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Inc
Original Assignee
Canon Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Inc filed Critical Canon Inc
Assigned to CANON KABUSHIKI KAISHA reassignment CANON KABUSHIKI KAISHA ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KOBAYASHI, MASAAKI
Publication of US20240168701A1 publication Critical patent/US20240168701A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/14Digital output to display device ; Cooperation and interconnection of the display device with other functional units
    • G06F3/147Digital output to display device ; Cooperation and interconnection of the display device with other functional units using display panels
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2354/00Aspects of interface with display user
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72409User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality by interfacing with external accessories
    • H04M1/724094Interfacing with a device worn on the user's body to provide access to telephonic functionalities, e.g. accepting a call, reading or composing a message
    • H04M1/724097Worn on the head
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/239Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/332Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
    • H04N13/344Displays for viewing with the aid of special glasses or head-mounted displays [HMD] with head-mounted left-right displays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/30Image reproducers
    • H04N13/398Synchronisation thereof; Control thereof

Definitions

  • the present disclosure relates to a head-mounted display control apparatus, a method for controlling the head-mounted display control apparatus, and a non-transitory computer readable storage medium.
  • XR information processing terminals of a head-mounted display (HMD) type have become widespread.
  • XR is a generic term for virtual reality (VR), augmented reality (AR), and mixed reality (MR).
  • HMD XR information processing terminals include those of a see-through type in which computer graphics (CG) are displayed on a transmissive display, and those of a non-see-through type in which CG and a real image captured by a camera are displayed on a non-transmissive display.
  • CG computer graphics
  • non-see-through type in which CG and a real image captured by a camera are displayed on a non-transmissive display.
  • CG computer graphics
  • non-see-through display a person wearing an HMD is able to determine the state of a real space.
  • the person is unable to determine the state of a real space but is highly immersed in a VR space.
  • there is a non-see-through display in which the state of a real space can be determined by displaying an image captured by a camera built in the HMD on the display.
  • a state in which a real space is displayed is called a video see-through mode, relative to a state in which a VR space is displayed.
  • a smartphone compatible with the headset profile (HSP) or the hands-free profile (HFP) of Bluetooth (registered trademark) and an audio terminal such as a headset enable a phone conversation with the audio terminal instead of a microphone and a speaker of the smartphone.
  • the HSP is the specifications of the function of transmitting and receiving audio data between a smartphone and an audio terminal.
  • the HFP is the specifications of the function of making and receiving phone calls.
  • U.S. patent Ser. No. 10/324,294 discloses that the reception of a message or a phone call in an external smartphone is indicated by an icon in an HMD.
  • a head-mounted display control apparatus includes one or more circuits, or one or more processors and at least one memory, the at least one memory being coupled to the one or more processors and having stored thereon instructions executable by the one or more processors. At least one of the one or more circuits or the execution of the instructions causes the head-mounted display control apparatus to function as an image capturing unit configured to capture an image; and a control unit configured to, based on phone call reception information, phone conversation information, or message reception information, perform control to display the captured image or a virtual space image on a display unit.
  • FIG. 1 is a diagram illustrating an example configuration of a head-mounted display (HMD).
  • HMD head-mounted display
  • FIG. 2 is a flowchart illustrating a method for controlling the HMD according to a first embodiment.
  • FIG. 3 is a flowchart illustrating a method for controlling the HMD according to a second embodiment.
  • FIG. 4 is a flowchart illustrating a method for controlling the HMD according to a third embodiment.
  • FIGS. 5 A and 5 B are each a diagram for describing a user interface screen displayed when a phone call is being received.
  • the present disclosure provides a technique of enabling a captured image to be displayed, for example, when a phone call is being received.
  • FIG. 1 is a diagram illustrating an example configuration of a head-mounted display (HMD) 120 according to a first embodiment.
  • the HMD 120 is an example of a display control apparatus that responds to a phone call reception signal.
  • the HMD 120 includes a bus 100 , a central processing unit (CPU) 101 , a graphics processing unit (GPU) 102 , a posture sensor 103 , an analog-to-digital/digital-to-analog (AD/DA) converter 104 for audio, and a wireless unit 105 .
  • CPU central processing unit
  • GPU graphics processing unit
  • D/DA analog-to-digital/digital-to-analog
  • the HMD 120 further includes an image processor 106 , a display 107 , a microphone 108 , a speaker 109 , a camera unit 110 , a random access memory (RAM) 111 , and a nonvolatile memory 112 .
  • an image processor 106 a display 107 , a microphone 108 , a speaker 109 , a camera unit 110 , a random access memory (RAM) 111 , and a nonvolatile memory 112 .
  • RAM random access memory
  • the bus 100 manages the data flow in the HMD 120 .
  • the CPU 101 executes a program stored in the nonvolatile memory 112 or the like to control the entire HMD 120 , and executes system software and application software.
  • the CPU 101 analyzes a digital audio signal to determine whether the digital audio signal is a phone call reception sound.
  • the GPU 102 generates a virtual reality image and outputs data of the image to the display 107 .
  • the posture sensor 103 includes a gyro sensor and an acceleration sensor, and acquires information regarding the posture and position of the HMD 120 .
  • the CPU 101 processes raw data received from the posture sensor 103 and acquires information regarding the posture and position of the HMD 120 .
  • the information regarding the posture and position is reflected in generation of a virtual reality image. That is, the HMD 120 generates a virtual reality image corresponding to the forward direction of the HMD 120 in accordance with a motion of the HMD 120 including the posture sensor 103 , and displays the virtual reality image on the display 107 .
  • the AD/DA converter 104 converts an analog audio signal received from the microphone 108 into a digital audio signal, converts the digital audio signal into an analog audio signal, and outputs the analog audio signal to the speaker 109 .
  • the wireless unit 105 is capable of connecting to another terminal by using a plurality of wireless standards (Bluetooth, Wi-Fi, and the like) to transmit and receive data. Specifically, the wireless unit 105 connects to a controller device to acquire operation information, or connects to a smartphone to transmit and receive audio data and acquire phone call reception information.
  • the image processor 106 acquires mosaic image data from the camera unit 110 and performs demosaicing and color correction processing.
  • the display 107 displays a real space image captured by the camera unit 110 or a virtual reality image generated by the GPU 102 . Which is to be displayed on the display 107 is controlled by the CPU 101 . The logic of the control will be described below.
  • the display 107 includes two panels, one is for the right eye and the other is for the left eye, and displays an input stereo image on the panels such that the stereo image is divided into partial images for the respective panels. A monocular image can also be input to the display 107 . In this case, the same image is displayed on the panels for the right eye and the left eye. Switching of the image to be displayed on the display 107 is controlled by the CPU 101 .
  • the microphone 108 physically acquires a sound and converts the sound into an analog electric signal.
  • the speaker 109 outputs an analog signal as a physical sound.
  • the camera unit 110 includes two lenses and two sensors, is disposed so as to be capable of capturing images for the positions of the two eyes of the user wearing the HMD 120 , and outputs captured and digitalized mosaic image data to the image processor 106 .
  • the image processor 106 combines the images received from the two sensors of the camera unit 110 into one stereo image.
  • the image processor 106 may correct a display image so that the user's perception of depth coincides with the reality.
  • the RAM 111 loads a program and data stored in the nonvolatile memory 112 at the startup of the HMD 120 and functions as a work area for the CPU 101 .
  • the RAM 111 stores an acquired image and a generated image.
  • the nonvolatile memory 112 stores a program and data. These operations are successively performed, and the user is able to view a moving image.
  • FIG. 2 is a flowchart illustrating a method for controlling the HMD 120 according to the first embodiment, and describes a phone call reception operation of the HMD 120 that responds to reception of a phone call.
  • the steps are performed in the order indicated by the arrows in FIG. 2 unless otherwise stated. If the input and output of data in the individual steps do not depend on each other, the steps may be performed in a different order.
  • the individual steps are performed by the CPU 101 , but part of processing, such as rendering processing, is performed by another processor such as the GPU 102 . In the present embodiment, a description will be given under the assumption that a VR game application is being executed.
  • step S 2010 the CPU 101 acquires phone call reception/phone conversation state information via the wireless unit 105 .
  • a smartphone connected to the HMD 120 via the wireless unit 105 transmits phone call reception/phone conversation state information to the HMD 120 , and thereby the CPU 101 acquires the phone call reception/phone conversation state information via the wireless unit 105 .
  • step S 2020 the CPU 101 determines, based on the phone call reception/phone conversation state information, whether a phone call is being received or a phone conversation is being performed. If a determination is made that a phone call is being received or a phone conversation is being performed, the process proceeds to step S 2030 . Otherwise, the process proceeds to step S 2060 .
  • step S 2030 the CPU 101 suspends the processing of the above-described application that is being executed.
  • the CPU 101 causes the game application to pause and causes the system to perform control.
  • step S 2040 the CPU 101 acquires a real space image by using the camera unit 110 and the image processor 106 .
  • the real space image is captured by the camera unit 110 built in the HMD 120 and is a stereo image of a region in front of the HMD 120 .
  • step S 2050 the CPU 101 renders the real space image. Specifically, the CPU 101 converts the format of the image acquired in step S 2040 into a format suitable for the display 107 , and transmits the converted stereo image data to the display 107 via the GPU 102 . The stereo image is displayed on the display 107 .
  • step S 2050 the CPU 101 also performs processing of waiting in synchronization with a video signal. Subsequently, the process proceeds to step S 2080 .
  • step S 2060 the CPU 101 executes the processing of the above-described application.
  • the CPU 101 executes processing for one frame and creates application data that is necessary for generating rendering data.
  • the application data may be acquired in synchronization with the processing of the other thread.
  • step S 2070 the CPU 101 renders a virtual space image that is based on the application processing.
  • the CPU 101 transmits, based on the application data generated in step S 2050 and information from the posture sensor 103 , an image of a virtual space in front to the display 107 via the GPU 102 .
  • the virtual space image is displayed on the display 107 .
  • the CPU 101 also performs processing of waiting in synchronization with a video signal.
  • the virtual space image may be a stereo image that is based on computer graphics or may be a stereo image actually captured by using a stereo fish eye lens. Subsequently, the process proceeds to step S 2080 .
  • step S 2080 the CPU 101 determines whether to end the application. If a determination is made that the application is not to be ended, the process returns to step S 2010 . If a determination is made that the application is to be ended, the process of the flowchart illustrated in FIG. 2 ends.
  • the user holds a controller in his/her hand and does not hold a smartphone when playing a game.
  • the CPU 101 in response to receipt of a phone call in the real space, the CPU 101 proceeds to step S 2030 , automatically performs switching to a video see-through mode, and displays a real space image. Accordingly, the user is able to view the real space, immediately find his/her smartphone around him/her, operate the smartphone, and answer the phone call, without taking off the HMD 120 .
  • the user is able to easily determine whether to answer the phone call or whether to avoid an important topic in a phone conversation by determining whether there is somebody around him/her.
  • the microphone and speaker used for a phone conversation by the user are not limited.
  • the user may use the microphone 108 and the speaker 109 of the HMD 120 , and the HMD 120 may transmit/receive audio information to/from the smartphone.
  • the user may use a microphone and a speaker built in the smartphone to perform a phone conversation.
  • the CPU 101 may display a phone call answering screen on the display 107 of the HMD 120 and may transmit/receive audio information to/from the smartphone.
  • the user performs a phone conversation by using the microphone 108 and the speaker 109 of the HMD 120 .
  • the target application is a game application.
  • the target application is not limited thereto and may be any application, such as a VR moving image player, a work support application, or a virtual office application.
  • the CPU 101 decodes video data to generate a frame image and renders the frame image as a virtual reality image.
  • the CPU 101 may perform switching to the video see-through mode.
  • the phone call that is received may be a phone call received by either a smartphone connected to a public line or a phone conversation application of the system.
  • a screen is not automatically switched in response to receipt of a phone call within the application.
  • a display image is switched in accordance with only whether a phone call is being received.
  • the HMD 120 is equipped with various sensors and a camera.
  • the CPU 101 may use information obtained therefrom to detect whether there is a person near the user, may proceed to step S 2030 if a phone call is being received and there is a person near the user, and may switch the display of a real space image.
  • the CPU 101 generates a virtual reality image corresponding to the forward direction of the HMD 120 in accordance with a motion of the HMD 120 including the posture sensor 103 .
  • the posture sensor 103 need not be included, and an image in a fixed direction may be displayed.
  • a stereo image is handled in the present embodiment, a monocular image may be handled instead of a stereo image.
  • the CPU 101 performs switching to display of a real space image in response to receipt of a phone call, but the present disclosure is not limited thereto.
  • the CPU 101 may display a virtual space image and a real space image in combination, or may make a small window on a virtual space image in response to receipt of a phone call and display a real space image in the video see-through mode only in the small window.
  • the camera unit 110 and the image processor 106 function as an image capturing unit, which captures an image.
  • the display 107 is an example of a display unit.
  • the CPU 101 acquires phone call reception information, phone conversation information, or message reception information from an external apparatus.
  • the external apparatus is a smartphone, for example.
  • the CPU 101 functions as a control unit, and performs control to display a captured image or a virtual space image on the display 107 , based on the phone call reception information, the phone conversation information, or the message reception information.
  • the CPU 101 When a phone call is not being received or a phone conversation is not being performed, the CPU 101 performs control to display a virtual space image on the display 107 in step S 2070 . When a phone call is being received or a phone conversation is being performed, the CPU 101 performs control to display a captured image on the display 107 in step S 2050 .
  • the CPU 101 may perform control to display a captured image and a virtual space image on the display 107 .
  • the CPU 101 is capable of displaying a captured image and a virtual space image in combination, or displaying a captured image in a small window on a virtual space image.
  • the present embodiment may be applied to reception of a message in a smartphone.
  • the CPU 101 may perform control to display a virtual space image on the display 107 .
  • the CPU 101 may perform control to display a captured image on the display 107 for a predetermined period.
  • the CPU 101 may perform control to display a captured image and a virtual space image on the display 107 for a predetermined period, similarly to the above.
  • the CPU 101 may perform control to display a captured image or a virtual space image on the display 107 , based on phone call reception information, phone conversation information, or message reception information.
  • the user is able to determine the situation of a real space and immediately determine whether he/she can start a phone conversation, without taking off the HMD 120 .
  • FIG. 3 is a flowchart illustrating a method for controlling the HMD 120 according to a second embodiment, and describes a phone call reception operation of the HMD 120 that responds to reception of a phone call.
  • FIG. 3 is different from FIG. 2 in that steps S 3000 , S 3005 , and S 3010 are provided instead of steps S 2010 and S 2020 . The operation is the same as in the first embodiment unless otherwise stated.
  • step S 3000 the CPU 101 acquires audio information via the microphone 108 and the AD/DA converter 104 .
  • the audio information is acquired as digital audio data, which is generated by AD converting a sound received by the microphone 108 built in the HMD 120 by using the AD/DA converter 104 .
  • the CPU 101 acquires audio data for 10 seconds before the time point at which step S 3000 is performed.
  • step S 3005 the CPU 101 analyzes the audio information acquired in step S 3000 to generate an audio information analysis result.
  • the CPU 101 performs fast Fourier transform (FFT) on the audio information to acquire information in a range of 15 Hz to 16 kHz as an analysis result.
  • FFT fast Fourier transform
  • step S 3010 the CPU 101 determines whether the audio information analysis result generated in step S 3005 indicates a phone call reception sound. If the audio information analysis result indicates a phone call reception sound, the process proceeds to step S 2030 . Otherwise, the process proceeds to step S 2060 .
  • the sound pressure of the peak of at least one frequency is 60 dB or more as a result of FFT, the CPU 101 determines that the audio information analysis result indicates a phone call reception sound.
  • the peak may be determined in the following manner. For example, if there is an amplitude ten times the median of the amplitude in a Log scale of the FFT result, it is determined that there is a peak.
  • the CPU 101 is capable of detecting a phone call reception sound of a smartphone by using the microphone 108 in an analog manner and performing switching to the video see-through mode in step S 2050 , even if the CPU 101 is incapable of digitally transmitting/receiving a signal via the wireless unit 105 .
  • the present embodiment is capable of obtaining a merit similar to that of the first embodiment.
  • a result of FFT in the range of 15 Hz to 10 kHz is regarded as an audio analysis result, but the present disclosure is not limited thereto.
  • the number of successive identical patterns may be calculated with respect to a time-series change in waveform, and if the number is two or more, the successive identical patterns may be determined to be a phone call reception sound.
  • a warning sound such as a fire alarm sound often has audio information having repetition of an identical pattern, and causes automatic switching to the video see-through mode. This is an effective operation in terms of being able to notice the occurrence of a warning sound and visually recognize the cause of the occurrence.
  • a phone call reception sound may be recorded in advance, and whether a sound is identical to the recorded phone call reception sound may be determined. Accordingly, even if a melody is set as a phone call reception sound, the melody can be determined to be a phone call reception sound.
  • the CPU 101 performs control to display a captured image or a virtual space image on the display 107 , based on a phone call reception sound or a message reception sound collected by the microphone 108 .
  • FIG. 4 is a flowchart illustrating a method for controlling the HMD 120 according to a third embodiment, and describes a phone call reception operation of the HMD 120 that responds to reception of a phone call.
  • FIG. 4 is different from FIG. 2 in that step S 2020 is deleted and steps S 4000 , S 4011 , S 4012 , and S 4071 to S 4073 are added. The operation is the same as in the first embodiment unless otherwise stated.
  • step S 4000 the CPU 101 sets the video see-through mode to OFF. This corresponds to setting of an initial state.
  • step S 2010 the CPU 101 acquires phone call reception/phone conversation state information via the wireless unit 105 .
  • step S 4011 the CPU 101 acquires a video see-through mode. Specifically, the CPU 101 acquires the video see-through mode set in step S 4000 described above or step S 4073 described below.
  • step S 4012 the CPU 101 determines whether the video see-through mode is ON. If the video see-through mode is ON, the process proceeds to step S 4071 via steps S 2030 to S 2050 . If the video see-through mode is OFF, the process proceeds to step S 4071 via steps S 2060 and S 2070 .
  • step S 4071 the CPU 101 determines, based on the phone call reception/phone conversation state information acquired in step S 2010 , whether a phone call is being received or a phone conversation is being performed. If a determination is made that a phone call is being received or a phone conversation is being performed, the process proceeds to step S 4072 . Otherwise, the process proceeds to step S 2080 .
  • step S 4072 the CPU 101 renders a video see-through mode selection screen.
  • the CPU 101 displays, on the display 107 via the GPU 102 , a video see-through mode selection screen 5010 illustrated in FIG. 5 A such that the video see-through mode selection screen 5010 is superimposed on a virtual space image. Accordingly, the user views the video see-through mode selection screen 5010 floating in a three-dimensional space.
  • the video see-through mode selection screen 5010 includes a video see-through ON button 5001 , an accept button 5003 , and a decline button 5004 .
  • the video see-through ON button 5001 is a button for switching the video see-through mode from OFF to ON.
  • the CPU 101 displays, on the display 107 via the GPU 102 , a video see-through mode selection screen 5020 illustrated in FIG. 5 B such that the video see-through mode selection screen 5020 is superimposed on a real space image.
  • the video see-through mode selection screen 5020 includes a video see-through OFF button 5002 , the accept button 5003 , and the decline button 5004 .
  • the video see-through OFF button 5002 is a button for switching the video see-through mode from ON to OFF.
  • the accept button 5003 or the decline button 5004 being pressed, the CPU 101 starts or declines a phone conversation.
  • step S 4073 the CPU 101 acquires user operation information and sets the video see-through mode.
  • the CPU 101 sets the video see-through mode to ON.
  • the CPU 101 sets the video see-through mode to OFF.
  • step S 2080 the CPU 101 determines whether to end the application. If a determination is made that the application is not to be ended, the process returns to step S 2010 . If a determination is made that the application is to be ended, the process of the flowchart illustrated in FIG. 4 ends.
  • the CPU 101 in response to receipt of a phone call, displays the video see-through mode selection screen 5010 or 5020 for allowing the user to switch ON or OFF the vide see-through mode, instead of immediately setting the video see-through mode to ON.
  • the CPU 101 switches ON or OFF the video see-through mode. Accordingly, a sudden change of the entire screen (the space to the user) unintended by the user can be prevented.
  • the CPU 101 changes the screen in response to pressing of the video see-through ON button 5001 .
  • the CPU 101 may change the screen in response to pressing of the accept button 5003 .
  • step S 4071 if a determination is made in step S 4071 that a phone call is being received or a phone conversation is being performed, the process proceeds to step S 4072 .
  • step S 4072 the CPU 101 performs control to display an instruction button on the display 107 such that the instruction button is superimposed on a virtual space image.
  • the instruction button is the video see-through ON button 5001 or the accept button 5003 .
  • step S 4073 if an operation instruction for the instruction button is provided, the CPU 101 performs control to display an image captured by the camera unit 110 on the display 107 . If an operation instruction for the instruction button is provided, the CPU 101 may perform control to display an image captured by the camera unit 110 and a virtual space image on the display 107 . If an operation instruction for the instruction button is not provided, the CPU 101 performs control to display a virtual space image on the display 107 .
  • the present disclosure can be implemented by a process in which a program implementing one or more functions of the above-described embodiments is supplied to a system or an apparatus via a network or a storage medium and one or more processors in the system or the apparatus read and execute the program.
  • the present disclosure can also be implemented by a circuit that implements one or more functions (for example, an application specific integrated circuit (ASIC)).
  • ASIC application specific integrated circuit
  • a captured image for example, when a phone call is being received.
  • Embodiment(s) of the present disclosure can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions (e.g., one or more programs) recorded on a storage medium (which may also be referred to more fully as a ‘non-transitory computer-readable storage medium’) to perform the functions of one or more of the above-described embodiment(s) and/or that includes one or more circuits (e.g., application specific integrated circuit (ASIC)) for performing the functions of one or more of the above-described embodiment(s), and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s) and/or controlling the one or more circuits to perform the functions of one or more of the above-described embodiment(s).
  • computer executable instructions e.g., one or more programs
  • a storage medium which may also be referred to more fully as a
  • the computer may comprise one or more processors (e.g., central processing unit (CPU), micro processing unit (MPU)) and may include a network of separate computers or separate processors to read out and execute the computer executable instructions.
  • the computer executable instructions may be provided to the computer, for example, from a network or the storage medium.
  • the storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)TM), a flash memory device, a memory card, and the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Telephone Function (AREA)
  • Controls And Circuits For Display Device (AREA)

Abstract

A head-mounted display control apparatus includes one or more circuits, or one or more processors and at least one memory, the at least one memory being coupled to the one or more processors and having stored thereon instructions executable by the one or more processors. At least one of the one or more circuits or the execution of the instructions causes the head-mounted display control apparatus to function as an image capturing unit configured to capture an image; and a control unit configured to, based on phone call reception information, phone conversation information, or message reception information, perform control to display the captured image or a virtual space image on a display unit.

Description

    BACKGROUND Technical Field
  • The present disclosure relates to a head-mounted display control apparatus, a method for controlling the head-mounted display control apparatus, and a non-transitory computer readable storage medium.
  • Description of the Related Art
  • In recent years, XR information processing terminals of a head-mounted display (HMD) type have become widespread. XR is a generic term for virtual reality (VR), augmented reality (AR), and mixed reality (MR).
  • HMD XR information processing terminals include those of a see-through type in which computer graphics (CG) are displayed on a transmissive display, and those of a non-see-through type in which CG and a real image captured by a camera are displayed on a non-transmissive display. In a see-through display, a person wearing an HMD is able to determine the state of a real space. In a non-see-through display, the person is unable to determine the state of a real space but is highly immersed in a VR space. However, there is a non-see-through display in which the state of a real space can be determined by displaying an image captured by a camera built in the HMD on the display. A state in which a real space is displayed is called a video see-through mode, relative to a state in which a VR space is displayed.
  • A smartphone compatible with the headset profile (HSP) or the hands-free profile (HFP) of Bluetooth (registered trademark) and an audio terminal such as a headset enable a phone conversation with the audio terminal instead of a microphone and a speaker of the smartphone. The HSP is the specifications of the function of transmitting and receiving audio data between a smartphone and an audio terminal. The HFP is the specifications of the function of making and receiving phone calls.
  • U.S. patent Ser. No. 10/324,294 discloses that the reception of a message or a phone call in an external smartphone is indicated by an icon in an HMD.
  • SUMMARY
  • A head-mounted display control apparatus includes one or more circuits, or one or more processors and at least one memory, the at least one memory being coupled to the one or more processors and having stored thereon instructions executable by the one or more processors. At least one of the one or more circuits or the execution of the instructions causes the head-mounted display control apparatus to function as an image capturing unit configured to capture an image; and a control unit configured to, based on phone call reception information, phone conversation information, or message reception information, perform control to display the captured image or a virtual space image on a display unit.
  • Further features of the present disclosure will become apparent from the following description of exemplary embodiments with reference to the attached drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a diagram illustrating an example configuration of a head-mounted display (HMD).
  • FIG. 2 is a flowchart illustrating a method for controlling the HMD according to a first embodiment.
  • FIG. 3 is a flowchart illustrating a method for controlling the HMD according to a second embodiment.
  • FIG. 4 is a flowchart illustrating a method for controlling the HMD according to a third embodiment.
  • FIGS. 5A and 5B are each a diagram for describing a user interface screen displayed when a phone call is being received.
  • DESCRIPTION OF THE EMBODIMENTS First Embodiment
  • In the above-described related art, when a phone call is received in a real space while a virtual space is displayed on a non-see-through head-mounted display (HMD), the situation of the real space, for example, whether there is another person, is unrecognizable, and it is impossible to immediately determine whether a phone conversation can be started.
  • The present disclosure provides a technique of enabling a captured image to be displayed, for example, when a phone call is being received.
  • FIG. 1 is a diagram illustrating an example configuration of a head-mounted display (HMD) 120 according to a first embodiment. The HMD 120 is an example of a display control apparatus that responds to a phone call reception signal. The HMD 120 includes a bus 100, a central processing unit (CPU) 101, a graphics processing unit (GPU) 102, a posture sensor 103, an analog-to-digital/digital-to-analog (AD/DA) converter 104 for audio, and a wireless unit 105. The HMD 120 further includes an image processor 106, a display 107, a microphone 108, a speaker 109, a camera unit 110, a random access memory (RAM) 111, and a nonvolatile memory 112.
  • Next, the roles of the individual components will be described. The bus 100 manages the data flow in the HMD 120. The CPU 101 executes a program stored in the nonvolatile memory 112 or the like to control the entire HMD 120, and executes system software and application software. In addition, the CPU 101 analyzes a digital audio signal to determine whether the digital audio signal is a phone call reception sound.
  • The GPU 102 generates a virtual reality image and outputs data of the image to the display 107. The posture sensor 103 includes a gyro sensor and an acceleration sensor, and acquires information regarding the posture and position of the HMD 120. In the present embodiment, the CPU 101 processes raw data received from the posture sensor 103 and acquires information regarding the posture and position of the HMD 120. The information regarding the posture and position is reflected in generation of a virtual reality image. That is, the HMD 120 generates a virtual reality image corresponding to the forward direction of the HMD 120 in accordance with a motion of the HMD 120 including the posture sensor 103, and displays the virtual reality image on the display 107.
  • The AD/DA converter 104 converts an analog audio signal received from the microphone 108 into a digital audio signal, converts the digital audio signal into an analog audio signal, and outputs the analog audio signal to the speaker 109. The wireless unit 105 is capable of connecting to another terminal by using a plurality of wireless standards (Bluetooth, Wi-Fi, and the like) to transmit and receive data. Specifically, the wireless unit 105 connects to a controller device to acquire operation information, or connects to a smartphone to transmit and receive audio data and acquire phone call reception information.
  • The image processor 106 acquires mosaic image data from the camera unit 110 and performs demosaicing and color correction processing. The display 107 displays a real space image captured by the camera unit 110 or a virtual reality image generated by the GPU 102. Which is to be displayed on the display 107 is controlled by the CPU 101. The logic of the control will be described below. The display 107 includes two panels, one is for the right eye and the other is for the left eye, and displays an input stereo image on the panels such that the stereo image is divided into partial images for the respective panels. A monocular image can also be input to the display 107. In this case, the same image is displayed on the panels for the right eye and the left eye. Switching of the image to be displayed on the display 107 is controlled by the CPU 101.
  • The microphone 108 physically acquires a sound and converts the sound into an analog electric signal. The speaker 109 outputs an analog signal as a physical sound.
  • The camera unit 110 includes two lenses and two sensors, is disposed so as to be capable of capturing images for the positions of the two eyes of the user wearing the HMD 120, and outputs captured and digitalized mosaic image data to the image processor 106. The image processor 106 combines the images received from the two sensors of the camera unit 110 into one stereo image. When the distance between the centers of the lenses of the camera unit 110 is different from the distance between the eyes of the user, the image processor 106 may correct a display image so that the user's perception of depth coincides with the reality.
  • The RAM 111 loads a program and data stored in the nonvolatile memory 112 at the startup of the HMD 120 and functions as a work area for the CPU 101. The RAM 111 stores an acquired image and a generated image. The nonvolatile memory 112 stores a program and data. These operations are successively performed, and the user is able to view a moving image.
  • FIG. 2 is a flowchart illustrating a method for controlling the HMD 120 according to the first embodiment, and describes a phone call reception operation of the HMD 120 that responds to reception of a phone call. The steps are performed in the order indicated by the arrows in FIG. 2 unless otherwise stated. If the input and output of data in the individual steps do not depend on each other, the steps may be performed in a different order. The individual steps are performed by the CPU 101, but part of processing, such as rendering processing, is performed by another processor such as the GPU 102. In the present embodiment, a description will be given under the assumption that a VR game application is being executed.
  • In step S2010, the CPU 101 acquires phone call reception/phone conversation state information via the wireless unit 105. In the present embodiment, a smartphone connected to the HMD 120 via the wireless unit 105 transmits phone call reception/phone conversation state information to the HMD 120, and thereby the CPU 101 acquires the phone call reception/phone conversation state information via the wireless unit 105.
  • In step S2020, the CPU 101 determines, based on the phone call reception/phone conversation state information, whether a phone call is being received or a phone conversation is being performed. If a determination is made that a phone call is being received or a phone conversation is being performed, the process proceeds to step S2030. Otherwise, the process proceeds to step S2060.
  • In step S2030, the CPU 101 suspends the processing of the above-described application that is being executed. In the present embodiment, the CPU 101 causes the game application to pause and causes the system to perform control.
  • In step S2040, the CPU 101 acquires a real space image by using the camera unit 110 and the image processor 106. The real space image is captured by the camera unit 110 built in the HMD 120 and is a stereo image of a region in front of the HMD 120.
  • In step S2050, the CPU 101 renders the real space image. Specifically, the CPU 101 converts the format of the image acquired in step S2040 into a format suitable for the display 107, and transmits the converted stereo image data to the display 107 via the GPU 102. The stereo image is displayed on the display 107. In step S2050, the CPU 101 also performs processing of waiting in synchronization with a video signal. Subsequently, the process proceeds to step S2080.
  • In step S2060, the CPU 101 executes the processing of the above-described application. In the present embodiment, the CPU 101 executes processing for one frame and creates application data that is necessary for generating rendering data. In a case where the processing of the application is executed in another thread, the application data may be acquired in synchronization with the processing of the other thread.
  • In step S2070, the CPU 101 renders a virtual space image that is based on the application processing. The CPU 101 transmits, based on the application data generated in step S2050 and information from the posture sensor 103, an image of a virtual space in front to the display 107 via the GPU 102. The virtual space image is displayed on the display 107. In step S2070, the CPU 101 also performs processing of waiting in synchronization with a video signal. The virtual space image may be a stereo image that is based on computer graphics or may be a stereo image actually captured by using a stereo fish eye lens. Subsequently, the process proceeds to step S2080.
  • In step S2080, the CPU 101 determines whether to end the application. If a determination is made that the application is not to be ended, the process returns to step S2010. If a determination is made that the application is to be ended, the process of the flowchart illustrated in FIG. 2 ends.
  • Typically, the user holds a controller in his/her hand and does not hold a smartphone when playing a game. In the present embodiment, in response to receipt of a phone call in the real space, the CPU 101 proceeds to step S2030, automatically performs switching to a video see-through mode, and displays a real space image. Accordingly, the user is able to view the real space, immediately find his/her smartphone around him/her, operate the smartphone, and answer the phone call, without taking off the HMD 120. In addition, the user is able to easily determine whether to answer the phone call or whether to avoid an important topic in a phone conversation by determining whether there is somebody around him/her.
  • In the present embodiment, the microphone and speaker used for a phone conversation by the user are not limited. The user may use the microphone 108 and the speaker 109 of the HMD 120, and the HMD 120 may transmit/receive audio information to/from the smartphone. Alternatively, the user may use a microphone and a speaker built in the smartphone to perform a phone conversation.
  • In the present embodiment, an example is illustrated in which the user operates his/her smartphone to answer an incoming call. The CPU 101 may display a phone call answering screen on the display 107 of the HMD 120 and may transmit/receive audio information to/from the smartphone. In this case, the user performs a phone conversation by using the microphone 108 and the speaker 109 of the HMD 120.
  • In the present embodiment, a description has been given under the assumption that the target application is a game application. The target application is not limited thereto and may be any application, such as a VR moving image player, a work support application, or a virtual office application. In a VR moving image player, the CPU 101 decodes video data to generate a frame image and renders the frame image as a virtual reality image.
  • In the present embodiment, a description has been given of a phone conversation. When the smartphone receives a message, the CPU 101 may perform switching to the video see-through mode.
  • In the present embodiment, the phone call that is received may be a phone call received by either a smartphone connected to a public line or a phone conversation application of the system. However, in a case where users communicate with each other in a game application or the like, a screen is not automatically switched in response to receipt of a phone call within the application.
  • In the present embodiment, a display image is switched in accordance with only whether a phone call is being received. The HMD 120 is equipped with various sensors and a camera. The CPU 101 may use information obtained therefrom to detect whether there is a person near the user, may proceed to step S2030 if a phone call is being received and there is a person near the user, and may switch the display of a real space image.
  • In the present embodiment, the CPU 101 generates a virtual reality image corresponding to the forward direction of the HMD 120 in accordance with a motion of the HMD 120 including the posture sensor 103. Alternatively, the posture sensor 103 need not be included, and an image in a fixed direction may be displayed.
  • Although a stereo image is handled in the present embodiment, a monocular image may be handled instead of a stereo image.
  • In the present embodiment, the CPU 101 performs switching to display of a real space image in response to receipt of a phone call, but the present disclosure is not limited thereto. For example, the CPU 101 may display a virtual space image and a real space image in combination, or may make a small window on a virtual space image in response to receipt of a phone call and display a real space image in the video see-through mode only in the small window.
  • As described above, the camera unit 110 and the image processor 106 function as an image capturing unit, which captures an image. The display 107 is an example of a display unit. In step S2010, the CPU 101 acquires phone call reception information, phone conversation information, or message reception information from an external apparatus. The external apparatus is a smartphone, for example. The CPU 101 functions as a control unit, and performs control to display a captured image or a virtual space image on the display 107, based on the phone call reception information, the phone conversation information, or the message reception information.
  • When a phone call is not being received or a phone conversation is not being performed, the CPU 101 performs control to display a virtual space image on the display 107 in step S2070. When a phone call is being received or a phone conversation is being performed, the CPU 101 performs control to display a captured image on the display 107 in step S2050.
  • In step S2050, the CPU 101 may perform control to display a captured image and a virtual space image on the display 107. For example, the CPU 101 is capable of displaying a captured image and a virtual space image in combination, or displaying a captured image in a small window on a virtual space image.
  • The present embodiment may be applied to reception of a message in a smartphone. When no message is received, the CPU 101 may perform control to display a virtual space image on the display 107. When a message is received, the CPU 101 may perform control to display a captured image on the display 107 for a predetermined period. When a message is received, the CPU 101 may perform control to display a captured image and a virtual space image on the display 107 for a predetermined period, similarly to the above.
  • When a person different from a person wearing the HMD 120 is present within a predetermined range, the CPU 101 may perform control to display a captured image or a virtual space image on the display 107, based on phone call reception information, phone conversation information, or message reception information.
  • As described above, according to the present embodiment, the user is able to determine the situation of a real space and immediately determine whether he/she can start a phone conversation, without taking off the HMD 120.
  • Second Embodiment
  • FIG. 3 is a flowchart illustrating a method for controlling the HMD 120 according to a second embodiment, and describes a phone call reception operation of the HMD 120 that responds to reception of a phone call. FIG. 3 is different from FIG. 2 in that steps S3000, S3005, and S3010 are provided instead of steps S2010 and S2020. The operation is the same as in the first embodiment unless otherwise stated.
  • In step S3000, the CPU 101 acquires audio information via the microphone 108 and the AD/DA converter 104. The audio information is acquired as digital audio data, which is generated by AD converting a sound received by the microphone 108 built in the HMD 120 by using the AD/DA converter 104. The CPU 101 acquires audio data for 10 seconds before the time point at which step S3000 is performed.
  • In step S3005, the CPU 101 analyzes the audio information acquired in step S3000 to generate an audio information analysis result. In the present embodiment, the CPU 101 performs fast Fourier transform (FFT) on the audio information to acquire information in a range of 15 Hz to 16 kHz as an analysis result.
  • In step S3010, the CPU 101 determines whether the audio information analysis result generated in step S3005 indicates a phone call reception sound. If the audio information analysis result indicates a phone call reception sound, the process proceeds to step S2030. Otherwise, the process proceeds to step S2060. In the present embodiment, if the sound pressure of the peak of at least one frequency is 60 dB or more as a result of FFT, the CPU 101 determines that the audio information analysis result indicates a phone call reception sound. The peak may be determined in the following manner. For example, if there is an amplitude ten times the median of the amplitude in a Log scale of the FFT result, it is determined that there is a peak.
  • In the present embodiment, the CPU 101 is capable of detecting a phone call reception sound of a smartphone by using the microphone 108 in an analog manner and performing switching to the video see-through mode in step S2050, even if the CPU 101 is incapable of digitally transmitting/receiving a signal via the wireless unit 105. The present embodiment is capable of obtaining a merit similar to that of the first embodiment.
  • In the present embodiment, a result of FFT in the range of 15 Hz to 10 kHz is regarded as an audio analysis result, but the present disclosure is not limited thereto. Alternatively, the number of successive identical patterns may be calculated with respect to a time-series change in waveform, and if the number is two or more, the successive identical patterns may be determined to be a phone call reception sound. Regarding this, a warning sound such as a fire alarm sound often has audio information having repetition of an identical pattern, and causes automatic switching to the video see-through mode. This is an effective operation in terms of being able to notice the occurrence of a warning sound and visually recognize the cause of the occurrence. A phone call reception sound may be recorded in advance, and whether a sound is identical to the recorded phone call reception sound may be determined. Accordingly, even if a melody is set as a phone call reception sound, the melody can be determined to be a phone call reception sound.
  • As described above, the CPU 101 performs control to display a captured image or a virtual space image on the display 107, based on a phone call reception sound or a message reception sound collected by the microphone 108.
  • Third Embodiment
  • FIG. 4 is a flowchart illustrating a method for controlling the HMD 120 according to a third embodiment, and describes a phone call reception operation of the HMD 120 that responds to reception of a phone call. FIG. 4 is different from FIG. 2 in that step S2020 is deleted and steps S4000, S4011, S4012, and S4071 to S4073 are added. The operation is the same as in the first embodiment unless otherwise stated.
  • In step S4000, the CPU 101 sets the video see-through mode to OFF. This corresponds to setting of an initial state.
  • In step S2010, the CPU 101 acquires phone call reception/phone conversation state information via the wireless unit 105.
  • In step S4011, the CPU 101 acquires a video see-through mode. Specifically, the CPU 101 acquires the video see-through mode set in step S4000 described above or step S4073 described below.
  • In step S4012, the CPU 101 determines whether the video see-through mode is ON. If the video see-through mode is ON, the process proceeds to step S4071 via steps S2030 to S2050. If the video see-through mode is OFF, the process proceeds to step S4071 via steps S2060 and S2070.
  • In step S4071, the CPU 101 determines, based on the phone call reception/phone conversation state information acquired in step S2010, whether a phone call is being received or a phone conversation is being performed. If a determination is made that a phone call is being received or a phone conversation is being performed, the process proceeds to step S4072. Otherwise, the process proceeds to step S2080.
  • In step S4072, the CPU 101 renders a video see-through mode selection screen. In a state in which the video see-through mode is OFF, the CPU 101 displays, on the display 107 via the GPU 102, a video see-through mode selection screen 5010 illustrated in FIG. 5A such that the video see-through mode selection screen 5010 is superimposed on a virtual space image. Accordingly, the user views the video see-through mode selection screen 5010 floating in a three-dimensional space. The video see-through mode selection screen 5010 includes a video see-through ON button 5001, an accept button 5003, and a decline button 5004. The video see-through ON button 5001 is a button for switching the video see-through mode from OFF to ON.
  • In a state in which the video see-through mode is ON, the CPU 101 displays, on the display 107 via the GPU 102, a video see-through mode selection screen 5020 illustrated in FIG. 5B such that the video see-through mode selection screen 5020 is superimposed on a real space image. The video see-through mode selection screen 5020 includes a video see-through OFF button 5002, the accept button 5003, and the decline button 5004. The video see-through OFF button 5002 is a button for switching the video see-through mode from ON to OFF.
  • The user presses a button on the video see-through mode selection screen 5010 or 5020 by using a controller or his/her finger. In response to the accept button 5003 or the decline button 5004 being pressed, the CPU 101 starts or declines a phone conversation.
  • In step S4073, the CPU 101 acquires user operation information and sets the video see-through mode. In response to the video see-through ON button 5001 being pressed, the CPU 101 sets the video see-through mode to ON. In response to the video see-through OFF button 5002 being pressed, the CPU 101 sets the video see-through mode to OFF.
  • In step S2080, the CPU 101 determines whether to end the application. If a determination is made that the application is not to be ended, the process returns to step S2010. If a determination is made that the application is to be ended, the process of the flowchart illustrated in FIG. 4 ends.
  • According to the present embodiment, in response to receipt of a phone call, the CPU 101 displays the video see-through mode selection screen 5010 or 5020 for allowing the user to switch ON or OFF the vide see-through mode, instead of immediately setting the video see-through mode to ON. In accordance with a selection result made on the video see-through mode selection screen 5010 or 5020, the CPU 101 switches ON or OFF the video see-through mode. Accordingly, a sudden change of the entire screen (the space to the user) unintended by the user can be prevented. In the present embodiment, the CPU 101 changes the screen in response to pressing of the video see-through ON button 5001. Alternatively, the CPU 101 may change the screen in response to pressing of the accept button 5003.
  • As described above, if a determination is made in step S4071 that a phone call is being received or a phone conversation is being performed, the process proceeds to step S4072. In step S4072, the CPU 101 performs control to display an instruction button on the display 107 such that the instruction button is superimposed on a virtual space image. The instruction button is the video see-through ON button 5001 or the accept button 5003.
  • In step S4073, if an operation instruction for the instruction button is provided, the CPU 101 performs control to display an image captured by the camera unit 110 on the display 107. If an operation instruction for the instruction button is provided, the CPU 101 may perform control to display an image captured by the camera unit 110 and a virtual space image on the display 107. If an operation instruction for the instruction button is not provided, the CPU 101 performs control to display a virtual space image on the display 107.
  • Other Embodiments
  • The present disclosure can be implemented by a process in which a program implementing one or more functions of the above-described embodiments is supplied to a system or an apparatus via a network or a storage medium and one or more processors in the system or the apparatus read and execute the program. The present disclosure can also be implemented by a circuit that implements one or more functions (for example, an application specific integrated circuit (ASIC)).
  • The above-described embodiments are merely specific examples for implementing the present disclosure, and the technical scope of the present disclosure is not to be interpreted in a limited manner based on these embodiments. That is, the present disclosure can be implemented in various ways without departing from the technical spirit thereof or the main features thereof.
  • According to the present disclosure, it is possible to display a captured image, for example, when a phone call is being received.
  • Embodiment(s) of the present disclosure can also be realized by a computer of a system or apparatus that reads out and executes computer executable instructions (e.g., one or more programs) recorded on a storage medium (which may also be referred to more fully as a ‘non-transitory computer-readable storage medium’) to perform the functions of one or more of the above-described embodiment(s) and/or that includes one or more circuits (e.g., application specific integrated circuit (ASIC)) for performing the functions of one or more of the above-described embodiment(s), and by a method performed by the computer of the system or apparatus by, for example, reading out and executing the computer executable instructions from the storage medium to perform the functions of one or more of the above-described embodiment(s) and/or controlling the one or more circuits to perform the functions of one or more of the above-described embodiment(s). The computer may comprise one or more processors (e.g., central processing unit (CPU), micro processing unit (MPU)) and may include a network of separate computers or separate processors to read out and execute the computer executable instructions. The computer executable instructions may be provided to the computer, for example, from a network or the storage medium. The storage medium may include, for example, one or more of a hard disk, a random-access memory (RAM), a read only memory (ROM), a storage of distributed computing systems, an optical disk (such as a compact disc (CD), digital versatile disc (DVD), or Blu-ray Disc (BD)™), a flash memory device, a memory card, and the like.
  • While the present disclosure has been described with reference to exemplary embodiments, it is to be understood that the disclosure is not limited to the disclosed exemplary embodiments. The scope of the following claims is to be accorded the broadest interpretation so as to encompass all such modifications and equivalent structures and functions.
  • This application claims the benefit of Japanese Patent Application No. 2022-184911, filed Nov. 18, 2022, which is hereby incorporated by reference herein in its entirety.

Claims (16)

What is claimed is:
1. A head-mounted display control apparatus comprising:
one or more circuits; or
one or more processors and at least one memory, the at least one memory being coupled to the one or more processors and having stored thereon instructions executable by the one or more processors,
wherein at least one of the one or more circuits or the execution of the instructions causes the head-mounted display control apparatus to function as:
an image capturing unit configured to capture an image; and
a control unit configured to, based on phone call reception information, phone conversation information, or message reception information, perform control to display the captured image or a virtual space image on a display unit.
2. The head-mounted display control apparatus according to claim 1, wherein
the control unit is configured to
when a phone call is not being received or a phone conversation is not being performed, perform control to display the virtual space image on the display unit, and
when a phone call is being received or a phone conversation is being performed, perform control to display the captured image on the display unit.
3. The head-mounted display control apparatus according to claim 1, wherein
the control unit is configured to
when a phone call is not being received or a phone conversation is not being performed, perform control to display the virtual space image on the display unit, and
when a phone call is being received or a phone conversation is being performed, perform control to display the captured image and the virtual space image on the display unit.
4. The head-mounted display control apparatus according to claim 1, wherein
the control unit is configured to
when no message is received, perform control to display the virtual space image on the display unit, and
when a message is received, perform control to display the captured image on the display unit for a predetermined period.
5. The head-mounted display control apparatus according to claim 1, wherein
the control unit is configured to
when no message is received, perform control to display the virtual space image on the display unit, and
when a message is received, perform control to display the captured image and the virtual space image on the display unit for a predetermined period.
6. The head-mounted display control apparatus according to claim 1, wherein
the control unit is configured to, when a person different from a person wearing the display control apparatus is present within a predetermined range,
perform control to display the captured image or the virtual space image on the display unit, based on the phone call reception information, the phone conversation information, or the message reception information.
7. The head-mounted display control apparatus according to claim 1, wherein the control unit is configured to acquire the phone call reception information, the phone conversation information, or the message reception information from an external apparatus.
8. The head-mounted display control apparatus according to claim 1, wherein the control unit is configured to, based on a phone call reception sound or a message reception sound, perform control to display the captured image or the virtual space image on the display unit.
9. The head-mounted display control apparatus according to claim 8, wherein the phone call reception sound or the message reception sound is a phone call reception sound or a message reception sound collected by a microphone.
10. The head-mounted display control apparatus according to claim 2, wherein
the control unit is configured to
when a phone call is being received or a phone conversation is being performed, and an instruction is provided, perform control to display the captured image on the display unit, and
when a phone call is being received or a phone conversation is being performed, and no instruction is provided, perform control to display the virtual space image on the display unit.
11. The head-mounted display control apparatus according to claim 10, wherein
the control unit is configured to
when a phone call is being received or a phone conversation is being performed, perform control to display an instruction button on the display unit such that the instruction button is superimposed on the virtual space image,
when a phone call is being received or a phone conversation is being performed, and an operation instruction for the instruction button is provided, perform control to display the captured image on the display unit, and
when a phone call is being received or a phone conversation is being performed, and no operation instruction for the instruction button is provided, perform control to display the virtual space image on the display unit.
12. The head-mounted display control apparatus according to claim 3, wherein
the control unit is configured to
when a phone call is being received or a phone conversation is being performed, and an instruction is provided, perform control to display the captured image and the virtual space image on the display unit, and
when a phone call is being received or a phone conversation is being performed, and no instruction is provided, perform control to display the virtual space image on the display unit.
13. The head-mounted display control apparatus according to claim 12, wherein
the control unit is configured to
when a phone call is being received or a phone conversation is being performed, perform control to display an instruction button on the display unit such that the instruction button is superimposed on the virtual space image,
when a phone call is being received or a phone conversation is being performed, and an operation instruction for the instruction button is provided, perform control to display the captured image and the virtual space image on the display unit, and
when a phone call is being received or a phone conversation is being performed, and no operation instruction for the instruction button is provided, perform control to display the virtual space image on the display unit.
14. The head-mounted display control apparatus according to claim 1, wherein the display control apparatus is a head-mounted display.
15. A method for controlling a head-mounted display control apparatus, comprising:
capturing an image; and
performing control to display the captured image or a virtual space image on a display unit, based on phone call reception information, phone conversation information, or message reception information.
16. A non-transitory computer readable storage medium storing instructions that, when executed by one or more processors of a head-mounted display control apparatus configures the head-mounted display control apparatus to execute a process of:
capturing an image; and
performing control to display the captured image or a virtual space image on a display unit, based on phone call reception information, phone conversation information, or message reception information.
US18/511,788 2022-11-18 2023-11-16 Head-mounted display control apparatus, method for controlling head-mounted display control apparatus, and non-transitory computer readable storage medium Pending US20240168701A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022184911A JP2024073924A (en) 2022-11-18 2022-11-18 DISPLAY CONTROL DEVICE, CONTROL METHOD AND PROGRAM FOR DISPLAY CONTROL DEVICE
JP2022-184911 2022-11-18

Publications (1)

Publication Number Publication Date
US20240168701A1 true US20240168701A1 (en) 2024-05-23

Family

ID=91079806

Family Applications (1)

Application Number Title Priority Date Filing Date
US18/511,788 Pending US20240168701A1 (en) 2022-11-18 2023-11-16 Head-mounted display control apparatus, method for controlling head-mounted display control apparatus, and non-transitory computer readable storage medium

Country Status (2)

Country Link
US (1) US20240168701A1 (en)
JP (1) JP2024073924A (en)

Also Published As

Publication number Publication date
JP2024073924A (en) 2024-05-30

Similar Documents

Publication Publication Date Title
US20210082435A1 (en) Multi-mode guard for voice commands
JP6743691B2 (en) Display control device, display control method, and computer program
EP3143477B1 (en) System and method for providing haptic feedback to assist in capturing images
US10609196B2 (en) Method and apparatus for controlling application program, and electronic device
EP3299946B1 (en) Method and device for switching environment picture
CN109416562B (en) Apparatus, method and computer readable medium for virtual reality
CN110634189A (en) System and method for user alerts during immersive mixed reality experience
US20210201908A1 (en) Driving interaction object
KR102508286B1 (en) Electronic device and methodfor providing information in virtual reality
JP2016105534A (en) Imaging apparatus and imaging apparatus system
JPWO2018158852A1 (en) Call system
US20220021763A1 (en) Touch Operation Locking Method and Electronic Device
CN113821337A (en) Varying resource utilization associated with a media object based on engagement scores
CN105637851B (en) Computing system, method and computer readable medium for peripheral control
US20240168701A1 (en) Head-mounted display control apparatus, method for controlling head-mounted display control apparatus, and non-transitory computer readable storage medium
US11037519B2 (en) Display device having display based on detection value, program, and method of controlling device
KR20220006924A (en) Wearable device and operating method thereof
US11327576B2 (en) Information processing apparatus, information processing method, and program
CN109144461B (en) Sound production control method and device, electronic device and computer readable medium
JP2019197497A (en) Head-mounted display system, notification controller, method for controlling notification, and program
CN115686190A (en) Guiding a virtual agent based on eye behavior of a user
WO2023058451A1 (en) Information processing device, information processing method, and program
WO2018056169A1 (en) Interactive device, processing method, and program
US20230347527A1 (en) Communication system, control method, and storage medium
WO2023228391A1 (en) Portable information terminal and display control method for portable information terminal

Legal Events

Date Code Title Description
AS Assignment

Owner name: CANON KABUSHIKI KAISHA, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KOBAYASHI, MASAAKI;REEL/FRAME:065821/0846

Effective date: 20231107

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION