CN114205318B - Head portrait display method and electronic equipment - Google Patents

Head portrait display method and electronic equipment Download PDF

Info

Publication number
CN114205318B
CN114205318B CN202010899499.XA CN202010899499A CN114205318B CN 114205318 B CN114205318 B CN 114205318B CN 202010899499 A CN202010899499 A CN 202010899499A CN 114205318 B CN114205318 B CN 114205318B
Authority
CN
China
Prior art keywords
user
chat group
display area
chat
head portrait
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010899499.XA
Other languages
Chinese (zh)
Other versions
CN114205318A (en
Inventor
王龙
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Honor Device Co Ltd
Original Assignee
Honor Device Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Honor Device Co Ltd filed Critical Honor Device Co Ltd
Priority to CN202010899499.XA priority Critical patent/CN114205318B/en
Priority to PCT/CN2021/125916 priority patent/WO2022042774A1/en
Publication of CN114205318A publication Critical patent/CN114205318A/en
Application granted granted Critical
Publication of CN114205318B publication Critical patent/CN114205318B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L51/00User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
    • H04L51/04Real-time or near real-time messaging, e.g. instant messaging [IM]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/16Arrangements for providing special services to substations
    • H04L12/18Arrangements for providing special services to substations for broadcast or conference, e.g. multicast
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/02Details
    • H04L12/16Arrangements for providing special services to substations
    • H04L12/18Arrangements for providing special services to substations for broadcast or conference, e.g. multicast
    • H04L12/1813Arrangements for providing special services to substations for broadcast or conference, e.g. multicast for computer conferences, e.g. chat rooms
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/02Protocols based on web technology, e.g. hypertext transfer protocol [HTTP]

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • General Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses a head portrait display method and electronic equipment, wherein the head portrait display method comprises the following steps: acquiring the activity of a first user in a chat group in an instant messaging tool in the chat group; determining the display area size of the user head portrait of the first user in the chat group according to the liveness; and displaying the user head images of the first user in the chat group according to the size of the display area. Therefore, when a plurality of chat groups exist in the instant messaging tool, the sizes of the display areas of the user head portraits determined according to the liveness in different chat groups are different, the head portraits of the chat groups are large in difference, and users can quickly identify the chat groups to be searched by using the head portraits of the chat groups, so that the identification efficiency among the chat groups is improved, and the user experience is improved.

Description

Head portrait display method and electronic equipment
Technical Field
The present application relates to the field of electronic devices, and in particular, to a head portrait display method and an electronic device.
Background
With the development of the internet, people have increasingly demanded instant messaging, and the functions of instant messaging tools are also increasingly abundant. In order to meet the requirement of multi-person chat, the instant messaging tool supports the establishment of a chat group, and a plurality of users in the chat group can chat with multiple persons in the chat group.
Currently, chat groups in instant messaging tools are typically displayed as a conversation list that includes avatars and text, the avatars of the chat groups being made up of user avatars of users in the chat groups. The conversation list in the instant messaging tool includes a conversation list of chat groups and a conversation list of individual chats. When the conversation list in the instant messaging tool is more, the head portraits of the chat groups can only be used for rapidly distinguishing the chat groups from the individual chat, and the head portraits of the chat groups cannot be used for rapidly identifying among the plurality of chat groups. Thus, the recognition efficiency between chat groups is low, and the user experience is poor.
Disclosure of Invention
The head portrait display method and the electronic device provided by the application have the advantages that the identification efficiency among chat groups is improved, and the user experience is improved.
In a first aspect, the present application provides a head portrait display method, including: acquiring the activity of a first user in a chat group in an instant messaging tool in the chat group; determining the display area size of the user head portrait of the first user in the chat group according to the liveness; and displaying the user head images of the first user in the chat group according to the size of the display area.
According to the method, the activity of the user in the chat group is obtained, and the size of the display area of the user head portrait in the chat group is determined according to the activity, so that the size of the display area of the user head portrait in the chat group is intelligently adjusted. When a plurality of chat groups exist in the instant messaging tool, the sizes of the display areas of the user head portraits determined according to the liveness in different chat groups are different, the head portraits of the chat groups are large in difference, and a user can quickly identify the chat groups to be searched by using the head portraits of the chat groups. Therefore, the recognition efficiency between chat groups is improved, and the user experience is improved.
With reference to the first aspect, in a possible implementation manner of the first aspect, the activity of the first user in the chat group includes a frequency of sending messages by the first user in the chat group and/or an interaction frequency of the first user and a second user in the chat group, where the second user is an application user of the instant messaging tool.
With reference to the first aspect, in a possible implementation manner of the first aspect, the determining, according to the activity level, a display area size of a user avatar of the first user in the chat group includes: and when the activity of the first user in the chat group is highest, determining that the display area of the user head portrait of the first user in the chat group is largest.
With reference to the first aspect, in a possible implementation manner of the first aspect, the determining, according to the activity level, a display area size of a user avatar of the first user in the chat group further includes: when the activity of the first user in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is equal to the display area of the user head portrait of a third user in the chat group, wherein the third user is other users except the user with the highest activity in the chat group.
With reference to the first aspect, in a possible implementation manner of the first aspect, the determining, according to the activity level, a display area size of a user avatar of the first user in the chat group further includes: when the activity of the first user in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is larger than the display area of the user head portrait of a fourth user in the chat group, wherein the fourth user is a user with activity lower than that of the first user in the chat group.
With reference to the first aspect, in a possible implementation manner of the first aspect, the method further includes: determining the arrangement sequence of user head portraits of the first user in the chat group according to the liveness; displaying the user head images of the first user in the chat group according to the arrangement sequence.
With reference to the first aspect, in a possible implementation manner of the first aspect, the determining, according to the activity level, an arrangement order of user avatars of the first user in the chat group includes: and when the activity of the first user in the chat group is highest, determining that the arrangement sequence of the user head portraits of the first user in the chat group is the forefront.
With reference to the first aspect, in a possible implementation manner of the first aspect, before the determining, according to the activity level, a size of a display area of a user avatar of the first user in the chat group, the method further includes: and confirming that the activity degree of the first user in the chat group is not smaller than a preset activity degree threshold value.
In a second aspect, the present application provides an electronic device comprising: one or more processors, memory; the memory is coupled with the one or more processors, the memory is for storing computer program code, the computer program code comprising computer instructions, the one or more processors executing the computer instructions to perform: acquiring the activity of a first user in a chat group in an instant messaging tool in the chat group; determining the display area size of the user head portrait of the first user in the chat group according to the liveness; and displaying the user head images of the first user in the chat group according to the size of the display area.
In the electronic device, the activity of the user in the chat group is obtained, and the size of the display area of the user head portrait in the chat group is determined according to the activity, so that the size of the display area of the user head portrait in the chat group is intelligently adjusted. When a plurality of chat groups exist in the instant messaging tool, the sizes of the display areas of the user head portraits determined according to the liveness in different chat groups are different, the head portraits of the chat groups are large in difference, and a user can quickly identify the chat groups to be searched by using the head portraits of the chat groups. Therefore, the recognition efficiency between chat groups is improved, and the user experience is improved.
With reference to the second aspect, in a possible implementation manner of the second aspect, the activity of the first user in the chat group includes a frequency of sending messages by the first user in the chat group and/or an interaction frequency of the first user and a second user in the chat group, where the second user is an application user of the instant messaging tool.
With reference to the second aspect, in a possible implementation manner of the second aspect, the processor is configured to determine, according to the activity level, a size of a display area of a user avatar of the first user in the chat group, where the processor is specifically configured to: and when the activity of the first user in the chat group is highest, determining that the display area of the user head portrait of the first user in the chat group is largest.
With reference to the second aspect, in a possible implementation manner of the second aspect, the processor is configured to determine, according to the activity level, a display area size of a user avatar of the first user in the chat group, where the processor is further configured to: when the activity of the first user in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is equal to the display area of the user head portrait of a third user in the chat group, wherein the third user is other users except the user with the highest activity in the chat group.
With reference to the second aspect, in a possible implementation manner of the second aspect, the processor is configured to determine, according to the activity level, a display area size of a user avatar of the first user in the chat group, where the processor is further configured to: when the activity of the first user in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is larger than the display area of the user head portrait of a fourth user in the chat group, wherein the fourth user is a user with activity lower than that of the first user in the chat group.
With reference to the second aspect, in a possible implementation manner of the second aspect, the processor is further configured to: determining the arrangement sequence of user head portraits of the first user in the chat group according to the liveness; displaying the user head images of the first user in the chat group according to the arrangement sequence.
With reference to the second aspect, in a possible implementation manner of the second aspect, the processor is configured to, when determining, according to the liveness, an arrangement order of user avatars of the first user in the chat group, the processor is specifically configured to: and when the activity of the first user in the chat group is highest, determining that the arrangement sequence of the user head portraits of the first user in the chat group is the forefront.
With reference to the second aspect, in a possible implementation manner of the second aspect, the processor is configured to determine, according to the liveness, a display area size of a user avatar of the first user before the chat group, where the processor is further configured to: and confirming that the activity degree of the first user in the chat group is not smaller than a preset activity degree threshold value.
In a third aspect, the present application provides a computer readable storage medium storing a computer program comprising program instructions which, when run on an electronic device, cause the electronic device to perform a method as in the first aspect or any of the possible implementations of the first aspect.
In a fourth aspect, the application provides a computer program product which, when run on an electronic device, causes the electronic device to perform the method as in the first aspect or any of the possible implementations of the first aspect.
Drawings
FIG. 1 is a schematic diagram of a user interface of an instant messaging tool;
fig. 2 is a schematic structural diagram of an electronic device according to an embodiment of the present application;
fig. 3 is a software structural block diagram of an electronic device according to an embodiment of the present application;
fig. 4 is a schematic diagram of a head portrait display method according to an embodiment of the present application;
fig. 5 is a schematic diagram of a user interface of an instant messaging tool according to an embodiment of the present application;
fig. 6a is a schematic diagram of a user avatar of a user in a chat group according to an embodiment of the present application;
FIG. 6b is a schematic diagram of a user avatar of another user in a chat group according to an embodiment of the application;
fig. 7 is a schematic diagram of another head portrait display method according to an embodiment of the present application;
FIG. 8 is a diagram of a user interface of another instant messaging tool according to an embodiment of the present application;
fig. 9 is a schematic diagram of a user avatar of another user in a chat group according to an embodiment of the application;
Fig. 10 is a schematic diagram of another head portrait display method according to an embodiment of the present application;
FIG. 11a is a schematic diagram of a user avatar of another user in a chat group according to an embodiment of the application;
fig. 11b is a schematic diagram of a user avatar of another user in a chat group according to an embodiment of the application.
Detailed Description
The technical scheme of the application will be described below with reference to the accompanying drawings.
In the following, some terms in the embodiments of the present application are explained for easy understanding by those skilled in the art.
The electronic device involved in the embodiments of the present application may be a mobile phone, tablet computer, desktop, laptop, notebook, ultra-mobile personal computer (ultra-mobile personal computer, UMPC), handheld computer, netbook, personal digital assistant (personal digital assistant, PDA), wearable electronic device, virtual reality device, etc.
Reference to "at least one" in embodiments of the application means one or more, and "plurality" means two or more. "and/or", describes an association relationship of an association object, and indicates that there may be three relationships, for example, a and/or B, and may indicate: a alone, a and B together, and B alone, wherein a, B may be singular or plural. The character "/" generally indicates that the context-dependent object is an "or" relationship. "at least one of" or the like means any combination of these items, including any combination of single item(s) or plural items(s). For example, at least one (one) of a, b, or c may represent: a, b, c, a-b, a-c, b-c, or a-b-c, wherein a, b, c may be single or plural.
And, unless specified to the contrary, references to "first," "second," etc. ordinal words of embodiments of the present application are used for distinguishing between multiple objects and are not used for limiting the order, timing, priority, or importance of the multiple objects. For example, the first information and the second information are only for distinguishing different information, and are not indicative of the difference in content, priority, transmission order, importance, or the like of the two information.
In order to facilitate understanding of the present application, concepts related to the present application will be explained first:
instant Message (IM): instant messaging refers to a service capable of sending and receiving internet messages and the like in real time. The user can send and receive messages through the instant messaging tool, and the conversation list containing the head portrait and the text is a common composition mode in the instant messaging tool and is used for bearing each message window. The head portraits of the chat group in the instant messaging tool are composed of the head portraits of the users in the chat group, and the composition mode of the head portraits of the users in the chat group is usually determined according to the joining sequence of the users in the chat group. The instant messaging tool may be, for example, a WeChat or the like.
Referring to fig. 1, fig. 1 is a schematic diagram of a user interface of an instant messaging tool. As shown in fig. 1, taking WeChat as an example, the user interface 100 includes a conversation list of a chat group and a conversation list of an individual chat, where the head portrait of the chat group is composed of user head portraits of users in the chat group, and the head portraits of the individual chat are user head portraits of the users of the individual chat. As can be seen from fig. 1, the user interface includes a large number of session lists, specifically including the session lists of chat group 1, chat group 2, and chat group 3, and the session lists of chat with user 1, user 2, user 3, and user 4 separately. In the user interface 100, the avatars of chat groups can only be used to quickly distinguish chat groups from individual chats, and the avatars of chat groups cannot be used for quick identification among multiple chat groups. Thus, the recognition efficiency between chat groups is low, and the user experience is poor.
Having described the background of the application as above, exemplary electronic devices provided in the following embodiments of the application are described below.
Fig. 2 shows a schematic structural diagram of an electronic device 200.
The electronic device 200 may include a processor 210, an external memory interface 220, an internal memory 221, a universal serial bus (universal serial bus, USB) interface 230, a charge management module 240, a power management module 241, a battery 242, an antenna 1, an antenna 2, a mobile communication module 250, a wireless communication module 260, an audio module 270, a speaker 270A, a receiver 270B, a microphone 270C, an ear-piece interface 270D, a sensor module 280, keys 290, a motor 291, an indicator 292, a camera 293, a display 294, and a subscriber identity module (subscriber identification module, SIM) card interface 295, among others. The sensor module 280 may include a pressure sensor 280A, a gyroscope sensor 280B, a barometric sensor 280C, a magnetic sensor 280D, an acceleration sensor 280E, a distance sensor 280F, a proximity sensor 280G, a fingerprint sensor 280H, a temperature sensor 280J, a touch sensor 280K, an ambient light sensor 280L, a bone conduction sensor 280M, and the like.
It should be understood that the structure illustrated in the embodiments of the present application does not constitute a specific limitation on the electronic device 200. In other embodiments of the application, electronic device 200 may include more or fewer components than shown, or certain components may be combined, or certain components may be separated, or different arrangements of components. The illustrated components may be implemented in hardware, software, or a combination of software and hardware.
Processor 210 may include one or more processing units such as, for example: the processor 210 may include an application processor (application processor, AP), a modem processor, a graphics processor (graphics processing unit, GPU), an image signal processor (image signal processor, ISP), a controller, a video codec, a digital signal processor (digital signal processor, DSP), a baseband processor, and/or a neural network processor (neural-network processing unit, NPU), etc. Wherein the different processing units may be separate devices or may be integrated in one or more processors.
The controller can generate operation control signals according to the instruction operation codes and the time sequence signals to finish the control of instruction fetching and instruction execution.
A memory may also be provided in the processor 210 for storing instructions and data. In some embodiments, the memory in the processor 210 is a cache memory. The memory may hold instructions or data that the processor 210 has just used or recycled. If the processor 210 needs to reuse the instruction or data, it may be called directly from the memory. Repeated accesses are avoided and the latency of the processor 210 is reduced, thereby improving the efficiency of the system.
In some embodiments, processor 210 may include one or more interfaces. The interfaces may include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, a pulse code modulation (pulse code modulation, PCM) interface, a universal asynchronous receiver transmitter (universal asynchronous receiver/transmitter, UART) interface, a mobile industry processor interface (mobile industry processor interface, MIPI), a general-purpose input/output (GPIO) interface, a subscriber identity module (subscriber identity module, SIM) interface, and/or a universal serial bus (universal serial bus, USB) interface, among others.
The I2C interface is a bi-directional synchronous serial bus comprising a serial data line (SDA) and a serial clock line (derail clock line, SCL). In some embodiments, the processor 210 may contain multiple sets of I2C buses. The processor 210 may be coupled to the touch sensor 280K, charger, flash, camera 293, etc., respectively, through different I2C bus interfaces. For example: the processor 210 may couple the touch sensor 280K through an I2C interface, so that the processor 210 communicates with the touch sensor 280K through an I2C bus interface to implement a touch function of the electronic device 200.
The I2S interface may be used for audio communication. In some embodiments, the processor 210 may contain multiple sets of I2S buses. The processor 210 may be coupled to the audio module 270 via an I2S bus to enable communication between the processor 210 and the audio module 270. In some embodiments, the audio module 270 may communicate audio signals to the wireless communication module 260 through the I2S interface to implement a function of answering a call through a bluetooth headset.
PCM interfaces may also be used for audio communication to sample, quantize and encode analog signals. In some embodiments, the audio module 270 and the wireless communication module 260 may be coupled by a PCM bus interface. In some embodiments, the audio module 270 may also transmit audio signals to the wireless communication module 260 through the PCM interface to implement a function of answering a call through the bluetooth headset. Both the I2S interface and the PCM interface may be used for audio communication.
The UART interface is a universal serial data bus for asynchronous communications. The bus may be a bi-directional communication bus. It converts the data to be transmitted between serial communication and parallel communication. In some embodiments, a UART interface is typically used to connect the processor 210 with the wireless communication module 260. For example: the processor 210 communicates with a bluetooth module in the wireless communication module 260 through a UART interface to implement a bluetooth function. In some embodiments, the audio module 270 may transmit an audio signal to the wireless communication module 260 through a UART interface, implementing a function of playing music through a bluetooth headset.
The MIPI interface may be used to connect the processor 210 to peripheral devices such as the display 294, the camera 293, and the like. The MIPI interfaces include camera serial interfaces (camera serial interface, CSI), display serial interfaces (display serial interface, DSI), and the like. In some embodiments, processor 210 and camera 293 communicate through a CSI interface to implement the photographing functions of electronic device 200. The processor 210 and the display 294 communicate via a DSI interface to implement the display functions of the electronic device 200.
The GPIO interface may be configured by software. The GPIO interface may be configured as a control signal or as a data signal. In some embodiments, a GPIO interface may be used to connect the processor 210 with the camera 293, display 294, wireless communication module 260, audio module 270, sensor module 280, and the like. The GPIO interface may also be configured as an I2C interface, an I2S interface, a UART interface, an MIPI interface, etc.
The USB interface 230 is an interface conforming to the USB standard specification, and may specifically be a Mini USB interface, a Micro USB interface, a USB Type C interface, or the like. The USB interface 230 may be used to connect a charger to charge the electronic device 200, or may be used to transfer data between the electronic device 200 and a peripheral device. And can also be used for connecting with a headset, and playing audio through the headset. The interface may also be used to connect other electronic devices, such as AR devices, etc.
It should be understood that the connection relationship between the modules illustrated in the embodiment of the present application is only illustrative, and does not limit the structure of the electronic device 200. In other embodiments of the present application, the electronic device 200 may also employ different interfacing manners, or a combination of interfacing manners, as in the above embodiments.
The charge management module 240 is configured to receive a charge input from a charger. The charger can be a wireless charger or a wired charger. In some wired charging embodiments, the charge management module 240 may receive a charging input of a wired charger through the USB interface 230. In some wireless charging embodiments, the charge management module 240 may receive wireless charging input through a wireless charging coil of the electronic device 200. The charging management module 240 may also provide power to the electronic device through the power management module 241 while charging the battery 242.
The power management module 241 is used for connecting the battery 242, and the charge management module 240 and the processor 210. The power management module 241 receives input from the battery 242 and/or the charge management module 240 and provides power to the processor 210, the internal memory 221, the display 294, the camera 293, the wireless communication module 260, and the like. The power management module 241 may also be configured to monitor battery capacity, battery cycle times, battery health (leakage, impedance), and other parameters. In other embodiments, the power management module 241 may also be disposed in the processor 210. In other embodiments, the power management module 241 and the charge management module 240 may be disposed in the same device.
The wireless communication function of the electronic device 200 can be implemented by the antenna 1, the antenna 2, the mobile communication module 250, the wireless communication module 260, a modem processor, a baseband processor, and the like.
The antennas 1 and 2 are used for transmitting and receiving electromagnetic wave signals. Each antenna in the electronic device 200 may be used to cover a single or multiple communication bands. Different antennas may also be multiplexed to improve the utilization of the antennas. For example: the antenna 1 may be multiplexed into a diversity antenna of a wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
The mobile communication module 250 may provide a solution for wireless communication including 2G/3G/4G/5G, etc., applied on the electronic device 200. The mobile communication module 250 may include at least one filter, switch, power amplifier, low noise amplifier (low noise amplifier, LNA), etc. The mobile communication module 250 may receive electromagnetic waves from the antenna 1, perform processes such as filtering, amplifying, and the like on the received electromagnetic waves, and transmit the processed electromagnetic waves to the modem processor for demodulation. The mobile communication module 250 can amplify the signal modulated by the modem processor, and convert the signal into electromagnetic waves through the antenna 1 to radiate. In some embodiments, at least some of the functional modules of the mobile communication module 250 may be disposed in the processor 210. In some embodiments, at least some of the functional modules of the mobile communication module 250 may be provided in the same device as at least some of the modules of the processor 210.
The modem processor may include a modulator and a demodulator. The modulator is used for modulating the low-frequency baseband signal to be transmitted into a medium-high frequency signal. The demodulator is used for demodulating the received electromagnetic wave signal into a low-frequency baseband signal. The demodulator then transmits the demodulated low frequency baseband signal to the baseband processor for processing. The low frequency baseband signal is processed by the baseband processor and then transferred to the application processor. The application processor outputs sound signals through an audio device (not limited to speaker 270A, receiver 270B, etc.), or displays images or video through display screen 294. In some embodiments, the modem processor may be a stand-alone device. In other embodiments, the modem processor may be provided in the same device as the mobile communication module 250 or other functional module, independent of the processor 210.
The wireless communication module 260 may provide solutions for wireless communication including wireless local area network (wireless local area networks, WLAN) (e.g., wireless fidelity (wireless fidelity, wi-Fi) network), bluetooth (BT), global navigation satellite system (global navigation satellite system, GNSS), frequency modulation (frequency modulation, FM), near field wireless communication technology (near field communication, NFC), infrared technology (IR), etc., as applied on the electronic device 200. The wireless communication module 260 may be one or more devices that integrate at least one communication processing module. The wireless communication module 260 receives electromagnetic waves via the antenna 2, modulates the electromagnetic wave signals, filters the electromagnetic wave signals, and transmits the processed signals to the processor 210. The wireless communication module 260 may also receive a signal to be transmitted from the processor 210, frequency modulate it, amplify it, and convert it to electromagnetic waves for radiation via the antenna 2.
In some embodiments, antenna 1 and mobile communication module 250 of electronic device 200 are coupled, and antenna 2 and wireless communication module 260 are coupled, such that electronic device 200 may communicate with a network and other devices via wireless communication techniques. The wireless communication techniques may include the Global System for Mobile communications (global system for mobile communications, GSM), general packet radio service (general packet radio service, GPRS), code division multiple access (code division multiple access, CDMA), wideband code division multiple access (wideband code division multiple access, WCDMA), time division code division multiple access (time-division code division multiple access, TD-SCDMA), long term evolution (long term evolution, LTE), BT, GNSS, WLAN, NFC, FM, and/or IR techniques, among others. The GNSS may include a global satellite positioning system (global positioning system, GPS), a global navigation satellite system (global navigation satellite system, GLONASS), a beidou satellite navigation system (beidou navigation satellite system, BDS), a quasi zenith satellite system (quasi-zenith satellite system, QZSS) and/or a satellite based augmentation system (satellite based augmentation systems, SBAS).
The electronic device 200 implements display functions through a GPU, a display screen 294, an application processor, and the like. The GPU is a microprocessor for image processing, and is connected to the display screen 294 and the application processor. The GPU is used to perform mathematical and geometric calculations for graphics rendering. Processor 210 may include one or more GPUs that execute program instructions to generate or change display information.
The display 294 is used to display images, videos, and the like. The display 294 includes a display panel. The display panel may employ a liquid crystal display (liquid crystal display, LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode (AMOLED) or an active-matrix organic light-emitting diode (matrix organic light emitting diode), a flexible light-emitting diode (flex), a mini, a Micro led, a Micro-OLED, a quantum dot light-emitting diode (quantum dot light emitting diodes, QLED), or the like. In some embodiments, the electronic device 200 may include 1 or N display screens 294, N being a positive integer greater than 1.
The electronic device 200 may implement a photographing function through an ISP, a camera 293, a video codec, a GPU, a display 294, an application processor, and the like.
The ISP is used to process the data fed back by the camera 293. For example, when photographing, the shutter is opened, light is transmitted to the camera photosensitive element through the lens, the optical signal is converted into an electric signal, and the camera photosensitive element transmits the electric signal to the ISP for processing and is converted into an image visible to naked eyes. ISP can also optimize the noise, brightness and skin color of the image. The ISP can also optimize parameters such as exposure, color temperature and the like of a shooting scene. In some embodiments, the ISP may be provided in the camera 293.
The camera 293 is used to capture still images or video. The object generates an optical image through the lens and projects the optical image onto the photosensitive element. The photosensitive element may be a charge coupled device (charge coupled device, CCD) or a Complementary Metal Oxide Semiconductor (CMOS) phototransistor. The photosensitive element converts the optical signal into an electrical signal, which is then transferred to the ISP to be converted into a digital image signal. The ISP outputs the digital image signal to the DSP for processing. The DSP converts the digital image signal into an image signal in a standard RGB, YUV, or the like format. In some embodiments, the electronic device 200 may include 1 or N cameras 293, N being a positive integer greater than 1.
The digital signal processor is used for processing digital signals, and can process other digital signals besides digital image signals. For example, when the electronic device 200 is selecting a frequency bin, the digital signal processor is used to fourier transform the frequency bin energy, or the like.
Video codecs are used to compress or decompress digital video. The electronic device 200 may support one or more video codecs. In this way, the electronic device 200 may play or record video in a variety of encoding formats, such as: dynamic picture experts group (moving picture experts group, MPEG) 1, MPEG2, MPEG3, MPEG4, etc.
The NPU is a neural-network (NN) computing processor, and can rapidly process input information by referencing a biological neural network structure, for example, referencing a transmission mode between human brain neurons, and can also continuously perform self-learning. Applications such as intelligent cognition of the electronic device 200 may be implemented by the NPU, for example: image recognition, face recognition, speech recognition, text understanding, etc.
The external memory interface 220 may be used to connect an external memory card, such as a Micro SD card, to enable expansion of the memory capabilities of the electronic device 200. The external memory card communicates with the processor 210 through an external memory interface 220 to implement data storage functions. For example, files such as music, video, etc. are stored in an external memory card.
Internal memory 221 may be used to store computer executable program code that includes instructions. The internal memory 221 may include a storage program area and a storage data area. The storage program area may store an application program (such as a sound playing function, an image playing function, etc.) required for at least one function of the operating system, etc. The storage data area may store data created during use of the electronic device 200 (e.g., audio data, phonebook, etc.), and so on. In addition, the internal memory 221 may include a high-speed random access memory, and may further include a nonvolatile memory such as at least one magnetic disk storage device, a flash memory device, a universal flash memory (universal flash storage, UFS), and the like. The processor 210 performs various functional applications of the electronic device 200 and data processing by executing instructions stored in the internal memory 221 and/or instructions stored in a memory provided in the processor.
The electronic device 200 may implement audio functions through an audio module 270, a speaker 270A, a receiver 270B, a microphone 270C, an ear-headphone interface 270D, an application processor, and the like. Such as music playing, recording, etc.
The audio module 270 is used to convert digital audio information into an analog audio signal output and also to convert an analog audio input into a digital audio signal. The audio module 270 may also be used to encode and decode audio signals. In some embodiments, the audio module 270 may be disposed in the processor 210, or some functional modules of the audio module 270 may be disposed in the processor 210.
Speaker 270A, also referred to as a "horn," is used to convert audio electrical signals into sound signals. The electronic device 200 may listen to music, or to hands-free conversations, through the speaker 270A.
A receiver 270B, also referred to as a "earpiece", is used to convert the audio electrical signal into a sound signal. When electronic device 200 is answering a telephone call or voice message, voice may be received by placing receiver 270B close to the human ear.
Microphone 270C, also referred to as a "microphone" or "microphone," is used to convert sound signals into electrical signals. When making a call or transmitting voice information, the user can sound near the microphone 270C through the mouth, inputting a sound signal to the microphone 270C. The electronic device 200 may be provided with at least one microphone 270C. In other embodiments, the electronic device 200 may be provided with two microphones 270C, and may implement a noise reduction function in addition to collecting sound signals. In other embodiments, the electronic device 200 may also be provided with three, four, or more microphones 270C to enable collection of sound signals, noise reduction, identification of sound sources, directional recording, etc.
The earphone interface 270D is for connecting a wired earphone. Earphone interface 270D may be USB interface 230 or a 3.5mm open mobile electronic device platform (open mobile terminal platform, OMTP) standard interface, american cellular telecommunications industry association (cellular telecommunications industry association of the USA, CTIA) standard interface.
The pressure sensor 280A is used to sense a pressure signal, and may convert the pressure signal into an electrical signal. In some embodiments, pressure sensor 280A may be disposed on display 294. The pressure sensor 280A is of various types, such as a resistive pressure sensor, an inductive pressure sensor, a capacitive pressure sensor, and the like. The capacitive pressure sensor may be a capacitive pressure sensor comprising at least two parallel plates with conductive material. When a force is applied to the pressure sensor 280A, the capacitance between the electrodes changes. The electronic device 200 determines the strength of the pressure from the change in capacitance. When a touch operation is applied to the display panel 294, the electronic apparatus 200 detects the touch operation intensity from the pressure sensor 280A. The electronic device 200 may also calculate the location of the touch based on the detection signal of the pressure sensor 280A. In some embodiments, touch operations that act on the same touch location, but at different touch operation strengths, may correspond to different operation instructions. For example: and executing an instruction for checking the short message when the touch operation with the touch operation intensity smaller than the first pressure threshold acts on the short message application icon. And executing an instruction for newly creating the short message when the touch operation with the touch operation intensity being greater than or equal to the first pressure threshold acts on the short message application icon.
The gyro sensor 280B may be used to determine a motion gesture of the electronic device 200. In some embodiments, the angular velocity of electronic device 200 about three axes (i.e., x, y, and z axes) may be determined by gyro sensor 280B. The gyro sensor 280B may be used for photographing anti-shake. For example, when the shutter is pressed, the gyro sensor 280B detects the shake angle of the electronic device 200, calculates the distance to be compensated by the lens module according to the angle, and makes the lens counteract the shake of the electronic device 200 through the reverse motion, thereby realizing anti-shake. The gyro sensor 280B may also be used for navigating, somatosensory game scenes.
The air pressure sensor 280C is used to measure air pressure. In some embodiments, the electronic device 200 calculates altitude from barometric pressure values measured by the barometric pressure sensor 280C, aiding in positioning and navigation.
The magnetic sensor 280D includes a hall sensor. The electronic device 200 may detect the opening and closing of the flip holster using the magnetic sensor 280D. In some embodiments, when the electronic device 200 is a flip machine, the electronic device 200 may detect the opening and closing of the flip according to the magnetic sensor 280D. And then according to the detected opening and closing state of the leather sheath or the opening and closing state of the flip, the characteristics of automatic unlocking of the flip and the like are set.
The acceleration sensor 280E may detect the magnitude of acceleration of the electronic device 200 in various directions (typically three axes). The magnitude and direction of gravity may be detected when the electronic device 200 is stationary. The electronic equipment gesture recognition method can also be used for recognizing the gesture of the electronic equipment, and is applied to horizontal and vertical screen switching, pedometers and other applications.
A distance sensor 280F for measuring distance. The electronic device 200 may measure the distance by infrared or laser. In some embodiments, the electronic device 200 may range using the distance sensor 280F to achieve quick focus.
Proximity light sensor 280G may include, for example, a Light Emitting Diode (LED) and a light detector, such as a photodiode. The light emitting diode may be an infrared light emitting diode. The electronic device 200 emits infrared light outward through the light emitting diode. The electronic device 200 detects infrared reflected light from nearby objects using a photodiode. When sufficient reflected light is detected, it may be determined that an object is in the vicinity of the electronic device 200. When insufficient reflected light is detected, the electronic device 200 may determine that there is no object in the vicinity of the electronic device 200. The electronic device 200 can detect that the user holds the electronic device 200 close to the ear by using the proximity light sensor 280G, so as to automatically extinguish the screen for the purpose of saving power. The proximity light sensor 280G may also be used in holster mode, pocket mode to automatically unlock and lock the screen.
The ambient light sensor 280L is used to sense ambient light level. The electronic device 200 may adaptively adjust the brightness of the display 294 based on the perceived ambient light level. The ambient light sensor 280L may also be used to automatically adjust white balance during photographing. Ambient light sensor 280L may also cooperate with proximity light sensor 280G to detect whether electronic device 200 is in a pocket to prevent false touches.
The fingerprint sensor 280H is used to collect a fingerprint. The electronic device 200 can utilize the collected fingerprint characteristics to realize fingerprint unlocking, access an application lock, fingerprint photographing, fingerprint incoming call answering and the like.
The temperature sensor 280J is used to detect temperature. In some embodiments, the electronic device 200 performs a temperature processing strategy using the temperature detected by the temperature sensor 280J. For example, when the temperature reported by temperature sensor 280J exceeds a threshold, electronic device 200 performs a reduction in the performance of a processor located in the vicinity of temperature sensor 280J in order to reduce power consumption to implement thermal protection. In other embodiments, when the temperature is below another threshold, the electronic device 200 heats the battery 242 to avoid the low temperature causing the electronic device 200 to be abnormally shut down. In other embodiments, when the temperature is below a further threshold, the electronic device 200 performs boosting of the output voltage of the battery 242 to avoid abnormal shutdown caused by low temperatures.
The touch sensor 280K, also referred to as a "touch device". The touch sensor 280K may be disposed on the display screen 294, and the touch sensor 280K and the display screen 294 form a touch screen, which is also referred to as a "touch screen". The touch sensor 280K is used to detect a touch operation acting on or near it. The touch sensor may communicate the detected touch operation to the application processor to determine the touch event type. Visual output related to touch operations may be provided through the display 294. In other embodiments, the touch sensor 280K may also be disposed on the surface of the electronic device 200 at a different location than the display 294.
Bone conduction sensor 280M may acquire a vibration signal. In some embodiments, bone conduction sensor 280M may acquire a vibration signal of a human vocal tract vibrating bone pieces. The bone conduction sensor 280M may also contact the pulse of the human body to receive the blood pressure pulsation signal. In some embodiments, bone conduction sensor 280M may also be provided in a headset, in combination with an osteoinductive headset. The audio module 270 may analyze the voice signal based on the vibration signal of the sound portion vibration bone piece obtained by the bone conduction sensor 280M, so as to implement the voice function. The application processor can analyze heart rate information based on the blood pressure beat signal acquired by the bone conduction sensor 280M, so as to realize a heart rate detection function.
Keys 290 include a power on key, a volume key, etc. The keys 290 may be mechanical keys. Or may be a touch key. The electronic device 200 may receive key inputs, generating key signal inputs related to user settings and function controls of the electronic device 200.
The motor 291 may generate a vibration alert. The motor 291 may be used for incoming call vibration alerting or for touch vibration feedback. For example, touch operations acting on different applications (e.g., photographing, audio playing, etc.) may correspond to different vibration feedback effects. The motor 291 may also correspond to different vibration feedback effects by touch operations applied to different areas of the display 294. Different application scenarios (such as time reminding, receiving information, alarm clock, game, etc.) can also correspond to different vibration feedback effects. The touch vibration feedback effect may also support customization.
The indicator 292 may be an indicator light, which may be used to indicate a state of charge, a change in power, a message indicating a missed call, a notification, etc.
The SIM card interface 295 is for interfacing with a SIM card. The SIM card may be inserted into the SIM card interface 295 or removed from the SIM card interface 295 to enable contact and separation from the electronic device 200. The electronic device 200 may support 1 or N SIM card interfaces, N being a positive integer greater than 1. The SIM card interface 295 may support Nano SIM cards, micro SIM cards, and the like. The same SIM card interface 295 may be used to insert multiple cards simultaneously. The types of the plurality of cards may be the same or different. The SIM card interface 295 may also be compatible with different types of SIM cards. The SIM card interface 295 may also be compatible with external memory cards. The electronic device 200 interacts with the network through the SIM card to realize functions such as communication and data communication. In some embodiments, the electronic device 200 employs esims, namely: an embedded SIM card. The eSIM card can be embedded in the electronic device 200 and cannot be separated from the electronic device 100.
The software system of the electronic device 200 may employ a layered architecture, an event driven architecture, a microkernel architecture, a microservice architecture, or a cloud architecture. In the embodiment of the invention, taking an Android system with a layered architecture as an example, a software structure of the electronic device 200 is illustrated.
Referring to fig. 3, fig. 3 is a software architecture block diagram of an electronic device 200 according to an embodiment of the invention.
The layered architecture divides the software into several layers, each with distinct roles and branches. The layers communicate with each other through a software interface. In some embodiments, the Android system is divided into four layers, from top to bottom, an application layer, an application framework layer, an Zhuoyun row (Android run) and system libraries, and a kernel layer, respectively.
The application layer may include a series of application packages.
As shown in fig. 3, the application package may include applications for cameras, gallery, calendar, phone calls, maps, navigation, WLAN, bluetooth, music, video, short messages, etc.
The application framework layer provides an application programming interface (application programming interface, API) and programming framework for application programs of the application layer. The application framework layer includes a number of predefined functions.
As shown in FIG. 3, the application framework layer may include a window manager, a content provider, a view system, a telephony manager, a resource manager, a notification manager, and the like.
The window manager is used for managing window programs. The window manager can acquire the size of the display screen, judge whether a status bar exists, lock the screen, intercept the screen and the like.
The content provider is used to store and retrieve data and make such data accessible to applications. The data may include video, images, audio, calls made and received, browsing history and bookmarks, phonebooks, etc.
The view system includes visual controls, such as controls to display text, controls to display pictures, and the like. The view system may be used to build applications. The display interface may be composed of one or more views. For example, a display interface including a text message notification icon may include a view displaying text and a view displaying a picture.
The telephony manager is used to provide the communication functions of the electronic device 200. Such as the management of call status (including on, hung-up, etc.).
The resource manager provides various resources for the application program, such as localization strings, icons, pictures, layout files, video files, and the like.
The notification manager allows the application to display notification information in a status bar, can be used to communicate notification type messages, can automatically disappear after a short dwell, and does not require user interaction. Such as notification manager is used to inform that the download is complete, message alerts, etc. The notification manager may also be a notification in the form of a chart or scroll bar text that appears on the system top status bar, such as a notification of a background running application, or a notification that appears on the screen in the form of a dialog window. For example, a text message is prompted in a status bar, a prompt tone is emitted, the electronic device vibrates, and an indicator light blinks, etc.
Android run time includes a core library and virtual machines. Android run time is responsible for scheduling and management of the Android system.
The core library consists of two parts: one part is a function which needs to be called by java language, and the other part is a core library of android.
The application layer and the application framework layer run in a virtual machine. The virtual machine executes java files of the application program layer and the application program framework layer as binary files. The virtual machine is used for executing the functions of object life cycle management, stack management, thread management, security and exception management, garbage collection and the like.
The system library may include a plurality of functional modules. For example: surface manager (surface manager), media Libraries (Media Libraries), three-dimensional graphics processing Libraries (e.g., openGL ES), 2D graphics engines (e.g., SGL), etc.
The surface manager is used to manage the display subsystem and provides a fusion of 2D and 3D layers for multiple applications.
Media libraries support a variety of commonly used audio, video format playback and recording, still image files, and the like. The media library may support a variety of audio and video encoding formats, such as MPEG4, h.264, MP3, AAC, AMR, JPG, PNG, etc.
The three-dimensional graphic processing library is used for realizing three-dimensional graphic drawing, image rendering, synthesis, layer processing and the like.
The 2D graphics engine is a drawing engine for 2D drawing.
The kernel layer is a layer between hardware and software. The inner core layer at least comprises a display driver, a camera driver, an audio driver and a sensor driver.
The workflow of the electronic device 200 software and hardware is illustrated below in connection with capturing a photo scene.
When touch sensor 280K receives a touch operation, a corresponding hardware interrupt is issued to the kernel layer. The kernel layer processes the touch operation into the original input event (including information such as touch coordinates, time stamp of touch operation, etc.). The original input event is stored at the kernel layer. The application framework layer acquires an original input event from the kernel layer, and identifies a control corresponding to the input event. Taking the touch operation as a touch click operation, taking a control corresponding to the click operation as an example of a control of a camera application icon, the camera application calls an interface of an application framework layer, starts the camera application, further starts a camera driver by calling a kernel layer, and captures a still image or video through a camera 293.
Referring to fig. 4, fig. 4 is a schematic diagram of a head portrait display method according to an embodiment of the present application. As shown in fig. 4, the avatar display method is applied to an electronic device, which may be the electronic device 200 shown in fig. 2, and includes:
s401, the activity of a first user in a chat group in an instant messaging tool in the chat group is obtained.
Optionally, the activity of the first user in the chat group includes a frequency of sending a message by the first user in the chat group and/or an interaction frequency of the first user and the second user in the chat group, where the second user is an application user of the instant messaging tool.
Specifically, after the chat group in the instant messaging tool is established, the electronic device may obtain the activity of the first user in the chat group in a preset time period. The preset time period may be preset, for example, may be one week or one month, and when the preset time period arrives, the electronic device automatically obtains the activity of the first user in the chat group.
When the activity of the first user in the chat group is the frequency with which the first user sends messages in the chat group, the first user may be any one or more users in the chat group, including an application user of an instant messaging tool. When the activity of the first user in the chat group is the interaction frequency of the first user and the second user in the chat group, the second user is the application user of the instant messaging tool, and the first user is one or more other users except the second user in the chat group.
S402, determining the size of a display area of a user head portrait of the first user in the chat group according to the activity.
Optionally, when the activity of the first user in the chat group is highest, determining that the display area of the user head portrait of the first user in the chat group is largest.
In one possible implementation manner, when the activity of the first user in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is equal to the display area of the user head portrait of a third user in the chat group, where the third user is other users except the user with the highest activity in the chat group.
In another possible implementation manner, when the activity of the first user in the chat group is not the highest, it is determined that the display area of the user avatar of the first user in the chat group is larger than the display area of the user avatar of a fourth user in the chat group, where the fourth user is a user whose activity in the chat group is lower than the activity of the first user in the chat group.
Specifically, when the activity of the first user in the chat group is highest, the display area of the user head portrait of the first user in the chat group is determined to be the largest, so that the first user can be distinguished from other users except the first user in the chat group.
When the activity of the first user in the chat group is not the highest, determining that the display area of the user head portrait in the chat group of the user with the highest activity is the largest. The other users (including the first user) except the user with the highest activity in the chat group have the same display area of the user head portrait in the chat group, so that the user with the highest activity in the chat group can be distinguished from the other users except the user with the highest activity. Or, other users (including the first user) except the user with the highest activity in the chat group, wherein the display area of the user head portrait in the chat group is adjusted in size according to the activity, and the higher the activity is, the larger the display area is. Thus, not only the user with the highest activity in the chat group can be distinguished from other users except the user with the highest activity, but also other users except the user with the highest activity in the chat group can be distinguished.
S403, displaying the user head portraits of the first user in the chat group according to the size of the display area.
Specifically, according to the size of the display area, displaying the user head portrait of the first user in the chat group. That is, for the application user of the instant messaging tool, on the user interface of the instant messaging tool, the user with the largest display area of the user avatar in the chat group is the user with the highest activity in the chat group. For the same chat group, on the user interfaces of the instant messaging tools of all users in the chat group, the user with the largest display area of the user head portrait in the chat group is the user with the highest activity in the chat group. For different chat groups, the users with highest liveness may be different, so that on the user interface of the instant messaging tool, the users with the largest display areas of the user head portraits in the different chat groups are also different, and therefore, the different chat groups can be rapidly distinguished.
In one possible implementation, the activity of a first user in a chat group is the frequency with which the first user sends messages in the chat group. That is, for the same chat group, on the user interface of the instant messaging tool of all users in the chat group, the user with the largest display area of the user head portrait in the chat group is the same user with the highest frequency of sending messages in the chat group.
In another possible embodiment, the activity of the first user in the chat group is the interaction frequency of the first user and the second user in the chat group, and the second user is the application user of the instant messaging tool. That is, for the same chat group, the user with the largest display area of the user head portrait in the chat group is the user with the highest interaction frequency with the user on the user interface of the instant messaging tool of one user in the chat group. In the instant messaging tool, the user with the highest interaction frequency with a certain user in the chat group may be different, so that the user with the largest display area of the user head portrait in the chat group may be different on the user interface of the instant messaging tool of all users in the chat group.
According to the method, the activity of the user in the chat group is obtained, and the size of the display area of the user head portrait in the chat group is determined according to the activity, so that the size of the display area of the user head portrait in the chat group is intelligently adjusted. When a plurality of chat groups exist in the instant messaging tool, the sizes of the display areas of the user head portraits determined according to the liveness in different chat groups are different, the head portraits of the chat groups are large in difference, and a user can quickly identify the chat groups to be searched by using the head portraits of the chat groups. Therefore, the recognition efficiency between chat groups is improved, and the user experience is improved.
Referring to fig. 5, fig. 5 is a schematic diagram of a user interface of an instant messaging tool according to an embodiment of the present application. As shown in fig. 5, in the example of Wechat, the user interface 500 includes a list of sessions for a chat group and a list of sessions for an individual chat. As can be seen from fig. 5, the user interface includes a large number of session lists, specifically including the session lists of chat group 1 and chat group 2, and the session lists of chat with user 1, user 2, user 3, user 4, and user 5 separately.
In the user interface 500, the head portraits of the chat group 1 are composed of user head portraits of 6 users, and the display area of the user head portraits of one user is the largest, the display areas of the user head portraits of other 5 users are the same, and the user with the largest display area of the user head portraits is the user with the highest activity in the chat group 1. The head portraits of the chat group 2 are composed of user head portraits of 3 users, wherein the display area of the user head portraits of one user is the largest, the display areas of the user head portraits of other 2 users are the same, and the user with the largest display area of the user head portraits is the user with the highest activity in the chat group 2.
It can be seen that in the user interface 500 shown in fig. 5, the avatars of chat groups can be used to quickly distinguish between chat groups and individual chats, but also between chat group 1 and chat group 2 can be quickly identified using the avatars of chat groups. Therefore, the recognition efficiency between chat groups is improved, and the user experience is improved.
Referring to fig. 6a and fig. 6b, fig. 6a is a schematic diagram of a user head portrait of a user in a chat group according to an embodiment of the present application, and fig. 6b is a schematic diagram of a user head portrait of another user in a chat group according to an embodiment of the present application. As shown in fig. 6a and 6B, the head portraits of the chat group are composed of user head portraits of a user a, a user B, a user C and a user D, wherein the activity of the user a in the chat group is highest, the activity of the user B in the chat group is ranked second, and the activity of the user C and the user D in the chat group are the same. In fig. 6a, the user a has the largest display area of the user avatar in the chat group, and the user B, user C, and user D have not the same activity, but the display area of the user avatar in the chat group is equal. In fig. 6B, the display area of the user head portraits of user a in the chat group is largest, the display area of the user head portraits of user B in the chat group is smaller than the display area of the user head portraits of user a in the chat group, but larger than the display areas of the user head portraits of user C and user D in the chat group.
Referring to fig. 7, fig. 7 is a schematic diagram of another head portrait display method according to an embodiment of the present application. As shown in fig. 7, the avatar display method is applied to an electronic device, which may be the electronic device 200 shown in fig. 2, and includes:
S701, obtaining the frequency of sending messages in the chat group by a first user in the chat group in the instant messaging tool.
Specifically, after the chat group in the instant messaging tool is established, the electronic device may obtain a frequency of sending a message in the chat group by the first user in the chat group in a preset time period. The preset time period may be preset, for example, may be one week or one month, and when the preset time period arrives, the electronic device automatically obtains the frequency of sending the message by the first user in the chat group. The first user may be any one or more users in the chat group, including application users of instant messaging tools.
S702, determining the size of a display area of a user head portrait of the first user in the chat group according to the frequency of the first user sending messages in the chat group.
In one possible implementation, when the frequency of sending messages by the first user in the chat group is highest, it is determined that the display area of the user avatar of the first user in the chat group is largest. When the frequency of the first user sending the message in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is equal to the display area of the user head portrait of a third user in the chat group, wherein the third user is other users except the user with the highest message sending frequency in the chat group.
For example, the chat group includes a user a, a user B, and a user C, where the frequency of sending messages in the chat group by the user a is highest, and the frequency of sending messages in the chat group by the user B and the user C is different and is lower than that of the user a. The display area of the user head portraits of the user a in the chat group is largest, and the display areas of the user head portraits of the user B and the user C in the chat group are equal to each other regardless of whether the frequency of the messages sent by the user B in the chat group is higher or lower than the frequency of the messages sent by the user C in the chat group.
In another possible implementation, when the frequency of sending messages by the first user in the chat group is highest, it is determined that the display area of the user avatar of the first user in the chat group is largest. When the frequency of the first user sending the messages in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is larger than the display area of the user head portrait of a fourth user in the chat group, wherein the fourth user is a user with the frequency of sending the messages in the chat group being lower than the frequency of sending the messages in the chat group by the first user.
For example, the chat group includes a user a, a user B, and a user C, where the frequency of sending messages in the chat group by the user a is highest, the frequency of sending messages in the chat group by the user B and the user C is lower than the frequency of sending messages in the chat group by the user a, and the frequency of sending messages in the chat group by the user B is higher than the frequency of sending messages in the chat group by the user C. The display area of the user head portrait of the user a in the chat group is largest, and the display area of the user head portrait of the user B in the chat group is smaller than the display area of the user head portrait of the user a in the chat group, but larger than the display area of the user head portrait of the user C in the chat group.
Specifically, when the frequency of sending messages by the first user in the chat group is highest, the display area of the user head portrait of the first user in the chat group is determined to be the largest, so that the first user can be distinguished from other users except the first user in the chat group.
When the frequency of the first user sending the messages in the chat group is not the highest, the display area of the user head portrait in the chat group of the user with the highest frequency of sending the messages is determined to be the largest. The other users (including the first user) of the chat group except the user who sends the message with the highest frequency are large in the display area of the user head portrait of the chat group, so that the user who sends the message with the highest frequency in the chat group can be distinguished from the other users except the user who sends the message with the highest frequency. Or, other users (including the first user) except the user with the highest message sending frequency in the chat group, the display area of the user head portrait in the chat group is adjusted in size according to the message sending frequency, wherein the higher the message sending frequency is, the larger the display area is. This allows not only the most frequent users of the chat group that send messages to be distinguished from other users than the most frequent users that send messages, but also the other users of the chat group that are not the most frequent users that send messages to be distinguished.
S703, displaying the user head of the first user in the chat group according to the size of the display area.
Specifically, according to the size of the display area, displaying the user head portrait of the first user in the chat group. That is, for the application user of the instant messaging tool, on the user interface of the instant messaging tool, the user with the largest display area of the user avatar in the chat group is the user with the highest frequency of sending messages in the chat group.
For the same chat group, on the user interface of the instant messaging tool of all users in the chat group, the user with the largest display area of the user head portrait in the chat group is the same user with the highest message sending frequency in the chat group. For different chat groups, the user with highest message sending frequency may be different, so that on the user interface of the instant messaging tool, the user with the largest display area of the user head portrait in the different chat groups is also different, and thus, the different chat groups can be rapidly distinguished.
S704, determining the arrangement sequence of the user head images of the first user in the chat group according to the frequency of the first user sending the messages in the chat group.
In one possible implementation, when the frequency of sending messages by the first user in the chat group is highest, it is determined that the ranking order of the user head portraits of the first user in the chat group is the forefront. When the frequency of the first user sending the message in the chat group is not the highest, the front-back relation between the arrangement sequence of the user head images of the first user in the chat group and the arrangement sequence of the user head images of a third user in the chat group is not limited, wherein the third user is other users except the user with the highest message sending frequency in the chat group.
For example, the chat group includes a user a, a user B, and a user C, where the frequency of sending messages in the chat group by the user a is highest, and the frequency of sending messages in the chat group by the user B and the user C is different and is lower than that of the user a. The order of the user's head portraits in the chat group is foremost, and whether the frequency of the user's B sending messages in the chat group is higher or lower than the frequency of the user's C sending messages in the chat group, the order of the user's head portraits in the chat group may be earlier or later than the order of the user's C head portraits in the chat group.
In another possible implementation, when the frequency of sending messages by the first user in the chat group is highest, it is determined that the ranking order of the user head portraits of the first user in the chat group is the forefront. When the frequency of the first user sending the messages in the chat group is not the highest, determining that the arrangement sequence of the user head images of the first user in the chat group is earlier than the arrangement sequence of the user head images of a fourth user in the chat group, wherein the fourth user is a user with the frequency of sending the messages in the chat group being lower than the frequency of sending the messages in the chat group by the first user.
For example, the chat group includes a user a, a user B, and a user C, where the frequency of sending messages in the chat group by the user a is highest, the frequency of sending messages in the chat group by the user B and the user C is lower than the frequency of sending messages in the chat group by the user a, and the frequency of sending messages in the chat group by the user B is higher than the frequency of sending messages in the chat group by the user C. The arrangement order of the user head images of the user A in the chat group is the forefront, and the arrangement order of the user head images of the user B in the chat group is more front than the arrangement order of the user head images of the user C in the chat group.
S705, displaying the user head of the first user in the chat group according to the arrangement order.
Specifically, according to the arrangement sequence, user head portraits of the first user in the chat group are displayed. That is, for the application user of the instant messaging tool, on the user interface of the instant messaging tool, the user whose head is arranged in the chat group is the most frequent user to send messages in the chat group.
For the same chat group, on the user interface of the instant messaging tool of all users in the chat group, the user with the head portrait arranged in the first order in the chat group is the same user with the highest message sending frequency in the chat group. For different chat groups, the users with highest message sending frequency may be different, so that on the user interface of the instant messaging tool, the users with forefront user head images in different chat groups are also different, and therefore different chat groups can be rapidly distinguished.
In one possible implementation, when there are more users in the chat group, before determining the size and/or arrangement order of the display area of the user head portraits in the chat group, it is further necessary to determine how often the first user sends messages in the chat group. When the frequency of the first user sending the message in the chat group is not less than a preset frequency threshold, determining the size and/or arrangement sequence of the display area of the user head portrait of the first user in the chat group, and then displaying the user head portrait of the first user in the chat group. And when the frequency of the first user sending the messages in the chat group is smaller than a preset frequency threshold, not displaying the user head portrait of the first user in the chat group.
In the method, the frequency of sending the message by the user in the chat group is obtained, and the size and the arrangement sequence of the display area of the user head portrait in the chat group are determined according to the frequency of sending the message, so that the size and the arrangement sequence of the display area of the user head portrait in the chat group are intelligently adjusted. When a plurality of chat groups exist in the instant messaging tool, the sizes and the arrangement sequences of the display areas of the user head portraits determined according to the frequency of sending messages in different chat groups are different, the head portraits of the chat groups have large difference, and a user can quickly identify the chat groups to be searched by using the head portraits of the chat groups. Therefore, the recognition efficiency between chat groups is improved, and the user experience is improved.
Referring to fig. 8, fig. 8 is a schematic diagram of a user interface of another instant messaging tool according to an embodiment of the present application. As shown in fig. 8, in the example of Wechat, a list of sessions for a chat group and a list of sessions for an individual chat are included in user interface 800. As can be seen from fig. 8, the user interface includes a large number of session lists, specifically including the session lists of chat group 1 and chat group 2, and the session lists of chat with user 1, user 2, user 3, user 4, and user 5 separately.
In the user interface 800, the avatar of the chat group 1 is composed of user avatars of 4 users, and a display area of the user avatar of one of the users is largest and is arranged at the forefront. The display area of the user head portraits of the other 3 users is equal, the display area of the user head portraits is the largest, and the users arranged at the front are the users with the highest frequency of sending messages in the chat group 1. The head portraits of chat group 2 are made up of user head portraits of 6 users, and the display area of the user head portraits of one of the users is largest and arranged at the forefront. Wherein the display area of the user head portraits of 2 users is equal and the arrangement order is earlier. Wherein the display area of the user head portraits of the 3 users is the smallest, and the arrangement order is at the last. The user head has the largest display area and the most front users are the most frequent users in chat group 2 to send messages.
It can be seen that in the user interface 800 shown in fig. 8, the avatars of chat groups can be used to quickly distinguish between chat groups and individual chats, but also between chat group 1 and chat group 2 can be quickly identified using the avatars of chat groups. Therefore, the recognition efficiency between chat groups is improved, and the user experience is improved.
Referring to fig. 9, fig. 9 is a schematic diagram of a user head portrait of another user in a chat group according to an embodiment of the present application. As shown in fig. 9, the head portraits of the chat group are composed of the head portraits of the users a, B, C, D, E and F, where the frequency of sending messages in the chat group is highest for the users a and B. In fig. 9, the display area of the user head portraits of the user a and the user B in the chat group is larger than the display area of the user head portraits of other users in the chat group and is arranged at a front position, and the frequency of the user C, the user D, the user E, and the user F transmitting messages is not exactly the same, but the display area of the user head portraits in the chat group is equal and is arranged at a rear position.
Referring to fig. 10, fig. 10 is a schematic diagram of another head portrait display method according to an embodiment of the present application. As shown in fig. 10, the avatar display method is applied to an electronic device, which may be the electronic device 200 shown in fig. 2, and includes:
s1001, acquiring interaction frequency of a first user and a second user in a chat group in the instant messaging tool.
Specifically, the second user is an application user of the instant messaging tool, and after the chat group in the instant messaging tool is established, the electronic device can obtain the interaction frequency between the first user and the second user in the chat group in a preset time period. The preset time period may be preset, for example, may be one week or one month, and when the preset time period arrives, the electronic device automatically obtains the interaction frequency of the first user and the second user in the chat group. The first user is one or more other users in the chat group than the second user.
The frequency of interaction of the first user with the second user in the chat group comprises at least one of: the frequency of the second user replying to the first user's message in the chat group, the frequency of the second user forwarding the first user's message sent in the chat group outside the chat group, the frequency of the second user designating the first user to chat in the chat group, the frequency of the second user referencing the first user's message sent in the chat group.
In one possible implementation, the interaction frequency of the first user and the second user outside the chat group, such as the frequency at which the second user sends messages to the first user alone, may also be obtained.
S1002, determining the display area size of the user head portrait of the first user in the chat group according to the interaction frequency of the first user and the second user in the chat group.
In one possible implementation, when the interaction frequency of the first user and the second user in the chat group is highest, it is determined that the display area of the user head portrait of the first user in the chat group is largest. When the interaction frequency of the first user and the second user in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is equal to the display area of the user head portrait of a third user in the chat group, wherein the third user is other users except the user with the highest interaction frequency with the second user in the chat group.
For example, the chat group includes a user a, a user B, and a user C, where the interaction frequency between the user a and the user B in the chat group is the highest, and the interaction frequency between the user B and the user C in the chat group is the highest. Then, for the user a, when the user a is an application user of the instant messaging tool, the display area of the user head portrait of the user B in the chat group is the largest on the user interface of the instant messaging tool of the user a. When the user B is an application user of the instant messaging tool, the display area of the user head portrait of the user C in the chat group is the largest on the user interface of the instant messaging tool of the user B.
In another possible implementation, when the interaction frequency of the first user and the second user in the chat group is highest, it is determined that the display area of the user head portrait of the first user in the chat group is largest. When the interaction frequency of the first user and the second user in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is larger than the display area of the user head portrait of a fourth user in the chat group, wherein the fourth user is a user with the interaction frequency of the second user in the chat group lower than the interaction frequency of the first user and the second user in the chat group.
For example, the chat group includes a user a, a user B, a user C, and a user D, wherein the interaction frequency of the user a and the user B in the chat group is the highest, and the interaction frequency of the user a and the user C in the chat group is higher than the interaction frequency of the user a and the user D in the chat group. The interaction frequency of the user B and the user C in the chat group is highest, and the interaction frequency of the user B and the user A in the chat group is higher than the interaction frequency of the user B and the user D in the chat group. Then, when the user a is the application user of the instant messaging tool, the display area of the user head portrait of the user B in the chat group is the largest on the user interface of the instant messaging tool of the user a, and the display area of the user head portrait of the user C in the chat group is larger than the display area of the user head portrait of the user D in the chat group. When the user B is an application user of the instant messaging tool, the display area of the user head portrait of the user C in the chat group is the largest on the user interface of the instant messaging tool of the user B, and the display area of the user head portrait of the user a in the chat group is larger than the display area of the user head portrait of the user D in the chat group.
And S1003, displaying the user head of the first user in the chat group according to the size of the display area.
Specifically, according to the size of the display area, displaying the user head portrait of the first user in the chat group. That is, for the application user of the instant messaging tool, on the user interface of the instant messaging tool, the user with the largest display area of the user avatar in the chat group is the user with the highest interaction frequency with the second user in the chat group.
For the same chat group, on the user interface of the instant messaging tool of one user in the chat group, the user with the largest display area of the user head portrait in the chat group is the user with the highest interaction frequency with the user. In the instant messaging tool, the user with the highest interaction frequency with a certain user in the chat group may be different, so that the user with the largest display area of the user head portrait in the chat group may be different on the user interface of the instant messaging tool of all users in the chat group.
S1004, determining the arrangement sequence of the user head portraits of the first user in the chat group according to the interaction frequency of the first user and the second user in the chat group.
In one possible implementation, when the interaction frequency of the first user and the second user in the chat group is highest, determining that the arrangement order of the user head images of the first user in the chat group is the forefront. When the interaction frequency of the first user and the second user in the chat group is not the highest, the front-back relation between the arrangement sequence of the user head images of the first user in the chat group and the arrangement sequence of the user head images of a third user in the chat group is not limited, wherein the third user is other users except the user with the highest interaction frequency with the second user in the chat group.
In another possible implementation, when the interaction frequency of the first user and the second user in the chat group is highest, determining that the arrangement order of the user head images of the first user in the chat group is the forefront. When the interaction frequency of the first user and the second user in the chat group is not the highest, determining that the arrangement sequence of the user head images of the first user in the chat group is earlier than the arrangement sequence of the user head images of a fourth user in the chat group, wherein the fourth user is a user with lower interaction frequency with the second user in the chat group than the first user and the second user in the chat group.
For example, the chat group includes a user a, a user B, a user C, and a user D, wherein the interaction frequency of the user a and the user B in the chat group is the highest, and the interaction frequency of the user a and the user C in the chat group is higher than the interaction frequency of the user a and the user D in the chat group. The interaction frequency of the user B and the user C in the chat group is highest, and the interaction frequency of the user B and the user A in the chat group is higher than the interaction frequency of the user B and the user D in the chat group. Then, for the user a, on the user interface of the instant messaging tool of the user a, the arrangement order of the user head images of the user B in the chat group is the forefront, and the arrangement order of the user head images of the user C in the chat group is more front than the arrangement order of the user head images of the user D in the chat group. For user B, on the user interface of the instant messaging tool of user B, the arrangement order of the user avatars of user C in the chat group is the forefront, and the display area of the user avatars of user a in the chat group is more front than the arrangement order of the user avatars of user D in the chat group.
S1005, displaying the user head images of the first user in the chat group according to the arrangement sequence.
Specifically, according to the arrangement sequence, user head portraits of the first user in the chat group are displayed. That is, for the application user of the instant messaging tool, on the user interface of the instant messaging tool, the user with the head of the user in the chat group in the forefront order is the user with the highest interaction frequency with the second user in the chat group.
In one possible implementation, when there are more users in the chat group, before determining the size and/or the arrangement order of the display area of the user head portraits in the chat group, it is further required to determine the interaction frequency between the first user and the second user in the chat group. When the interaction frequency of the first user and the second user in the chat group is not less than a preset frequency threshold, determining the size and/or arrangement sequence of the display area of the user head portrait of the first user in the chat group, and then displaying the user head portrait of the first user in the chat group. And when the interaction frequency of the first user and the second user in the chat group is smaller than a preset frequency threshold, not displaying the user head portrait of the first user in the chat group.
In the method, the interactive frequency of the user in the chat group and the application user of the instant messaging tool in the chat group is obtained, and the size and the arrangement sequence of the display area of the user head portrait in the chat group are determined according to the interactive frequency, so that the size and the arrangement sequence of the display area of the user head portrait in the chat group are intelligently adjusted. When a plurality of chat groups are arranged in the instant messaging tool, the sizes and the arrangement sequences of the display areas of the user head portraits determined according to the interaction frequency in different chat groups are different, the head portraits of the chat groups have large difference, and a user can quickly identify the chat groups to be searched by using the head portraits of the chat groups. Therefore, the recognition efficiency between chat groups is improved, and the user experience is improved.
Referring to fig. 11a and 11b, fig. 11a is a schematic diagram of a user head portrait of another user in a chat group according to an embodiment of the present application, and fig. 11b is a schematic diagram of a user head portrait of another user in a chat group according to an embodiment of the present application. As shown in fig. 11a and 11B, the head portraits of the chat group are composed of user head portraits of user a, user B, user C and user D, wherein the interaction frequency of user a and user B in the chat group is highest, and the interaction frequency of user a and user C in the chat group is higher than the interaction frequency of user a and user D in the chat group. The interaction frequency of the user B and the user C in the chat group is highest, and the interaction frequency of the user B and the user A in the chat group is higher than the interaction frequency of the user B and the user D in the chat group. In fig. 11a, in the same chat group, for user a, user a is an application user of the instant messaging tool, and on the user interface of the instant messaging tool of user a, the display area of the user avatar of user B in the chat group is the largest, and the display areas of the user avatars of other users in the chat group are the same. In fig. 11B, in the same chat group, the user B is the application user of the instant messaging tool, and on the user interface of the instant messaging tool of the user B, the display area of the user avatar of the user C in the chat group is the largest, and the display areas of the user avatars of other users in the chat group are the same. That is, the user a and the user B see different users with the largest display area on the user interfaces of the respective instant messaging tools.
The embodiment of the application also provides a computer readable storage medium. All or part of the flow of the above method embodiments may be implemented by a computer program to instruct related hardware, where the program may be stored in the above computer storage medium, and when the program is executed, the program may include the flow of each method embodiment as described above. The computer-readable storage medium includes: a read-only memory (ROM) or a random access memory (random access memory, RAM), a magnetic disk or an optical disk, or the like.
In the above embodiments, it may be implemented in whole or in part by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When loaded and executed on a computer, produces a flow or function in accordance with embodiments of the present application, in whole or in part. The computer may be a general purpose computer, a special purpose computer, a computer network, or other programmable apparatus. The computer instructions may be stored in or transmitted across a computer-readable storage medium. The computer readable storage medium may be any available medium that can be accessed by a computer or a data storage device such as a server, data center, etc. that contains an integration of one or more available media. The usable medium may be a magnetic medium (e.g., a floppy disk, a hard disk, a magnetic tape), an optical medium (e.g., a DVD), or a semiconductor medium (e.g., a Solid State Disk (SSD)), or the like.
The steps in the method of the embodiment of the application can be sequentially adjusted, combined and deleted according to actual needs.
The modules in the device of the embodiment of the application can be combined, divided and deleted according to actual needs.
The above embodiments are only for illustrating the technical solution of the present application, and not for limiting the same; although the application has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical scheme described in the foregoing embodiments can be modified or some technical features thereof can be replaced by equivalents; such modifications and substitutions do not depart from the spirit of the application.

Claims (13)

1. A head portrait display method, comprising:
acquiring the activity of a first user in a chat group in an instant messaging tool in the chat group;
determining the display area size of the user head portrait of the first user in the chat group according to the liveness; the size of the display area of the user head portrait of the first user in the chat group is the size of the display area of the user head portrait of the first user in the head portrait of the chat group;
Displaying user head images of the first user in the chat group according to the size of the display area; displaying the user head portrait of the first user in the chat group, wherein the user head portrait of the first user is displayed in the head portrait of the chat group;
wherein the activity of the first user in the chat group comprises the interaction frequency of the first user and a second user in the chat group, wherein the second user is an application user of the instant messaging tool;
the determining, according to the liveness, a display area size of the user head portrait of the first user in the head portrait of the chat group includes: determining that the display area of the head portrait of the first user in the head portrait of the chat group is the largest under the condition that the activity of the first user in the chat group is the highest;
the interaction frequency comprises: at least one of a frequency with which the second user replies to the first user's message within the chat group, a frequency with which the second user forwards the first user's message sent within the chat group to other chat groups, a frequency with which the second user designates the first user to chat within the chat group, and a frequency with which the second user references the first user's message sent within the chat group.
2. The method of claim 1, wherein determining a display area size of a user avatar of the first user in the chat group based on the liveness further comprises:
when the activity of the first user in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is equal to the display area of the user head portrait of a third user in the chat group, wherein the third user is other users except the user with the highest activity in the chat group.
3. The method of claim 2, wherein determining a display area size of a user avatar of the first user in the chat group based on the liveness further comprises:
when the activity of the first user in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is larger than the display area of the user head portrait of a fourth user in the chat group, wherein the fourth user is a user with activity lower than that of the first user in the chat group.
4. A method according to any one of claims 1-3, wherein the method further comprises:
determining the arrangement sequence of user head portraits of the first user in the chat group according to the liveness;
displaying the user head images of the first user in the chat group according to the arrangement sequence.
5. The method of claim 4, wherein determining the order of user avatars of the first user in the chat group based on the liveness comprises:
and when the activity of the first user in the chat group is highest, determining that the arrangement sequence of the user head portraits of the first user in the chat group is the forefront.
6. The method of claim 1, wherein prior to said determining a display area size of a user avatar of the first user in the chat group based on the liveness, the method further comprises:
and confirming that the activity degree of the first user in the chat group is not smaller than a preset activity degree threshold value.
7. An electronic device, comprising: one or more processors, memory;
the memory is coupled with the one or more processors, the memory is for storing computer program code, the computer program code comprising computer instructions, the one or more processors executing the computer instructions to perform:
Acquiring the activity of a first user in a chat group in an instant messaging tool in the chat group;
determining the display area size of the user head portrait of the first user in the chat group according to the liveness; the size of the display area of the user head portrait of the first user in the chat group is the size of the display area of the user head portrait of the first user in the head portrait of the chat group;
displaying user head images of the first user in the chat group according to the size of the display area; displaying the user head portrait of the first user in the chat group, wherein the user head portrait of the first user is displayed in the head portrait of the chat group;
wherein the activity of the first user in the chat group comprises the interaction frequency of the first user and a second user in the chat group, wherein the second user is an application user of the instant messaging tool;
the determining, according to the liveness, a display area size of the user head portrait of the first user in the head portrait of the chat group includes: determining that the display area of the head portrait of the first user in the head portrait of the chat group is the largest under the condition that the activity of the first user in the chat group is the highest;
The interaction frequency comprises: at least one of a frequency with which the second user replies to the first user's message within the chat group, a frequency with which the second user forwards the first user's message sent within the chat group to other chat groups, a frequency with which the second user designates the first user to chat within the chat group, and a frequency with which the second user references the first user's message sent within the chat group.
8. The electronic device of claim 7, wherein the processor is configured to determine a size of a display area of a user avatar of the first user in the chat group based on the activity, the processor is further configured to:
when the activity of the first user in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is equal to the display area of the user head portrait of a third user in the chat group, wherein the third user is other users except the user with the highest activity in the chat group.
9. The electronic device of claim 8, wherein the processor is configured to determine a size of a display area of a user avatar of the first user in the chat group based on the activity, the processor is further configured to:
When the activity of the first user in the chat group is not the highest, determining that the display area of the user head portrait of the first user in the chat group is larger than the display area of the user head portrait of a fourth user in the chat group, wherein the fourth user is a user with activity lower than that of the first user in the chat group.
10. The electronic device of any one of claims 7-9, wherein the processor is further configured to:
determining the arrangement sequence of user head portraits of the first user in the chat group according to the liveness;
displaying the user head images of the first user in the chat group according to the arrangement sequence.
11. The electronic device of claim 10, wherein the processor is configured to, when determining the order of the user avatars of the first user in the chat group according to the liveness, the processor is specifically configured to:
and when the activity of the first user in the chat group is highest, determining that the arrangement sequence of the user head portraits of the first user in the chat group is the forefront.
12. The electronic device of claim 7, wherein the processor is configured to determine, based on the liveness, that the first user is prior to a display area size of a user avatar in the chat group, the processor is further configured to:
and confirming that the activity degree of the first user in the chat group is not smaller than a preset activity degree threshold value.
13. A computer readable storage medium, characterized in that the computer readable storage medium stores a computer program comprising program instructions which, when run on an electronic device, cause the electronic device to perform the method of any one of claims 1 to 6.
CN202010899499.XA 2020-08-31 2020-08-31 Head portrait display method and electronic equipment Active CN114205318B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202010899499.XA CN114205318B (en) 2020-08-31 2020-08-31 Head portrait display method and electronic equipment
PCT/CN2021/125916 WO2022042774A1 (en) 2020-08-31 2021-10-22 Profile picture display method and electronic device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010899499.XA CN114205318B (en) 2020-08-31 2020-08-31 Head portrait display method and electronic equipment

Publications (2)

Publication Number Publication Date
CN114205318A CN114205318A (en) 2022-03-18
CN114205318B true CN114205318B (en) 2023-12-08

Family

ID=80352711

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010899499.XA Active CN114205318B (en) 2020-08-31 2020-08-31 Head portrait display method and electronic equipment

Country Status (2)

Country Link
CN (1) CN114205318B (en)
WO (1) WO2022042774A1 (en)

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2001033327A1 (en) * 1999-10-29 2001-05-10 Electronic Arts Inc. Chat clusters for a virtual world application
CN105430473A (en) * 2015-11-30 2016-03-23 天脉聚源(北京)科技有限公司 Method and device for displaying head portrait of supporter
CN105787982A (en) * 2016-02-29 2016-07-20 腾讯科技(北京)有限公司 Method and device for manufacturing e-book
CN106209574A (en) * 2016-06-17 2016-12-07 广州爱九游信息技术有限公司 Group's head portrait display system based on instant communication software, method and electronic equipment
CN106534485A (en) * 2016-10-12 2017-03-22 乐视控股(北京)有限公司 Method and device for setting group head portrait
CN108196751A (en) * 2018-01-08 2018-06-22 深圳天珑无线科技有限公司 Update method, terminal and the computer readable storage medium of group chat head portrait
CN109656656A (en) * 2018-12-10 2019-04-19 上海掌门科技有限公司 Method and apparatus for generating group chat head portrait
CN110634168A (en) * 2018-06-21 2019-12-31 钉钉控股(开曼)有限公司 Method and device for generating group head portrait
CN110691027A (en) * 2019-08-29 2020-01-14 维沃移动通信有限公司 Information processing method and device, electronic equipment and medium

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7433895B2 (en) * 2005-06-24 2008-10-07 Microsoft Corporation Adding dominant media elements to search results
US8832552B2 (en) * 2008-04-03 2014-09-09 Nokia Corporation Automated selection of avatar characteristics for groups
US20110271209A1 (en) * 2010-04-30 2011-11-03 American Teleconferncing Services Ltd. Systems, Methods, and Computer Programs for Providing a Conference User Interface
US9619100B2 (en) * 2010-08-30 2017-04-11 Nokia Technologies Oy Method, apparatus, and computer program product for adapting a content segment based on an importance level
CN105991405B (en) * 2015-02-11 2019-02-26 腾讯科技(深圳)有限公司 A kind of method for building up and device of instant messaging
CN105681057B (en) * 2016-02-22 2019-01-15 北京橙鑫数据科技有限公司 Group's head portrait update method and device
CN108173742B (en) * 2017-12-08 2021-08-24 腾讯科技(深圳)有限公司 Image data processing method and device

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2001033327A1 (en) * 1999-10-29 2001-05-10 Electronic Arts Inc. Chat clusters for a virtual world application
CN105430473A (en) * 2015-11-30 2016-03-23 天脉聚源(北京)科技有限公司 Method and device for displaying head portrait of supporter
CN105787982A (en) * 2016-02-29 2016-07-20 腾讯科技(北京)有限公司 Method and device for manufacturing e-book
CN106209574A (en) * 2016-06-17 2016-12-07 广州爱九游信息技术有限公司 Group's head portrait display system based on instant communication software, method and electronic equipment
CN106534485A (en) * 2016-10-12 2017-03-22 乐视控股(北京)有限公司 Method and device for setting group head portrait
CN108196751A (en) * 2018-01-08 2018-06-22 深圳天珑无线科技有限公司 Update method, terminal and the computer readable storage medium of group chat head portrait
CN110634168A (en) * 2018-06-21 2019-12-31 钉钉控股(开曼)有限公司 Method and device for generating group head portrait
CN109656656A (en) * 2018-12-10 2019-04-19 上海掌门科技有限公司 Method and apparatus for generating group chat head portrait
WO2020119540A1 (en) * 2018-12-10 2020-06-18 上海掌门科技有限公司 Group profile picture generation method and device
CN110691027A (en) * 2019-08-29 2020-01-14 维沃移动通信有限公司 Information processing method and device, electronic equipment and medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
电视画面制作手段在大型直播节目中的运用――虚拟技术在《直通春晚》中的剖析;魏党生;;影视制作(第03期);全文 *

Also Published As

Publication number Publication date
WO2022042774A1 (en) 2022-03-03
CN114205318A (en) 2022-03-18

Similar Documents

Publication Publication Date Title
CN114467297B (en) Video call display method and related device applied to electronic equipment
CN113542485B (en) Notification processing method, electronic equipment and computer readable storage medium
CN115866121B (en) Application interface interaction method, electronic device and computer readable storage medium
CN111095723B (en) Wireless charging method and electronic equipment
CN114650363B (en) Image display method and electronic equipment
CN113254409B (en) File sharing method, system and related equipment
CN114079893B (en) Bluetooth communication method, terminal device and computer readable storage medium
CN113448382B (en) Multi-screen display electronic device and multi-screen display method of electronic device
CN111602108B (en) Application icon display method and terminal
CN113641271B (en) Application window management method, terminal device and computer readable storage medium
CN114089932B (en) Multi-screen display method, device, terminal equipment and storage medium
CN116048358B (en) Method and related device for controlling suspension ball
CN115426521A (en) Method, electronic device, medium, and program product for screen capture
CN115914461B (en) Position relation identification method and electronic equipment
CN116389884B (en) Thumbnail display method and terminal equipment
CN114911400A (en) Method for sharing pictures and electronic equipment
CN116055859B (en) Image processing method and electronic device
CN113645595B (en) Equipment interaction method and device
CN114205318B (en) Head portrait display method and electronic equipment
CN116048831B (en) Target signal processing method and electronic equipment
CN116233599B (en) Video mode recommendation method and electronic equipment
CN116051351B (en) Special effect processing method and electronic equipment
CN116095223B (en) Notification display method and terminal device
CN116048236B (en) Communication method and related device
CN117666810A (en) Input method and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant