CN112604274B - Virtual object display method, device, terminal and storage medium - Google Patents

Virtual object display method, device, terminal and storage medium Download PDF

Info

Publication number
CN112604274B
CN112604274B CN202011497277.1A CN202011497277A CN112604274B CN 112604274 B CN112604274 B CN 112604274B CN 202011497277 A CN202011497277 A CN 202011497277A CN 112604274 B CN112604274 B CN 112604274B
Authority
CN
China
Prior art keywords
virtual object
virtual
life value
player
current
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011497277.1A
Other languages
Chinese (zh)
Other versions
CN112604274A (en
Inventor
邓颖
潘佳绮
毛克
文晓晴
余伟祥
寇敬
杨泽锋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN202011497277.1A priority Critical patent/CN112604274B/en
Publication of CN112604274A publication Critical patent/CN112604274A/en
Application granted granted Critical
Publication of CN112604274B publication Critical patent/CN112604274B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/50Controlling the output signals based on the game progress
    • A63F13/52Controlling the output signals based on the game progress involving aspects of the displayed game scene
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/55Controlling game characters or game objects based on the game progress
    • A63F13/58Controlling game characters or game objects based on the game progress by computing conditions of game characters, e.g. stamina, strength, motivation or energy level
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/80Special adaptations for executing a specific game genre or game mode
    • A63F13/822Strategy games; Role-playing games
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/80Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game specially adapted for executing a specific type of game
    • A63F2300/807Role playing or strategy games

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Processing Or Creating Images (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application provides a virtual object display method, a virtual object display device, a virtual object display terminal and a virtual object storage medium, and belongs to the technical field of multimedia. The method comprises the following steps: displaying a first virtual scene image, wherein the first virtual scene image comprises a controlled virtual object in a virtual scene and at least one team member virtual object; in response to the trigger operation of detecting the recovery class skill of the controlled virtual object, displaying a second virtual scene image, wherein at least one team member virtual object in the second virtual scene image is respectively rendered into a color corresponding to the current virtual life value; and in response to the selection operation of any member virtual object, controlling the controlled virtual object to release the recovery class skills to the member virtual object. By the scheme, a user can visually determine the current virtual life value of each team member virtual object according to the color of each team member virtual object, so that the release target of the recovery skills can be determined without changing the sight back and forth, the operation is simple, and the human-computer interaction efficiency is improved.

Description

Virtual object display method, device, terminal and storage medium
Technical Field
The present application relates to the field of multimedia technologies, and in particular, to a method, an apparatus, a terminal, and a storage medium for displaying a virtual object.
Background
With the development of multimedia technology, more and more kinds of games can be experienced by users, such as MOBA (Multiplayer Online Battle Arena) games, FPS (First-Person shooter Game), RPG (Role-playing) games, and the like. In the game process, the situation that the virtual life value of the team member virtual object is attacked and lost can occur, and the user can control the controlled virtual object to use the treatment skill to recover the virtual life value for the team member virtual object.
At present, when a user can control a controlled virtual object to treat a team member virtual object, a team list displayed on the left side of an interface needs to be opened, the current virtual life value of each team member is checked, then the team member virtual object needing treatment is selected, and finally the controlled virtual object is controlled to use treatment skills to recover the virtual life value of the designated team member virtual object.
The gaze of the user needs to be moved to the team list on the left side from the center of the screen, namely the sight needs to be adjusted and changed back and forth, and the operation is complex, so that the human-computer interaction efficiency is low, and the game experience of the user is influenced.
Disclosure of Invention
The embodiment of the application provides a display method, a display device, a display terminal and a storage medium of virtual objects, which can enable a user to visually determine the current virtual life value of each team member virtual object by the color of each team member virtual object, so that the release target of recovery skills can be determined without changing the sight back and forth, the operation is simple, and the human-computer interaction efficiency is improved. The technical scheme is as follows:
in one aspect, a method for displaying a virtual object is provided, where the method includes:
displaying a first virtual scene image, wherein the first virtual scene image comprises a controlled virtual object and at least one team member virtual object in a virtual scene, the controlled virtual object is a virtual object controlled by a current terminal, and the team member virtual object and the controlled virtual object belong to the same team;
in response to detecting a triggering operation of recovery class skills of the controlled virtual objects, displaying a second virtual scene image, wherein the at least one team member virtual object in the second virtual scene image is respectively rendered into a color corresponding to a current virtual life value;
and in response to the selection operation of any member virtual object, controlling the controlled virtual object to release the recovery-class skills to the member virtual object.
In another aspect, there is provided an apparatus for displaying a virtual object, the apparatus including:
the system comprises a first display module, a second display module and a third display module, wherein the first display module is used for displaying a first virtual scene image, the first virtual scene image comprises a controlled virtual object and at least one team member virtual object in a virtual scene, the controlled virtual object is a virtual object controlled by a current terminal, and the team member virtual object and the controlled virtual object belong to the same team;
a second display module, configured to display a second virtual scene image in response to detecting a trigger operation of a recovery class skill of the controlled virtual object, where the at least one team member virtual object in the second virtual scene image is rendered to a color corresponding to a current virtual life value;
and the control module is used for responding to the selection operation of any player virtual object and controlling the controlled virtual object to release the recovery class skills to the player virtual object.
In an alternative implementation, the second display module includes:
the acquisition unit is used for responding to the trigger operation of detecting the recovery class skill of the controlled virtual object and acquiring the current virtual life value of the at least one team member virtual object;
the rendering unit is used for rendering any player virtual object into a color corresponding to the current virtual life value according to the current virtual life value of the player virtual object;
and the display unit is used for displaying the second virtual scene image.
In an optional implementation manner, the rendering unit is configured to determine a target proportion according to a current virtual life value of the team member virtual object, where the target proportion is a proportion of the current virtual life value to a total life value; determining target color depth according to a proportional interval in which the target proportion is positioned, wherein one proportional interval corresponds to one color depth, and the depth of the target color depth is inversely related to the size of the virtual life value; and rendering the player virtual object to be a color corresponding to the target color depth.
In an optional implementation manner, the rendering unit is configured to determine a target proportion according to a current virtual life value of the team member virtual object, where the target proportion is a proportion of the current virtual life value to a total life value; and rendering the target proportion of the player virtual object into a target color.
In an optional implementation manner, the rendering unit is further configured to determine the target color corresponding to a scale interval in which the target proportion is located, where one scale interval corresponds to one color.
In an optional implementation, the apparatus further includes:
the first determining module is used for determining skills and pressing time length corresponding to any skill icon in response to the fact that pressing operation on the skill icon is detected;
and the second determination module is used for determining the trigger operation of detecting the recovery skill in response to the fact that the skill belongs to the recovery skill and the pressing time length is not less than the target time length.
In an optional implementation manner, the control module is configured to determine that any player virtual object is a target object of the recovery-class skill in response to detecting a click operation on the player virtual object; and controlling the controlled virtual object to release the recovery class skills to the team member virtual object.
In an optional implementation manner, the control module is configured to determine a selected player virtual object according to a dragging operation of the skill icon of the recovery-class skill within a target range; in response to a press ending operation, controlling the controlled virtual object to release the recovery class skills to the player virtual object.
In an optional implementation manner, the second display module is further configured to render the virtual scene and the controlled virtual object in the second virtual scene image to be lighter than the color of the at least one team member virtual object in response to detecting the triggering operation of the recovery class skill of the controlled virtual object.
In another aspect, a terminal is provided, where the terminal includes a processor and a memory, where the memory is used to store at least one piece of computer program, and the at least one piece of computer program is loaded by the processor and executed to implement the operations performed in the display method of a virtual object in the embodiments of the present application.
In another aspect, a computer-readable storage medium is provided, in which at least one piece of computer program is stored, and the at least one piece of computer program is loaded and executed by a processor to implement the operations performed in the display method of a virtual object in the embodiments of the present application.
In another aspect, a computer program product or a computer program is provided, the computer program product or the computer program comprising computer program code, the computer program code being stored in a computer readable storage medium. The processor of the terminal reads the computer program code from the computer-readable storage medium, and the processor executes the computer program code, so that the terminal performs the display method of the virtual object provided in the above-described aspects or various alternative implementations of the aspects.
The technical scheme provided by the embodiment of the application has the following beneficial effects:
in the embodiment of the application, the display method of the virtual object is provided, when the recovery type skill is triggered, the team member virtual object is rendered into the color corresponding to the current virtual life value, so that a user can intuitively determine the current virtual life value of each team member virtual object according to the color of each team member virtual object, the release target of the recovery type skill can be determined without changing the sight back and forth, the operation is simple, and the human-computer interaction efficiency is improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
FIG. 1 is a schematic diagram of an implementation environment of a display method of a virtual object according to an embodiment of the present application;
FIG. 2 is a flowchart of a method for displaying a virtual object according to an embodiment of the present application;
FIG. 3 is a flowchart of another method for displaying a virtual object according to an embodiment of the present disclosure;
FIG. 4 is a schematic diagram of a first virtual scene image provided in accordance with an embodiment of the present application;
FIG. 5 is a schematic diagram of a second virtual scene image provided in accordance with an embodiment of the present application;
FIG. 6 is a schematic diagram of selecting a team member virtual object according to an embodiment of the present disclosure;
FIG. 7 is a flowchart of another method for displaying a virtual object according to an embodiment of the present disclosure;
FIG. 8 is a block diagram of a display device for a virtual object according to an embodiment of the present application;
fig. 9 is a block diagram of a terminal according to an embodiment of the present application.
Detailed Description
To make the objects, technical solutions and advantages of the present application more clear, embodiments of the present application will be described in further detail below with reference to the accompanying drawings.
Reference will now be made in detail to the exemplary embodiments, examples of which are illustrated in the accompanying drawings. When the following description refers to the accompanying drawings, like numbers in different drawings represent the same or similar elements unless otherwise indicated. The embodiments described in the following exemplary embodiments do not represent all embodiments consistent with the present application. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present application, as detailed in the appended claims.
In order to facilitate understanding of the technical processes of the embodiments of the present application, some terms referred to in the embodiments of the present application are explained below:
virtual scene: is a virtual scene that is displayed (or provided) by an application program when the application program runs on a terminal. The virtual scene may be a simulation environment of a real world, a semi-simulation semi-fictional virtual environment, or a pure fictional virtual environment. The virtual scene may be any one of a two-dimensional virtual scene, a 2.5-dimensional virtual scene, or a three-dimensional virtual scene, and the dimension of the virtual scene is not limited in the embodiment of the present application. For example, a virtual scene may include sky, land, ocean, etc., the land may include environmental elements such as deserts, cities, etc., and a user may control a virtual object to move in the virtual scene. Optionally, the virtual scene may also be used for virtual scene engagement between at least two virtual objects, in which virtual resources are available for use by the at least two virtual objects. Optionally, the virtual scene may include two symmetric regions, and the virtual objects belonging to the two enemy camps occupy one of the regions respectively, and destroy a target building/site/base/crystal deep in the other region as a winning target, where the symmetric regions include a lower left corner region and an upper right corner region, and further include a middle left side region and a middle right side region.
Virtual object: refers to a movable object in a virtual scene. The movable object can be a virtual character, a virtual animal, an animation character, etc., such as: characters, animals, plants, oil drums, walls, stones, etc. displayed in the virtual scene. The controlled virtual object may be an avatar in the virtual scene that is virtual to represent the user. A plurality of virtual objects may be included in the virtual scene, each virtual object having its own shape and volume in the virtual scene, occupying a portion of the space in the virtual scene. Optionally, when the virtual scene is a three-dimensional virtual scene, optionally, the virtual object may be a three-dimensional stereo model, the three-dimensional stereo model may be a three-dimensional character constructed based on a three-dimensional human skeleton technology, and the same virtual object may exhibit different external images by wearing different skins. In some embodiments, the virtual object may also be implemented by using a 2.5-dimensional or 2-dimensional model, which is not limited in this application.
Alternatively, the controlled virtual object may be a Player Character controlled by an operation on the client, an Artificial Intelligence (AI) set in the virtual scene battle by training, or a Non-Player Character (NPC) set in the virtual scene interaction. Alternatively, the controlled virtual object may be a virtual character playing a game in a virtual scene. Optionally, the number of virtual objects participating in the interaction in the virtual scene may be preset, or may be dynamically determined according to the number of clients participating in the interaction.
Hereinafter, an implementation environment of the display method of the virtual object provided in the embodiment of the present application is described. Fig. 1 is a schematic diagram of an implementation environment of a display method of a virtual object according to an embodiment of the present application. Referring to fig. 1, the implementation environment includes a terminal 101 and a server 102.
The terminal 101 and the server 102 can be directly or indirectly connected through wired or wireless communication, and the application is not limited herein.
Optionally, the terminal 101 is a smart phone, a tablet computer, a notebook computer, a desktop computer, a smart speaker, a smart watch, or the like, but is not limited thereto. The terminal 101 is installed and operated with an application program supporting a virtual scene. The application program may be any one of a First-Person shooter game (FPS), a third-Person shooter game, a Multiplayer Online Battle Arena game (MOBA), a virtual reality application program, a three-dimensional map program, or a Multiplayer gunfight type survival game. Illustratively, the terminal 101 is a terminal used by a user, and the user uses the terminal 101 to operate a virtual object located in a virtual scene for activities, including but not limited to: adjusting at least one of body posture, crawling, walking, running, riding, jumping, driving, picking, shooting, attacking, throwing. Illustratively, the virtual object is a virtual character, such as a simulated character or an animated character.
Optionally, the server 102 is an independent physical server, may also be a server cluster or a distributed system formed by a plurality of physical servers, and may also be a cloud server providing basic cloud computing services such as a cloud service, a cloud database, cloud computing, a cloud function, cloud storage, a Network service, cloud communication, a middleware service, a domain name service, a security service, a CDN (Content Delivery Network), a big data and artificial intelligence platform, and the like. The server 102 is used for providing background services for the application programs supporting the virtual scenes. Alternatively, the server 102 may undertake primary computational tasks and the terminal 101 may undertake secondary computational tasks; or, the server 102 undertakes the secondary computing work, and the terminal 101 undertakes the primary computing work; alternatively, the server 102 and the terminal 101 perform cooperative computing by using a distributed computing architecture.
Alternatively, the virtual object controlled by the terminal 101 (hereinafter referred to as the controlled virtual object) and the virtual object controlled by the other terminal 101 (hereinafter referred to as the player virtual object) are in the same virtual scene, and at this time, the controlled virtual object can interact with the player virtual object in the virtual scene. In some embodiments, the controlled virtual object and the team member virtual object are in a cooperative relationship, for example, the controlled virtual object and the team member virtual object may belong to the same team and organization, and the virtual objects in the cooperative relationship may interact by releasing the recovery-class skills from each other.
Those skilled in the art will appreciate that the number of terminals described above may be greater or fewer. For example, the number of the terminals may be only one, or several tens or hundreds of the terminals, or more. The number of terminals and the type of the device are not limited in the embodiments of the present application.
Optionally, the wireless network or wired network described above uses standard communication techniques and/or protocols. The Network is typically the Internet, but can be any Network including, but not limited to, a Local Area Network (LAN), a Metropolitan Area Network (MAN), a Wide Area Network (WAN), a mobile, wired or wireless Network, a private Network, or any combination of virtual private networks. In some embodiments, data exchanged over a network is represented using techniques and/or formats including Hypertext Mark-up Language (HTML), Extensible Markup Language (XML), and the like. All or some of the links can also be encrypted using conventional encryption techniques such as Secure Socket Layer (SSL), Transport Layer Security (TLS), Virtual Private Network (VPN), Internet Protocol Security (IPsec). In other embodiments, custom and/or dedicated data communication techniques can also be used in place of or in addition to the data communication techniques described above.
Fig. 2 is a flowchart of a method for displaying a virtual object according to an embodiment of the present application, and as shown in fig. 2, the method is described in the embodiment of the present application by taking an application to a terminal as an example. The display method of the virtual object comprises the following steps:
201. the terminal displays a first virtual scene image, wherein the first virtual scene image comprises a controlled virtual object and at least one team member virtual object in a virtual scene, the controlled virtual object is a virtual object controlled by the current terminal, and the team member virtual object and the controlled virtual object belong to the same team.
In the embodiment of the application, during the game played by the user through the terminal, the terminal displays a first virtual scene image including a virtual scene, a controlled virtual object controlled by the user through the terminal, and a team member virtual object controlled by a team member of the user through a terminal screen. The controlled virtual object and the at least one team member virtual object are in a cooperative relationship. Optionally, the team member virtual object and the controlled virtual object belong to the same team or the same camp, and the like.
202. And in response to the triggering operation of detecting the recovery class skill of the controlled virtual object, the terminal displays a second virtual scene image, and the at least one team member virtual object in the second virtual scene image is respectively rendered into a color corresponding to the current virtual life value.
In the embodiment of the application, the user can release the corresponding skill by pressing the skill icon, and if the skill icon pressed by the user corresponds to the recovery-class skill, the terminal displays a second scene image in which the at least one team member virtual object is not a colored three-dimensional stereo model any more, but is rendered in a single color, for example, the team member virtual object is rendered in red. Optionally, the shade of the rendered color of the player virtual object is inversely related to the current virtual life value of the player virtual object, that is, the higher the current virtual life value is, the lighter the rendered color is; the lower the current virtual life value, the darker the rendered color. Correspondingly, the user can intuitively know the current virtual life value of each team member virtual object according to the shade of the color of the team member virtual object, so that the team member virtual object is recovered.
203. And responding to the selection operation of any member virtual object, and controlling the controlled virtual object to release the recovery class skill to the member virtual object by the terminal.
In the embodiment of the application, after triggering the recovery-class skills, a user can select any player virtual object as a release target of the skills, the terminal determines the player virtual object selected by the user according to the selection operation of the user, and then the controlled virtual object is controlled to release the recovery-class skills to the selected player virtual object so as to recover the virtual life value of the corresponding player virtual object.
In the embodiment of the application, the display method of the virtual object is provided, when the recovery type skill is triggered, the team member virtual object is rendered into the color corresponding to the current virtual life value, so that a user can intuitively determine the current virtual life value of each team member virtual object according to the color of each team member virtual object, the release target of the recovery type skill can be determined without changing the sight back and forth, the operation is simple, and the human-computer interaction efficiency is improved.
Fig. 3 is a flowchart of another method for displaying a virtual object according to an embodiment of the present application, and as shown in fig. 3, the method is described in the embodiment of the present application by taking an application to a terminal as an example. The display method of the virtual object comprises the following steps:
301. the terminal displays a first virtual scene image, wherein the first virtual scene image comprises a controlled virtual object and at least one team member virtual object in a virtual scene, the controlled virtual object is a virtual object controlled by the current terminal, and the team member virtual object and the controlled virtual object belong to the same team.
In the embodiment of the application, a user can run a game program through a terminal, and during the game, the terminal displays a virtual scene image of the game on a terminal screen, which is called a first scene image for the convenience of distinguishing and describing. The first scene image comprises a virtual scene of a game and a virtual object positioned in the virtual scene, wherein the virtual object comprises a controlled virtual object controlled by a current terminal and at least one team member virtual object belonging to the same team or the same camp as the controlled virtual object.
For example, referring to fig. 4, fig. 4 is a schematic diagram of a first virtual scene image provided according to an embodiment of the present application. As shown in fig. 4, the first virtual scene image exemplarily shows the controlled virtual object and three team member virtual objects. The user can release the recovery-type skills by pressing the skill icon.
302. And in response to the trigger operation of detecting the recovery class skill of the controlled virtual object, the terminal acquires the current virtual life value of the at least one team member virtual object.
In the embodiment of the application, the virtual object generally has at least one skill, a skill icon of the at least one skill owned by the controlled virtual object is displayed on a terminal screen, and a user can release the skill by pressing the skill icon of any one skill. The skills are divided into various types, such as attack skills, defense skills, recovery skills, displacement skills and the like.
Optionally, when the terminal detects a pressing operation of any skill icon of the controlled virtual object, the terminal may determine a skill and a pressing duration corresponding to the skill icon, and in response to that the skill belongs to the recovery-type skill and the pressing duration is not less than the target duration, the terminal may determine that the triggering operation of the recovery-type skill is detected. Then, the terminal respectively obtains the current virtual life value of the at least one team member virtual object. Through detecting the skill type and the pressing duration corresponding to the skill icon, the release of recovery skills caused by misoperation of a user can be avoided, the accuracy of the skill operation is improved, and the human-computer interaction efficiency is improved.
For example, a user presses a skill icon of a treatment skill, the terminal starts timing when detecting the pressing operation and determines that the treatment skill belongs to a recovery-type skill, and if the pressing time of the skill icon is not more than two seconds, the terminal determines that the misoperation is detected; and if the pressing time of the skill icon is longer than two seconds, the terminal determines that the trigger operation of the recovery-type skill is detected, and acquires the current virtual life value of at least one team member virtual object.
303. For any member virtual object, the terminal renders the member virtual object into a color corresponding to the current virtual life value according to the current virtual life value of the member virtual object.
In the embodiment of the application, the rendering of the virtual object by the terminal is to render the three-dimensional model of the virtual object, and because the current virtual life values of different team member virtual objects are different, the colors of the rendered team member virtual objects are not completely the same.
Optionally, the terminal can render each member virtual object to be the same color, and the size of the current virtual life value of the member virtual object is inversely related to the color depth of the color. Correspondingly, the terminal determines a target proportion according to the current virtual life value of the team member virtual object, wherein the target proportion is the proportion of the current virtual life value of the team member virtual object to the total life value of the team member virtual object. And the terminal determines the target color depth according to the proportional interval in which the target proportion is positioned, wherein one proportional interval corresponds to one color depth, and the depth of the target color depth is inversely related to the size of the virtual life value. And finally, rendering the virtual object of the team member into a color corresponding to the target color depth by the terminal. The team member virtual objects are displayed in different color depths according to the current virtual life values of the team member virtual objects, so that a user can visually determine the current virtual object values of the team member virtual objects according to the color depths of the team member virtual objects, the operation of ordering a team list by the user is avoided, the follow-up selection of the team member virtual objects needing treatment is facilitated, and the human-computer interaction efficiency is improved.
For example, the current virtual life value of the player virtual object a accounts for 70% of the total life value of the player virtual object a, the current virtual life value of the player virtual object B accounts for 50% of the total life value of the player virtual object B, and the current virtual life value of the player virtual object C accounts for 20% of the total life value of the player virtual object C. The terminal renders the player virtual object A into light red, renders the player virtual object B into red, and renders the player virtual object C into dark red.
Optionally, the terminal may render each member virtual object to be the same color, and it should be noted that the ratio of the current virtual life value of the member virtual object to the total life value is the ratio of the member virtual object rendered in a single color. Correspondingly, the terminal determines a target proportion according to the current virtual life value of the team member virtual object, wherein the target proportion is the proportion of the current virtual life value to the total life value. And rendering the target proportion of the virtual object of the team member as a target color by the terminal.
For example, the current virtual life value of the player virtual object a accounts for 70% of the total life value of the player virtual object a, the current virtual life value of the player virtual object B accounts for 50% of the total life value of the player virtual object B, and the current virtual life value of the player virtual object C accounts for 20% of the total life value of the player virtual object C. Rendering the part of the player virtual object A, which is close to the bottom 70%, into red by the terminal, namely rendering the part which is close to the lower part of the head into red; rendering the part of the player virtual object B close to the bottom 50% to be red by the terminal, namely rendering the part close to the lower part of the waist to be red; the terminal renders the portion of the player virtual object C near the bottom 20% red, that is, the portion near below the knee, red. Note that the remaining part of the player virtual object is in a state where no color is rendered.
Optionally, the terminal can also render each of the player virtual objects in a different color. Correspondingly, the terminal determines a target proportion according to the current virtual life value of the team member virtual object, wherein the target proportion is the proportion of the current virtual life value to the total life value. The terminal then determines a target color corresponding to a scale interval in which the target scale is located, one scale interval corresponding to one color. And finally, rendering the target proportion of the virtual object of the team member into a target color by the terminal.
For example, the current virtual life value of the player virtual object a accounts for 70% of the total life value of the player virtual object a, the current virtual life value of the player virtual object B accounts for 50% of the total life value of the player virtual object B, and the current virtual life value of the player virtual object C accounts for 20% of the total life value of the player virtual object C. The terminal renders the team member virtual object A into green, renders the team member virtual object B into orange, and renders the team member virtual object C into red.
304. And the terminal displays a second virtual scene image, and the at least one team member virtual object in the second virtual scene image is respectively rendered into a color corresponding to the current virtual life value.
In the embodiment of the application, after the terminal renders at least one member virtual object, the rendered member virtual object is displayed in the second virtual scene image. Optionally, in the second virtual scene image, the virtual scene and the controlled virtual object except for the at least one player virtual object are rendered to be lighter than the at least one player virtual object, so that the player virtual object can be more prominent.
For example, referring to fig. 5, fig. 5 is a schematic diagram of a second virtual scene image provided according to an embodiment of the present application. As shown in fig. 5, the current virtual life value of the player virtual object a accounts for 70% of the total life value of the player virtual object a, the current virtual life value of the player virtual object B accounts for 50% of the total life value of the player virtual object B, and the current virtual life value of the player virtual object C accounts for 20% of the total life value of the player virtual object C. Team virtual object A is rendered in light red, team virtual object B is rendered in red, and team virtual object C is rendered in dark red.
305. And responding to the selection operation of any player virtual object, and controlling the controlled virtual object to release the recovery class skill to the player virtual object by the terminal.
In the embodiment of the application, the user can select the player virtual object in various ways, such as directly clicking the player virtual object, or dragging the skill icon to select the player virtual object, and the like.
Optionally, in response to detecting the click operation on any player virtual object, the terminal determines that the player virtual object is a target object of the recovery-type skills, and then controls the controlled virtual object to release the recovery-type skills to the player virtual object. It should be noted that, the user can select the team member virtual object in a direct click mode, and the finger of the user triggering the recovery-type skill may be removed or not removed, which is not limited in the embodiment of the present application.
For example, referring to fig. 6, fig. 6 is a schematic diagram illustrating a selection of a team member virtual object according to an embodiment of the present disclosure. As shown in fig. 6, the user can point with the other hand to select by any player virtual object.
Optionally, the terminal determines the selected player virtual object according to the dragging operation of the skill icon of the recovery class skill in the target range, and then, in response to the pressing ending operation, the terminal controls the controlled virtual object to release the recovery class skill to the player virtual object. The target range is a sector range centered on the recovery-type skill. Of course, the user can also drag the skill icon directly onto any team member virtual object to select the team member virtual object, which is not limited in the embodiment of the present application.
It should be noted that, the foregoing steps 301 to 305 are optional implementations of the display method for virtual objects provided in the present application, and accordingly, there are other optional implementations, as shown in fig. 7, that include the following steps: after the game starts, 701, detecting the skill of pressing treatment again; 702 determining whether the compression treatment skill exceeds 2 seconds; 703. not more than 2 seconds, treatment skills are not triggered; 704. displaying a player virtual life value sensing state diagram when the time exceeds 2 seconds, wherein in the virtual life value sensing state diagram, player virtual objects are rendered into a single color; 705. determining a team member virtual object selected by a user according to the virtual life value state; 706. and controlling the therapeutic skill of the therapy application of the controlled virtual object.
In the embodiment of the application, the display method of the virtual object is provided, and when the recovery type skill is triggered, the virtual object of the team members is rendered into the color corresponding to the current virtual life value, so that a user can intuitively determine the current virtual life value of each virtual object of the team members according to the color of each virtual object of the team members, the release target of the recovery type skill can be determined without changing the sight back and forth, the operation is simple, and the human-computer interaction efficiency is improved.
Fig. 8 is a block diagram of a display device for a virtual object according to an embodiment of the present application. The apparatus is used for executing the steps when the display method of the virtual object is executed, referring to fig. 8, and the apparatus includes: a first display module 801, a second display module 802, and a control module 803.
A first display module 801, configured to display a first virtual scene image, where the first virtual scene image includes a controlled virtual object in a virtual scene and at least one team member virtual object, the controlled virtual object is a virtual object controlled by a current terminal, and the team member virtual object and the controlled virtual object belong to a same team;
a second display module 802, configured to, in response to detecting a trigger operation of a recovery-class skill of the controlled virtual object, display a second virtual scene image, where the at least one team member virtual object in the second virtual scene image is rendered to a color corresponding to the current virtual life value;
and the control module 803 is used for responding to the selection operation of any player virtual object, and controlling the controlled virtual object to release the recovery class skill to the player virtual object.
In an alternative implementation, the second display module 802 includes:
the acquisition unit is used for responding to the trigger operation of detecting the recovery class skill of the controlled virtual object and acquiring the current virtual life value of the at least one team member virtual object;
the rendering unit is used for rendering any player virtual object into a color corresponding to the current virtual life value of the player virtual object according to the current virtual life value of the player virtual object;
and the display unit is used for displaying the second virtual scene image.
In an optional implementation manner, the rendering unit is configured to determine a target proportion according to a current virtual life value of the team member virtual object, where the target proportion is a proportion of the current virtual life value to a total life value; determining a target color depth according to a proportional interval in which the target proportion is positioned, wherein one proportional interval corresponds to one color depth, and the depth of the target color depth is inversely related to the size of the virtual life value; and rendering the virtual object of the team member into a color corresponding to the target color depth.
In an optional implementation manner, the rendering unit is configured to determine a target proportion according to a current virtual life value of the team member virtual object, where the target proportion is a proportion of the current virtual life value to a total life value; and rendering the target proportion of the player virtual object as a target color.
In an optional implementation manner, the rendering unit is further configured to determine the target color corresponding to a scale interval in which the target scale is located, where one scale interval corresponds to one color.
In an optional implementation, the apparatus further includes:
the first determining module is used for determining skills and pressing time length corresponding to any skill icon in response to the fact that pressing operation on the skill icon is detected;
and the second determination module is used for determining the trigger operation of detecting the recovery skill in response to the fact that the skill belongs to the recovery skill and the pressing time length is not less than the target time length.
In an alternative implementation, the control module 803 is configured to determine that any player virtual object is a target object of the recovery-class skill in response to detecting a click operation on the player virtual object; and controlling the controlled virtual object to release the recovery class skills to the team member virtual object.
In an alternative implementation manner, the control module 803 is configured to determine the selected player virtual object according to the dragging operation of the skill icon of the recovery-class skill within the target range; and controlling the controlled virtual object to release the recovery-class skills to the team member virtual object in response to the pressing of the ending operation.
In an alternative implementation, the second display module 802 is further configured to render the virtual scene and the controlled virtual object in the second virtual scene image to be lighter than the color of the at least one team member virtual object in response to detecting the triggering operation of the recovery class skill of the controlled virtual object.
In the embodiment of the application, the display method of the virtual object is provided, when the recovery type skill is triggered, the team member virtual object is rendered into the color corresponding to the current virtual life value, so that a user can intuitively determine the current virtual life value of each team member virtual object according to the color of each team member virtual object, the release target of the recovery type skill can be determined without changing the sight back and forth, the operation is simple, and the human-computer interaction efficiency is improved.
It should be noted that: in the display apparatus for virtual objects provided in the foregoing embodiments, when running an application program, the division of each function module is only used for illustration, and in practical applications, the function distribution may be completed by different function modules according to needs, that is, the internal structure of the apparatus is divided into different function modules, so as to complete all or part of the functions described above. In addition, the display apparatus of the virtual object provided in the above embodiments and the display method embodiment of the virtual object belong to the same concept, and specific implementation processes thereof are described in the method embodiments and are not described herein again.
Fig. 9 is a block diagram of a terminal 900 according to an embodiment of the present application. The terminal 900 may be a portable mobile terminal such as: a smart phone, a tablet computer, an MP3 player (Moving Picture Experts Group Audio Layer III, motion video Experts compression standard Audio Layer 3), an MP4 player (Moving Picture Experts Group Audio Layer IV, motion video Experts compression standard Audio Layer 4), a notebook computer, or a desktop computer. Terminal 900 may also be referred to by other names such as user equipment, portable terminals, laptop terminals, desktop terminals, and the like.
In general, terminal 900 includes: a processor 901 and a memory 902.
Processor 901 may include one or more processing cores, such as a 4-core processor, an 8-core processor, and so forth. The processor 901 may be implemented in at least one hardware form of a DSP (Digital Signal Processing), an FPGA (Field-Programmable Gate Array), and a PLA (Programmable Logic Array). The processor 901 may also include a main processor and a coprocessor, where the main processor is a processor for Processing data in an awake state, and is also called a Central Processing Unit (CPU); a coprocessor is a low power processor for processing data in a standby state. In some embodiments, the processor 901 may be integrated with a GPU (Graphics Processing Unit), which is responsible for rendering and drawing the content required to be displayed by the display screen. In some embodiments, the processor 901 may further include an AI (Artificial Intelligence) processor for processing computing operations related to machine learning.
Memory 902 may include one or more computer-readable storage media, which may be non-transitory. The memory 902 may also include high-speed random access memory, as well as non-volatile memory, such as one or more magnetic disk storage devices, flash memory storage devices. In some embodiments, a non-transitory computer readable storage medium in the memory 902 is used to store at least one computer program for execution by the processor 901 to implement the display method of virtual objects provided by the method embodiments in the present application.
In some embodiments, terminal 900 can also optionally include: a peripheral interface 903 and at least one peripheral. The processor 901, memory 902, and peripheral interface 903 may be connected by buses or signal lines. Various peripheral devices may be connected to the peripheral interface 903 via a bus, signal line, or circuit board. Specifically, the peripheral device includes: at least one of a radio frequency circuit 904, a display screen 905, a camera assembly 906, an audio circuit 907, and a power supply 909.
The peripheral interface 903 may be used to connect at least one peripheral related to I/O (Input/Output) to the processor 901 and the memory 902. In some embodiments, the processor 901, memory 902, and peripheral interface 903 are integrated on the same chip or circuit board; in some other embodiments, any one or two of the processor 901, the memory 902 and the peripheral interface 903 may be implemented on a separate chip or circuit board, which is not limited by this embodiment.
The Radio Frequency circuit 904 is used for receiving and transmitting RF (Radio Frequency) signals, also called electromagnetic signals. The radio frequency circuitry 904 communicates with communication networks and other communication devices via electromagnetic signals. The radio frequency circuit 904 converts an electrical signal into an electromagnetic signal for transmission, or converts a received electromagnetic signal into an electrical signal. Optionally, the radio frequency circuit 904 comprises: an antenna system, an RF transceiver, one or more amplifiers, a tuner, an oscillator, a digital signal processor, a codec chipset, a subscriber identity module card, and so forth. The radio frequency circuit 904 may communicate with other terminals via at least one wireless communication protocol. The wireless communication protocols include, but are not limited to: the world wide web, metropolitan area networks, intranets, generations of mobile communication networks (2G, 3G, 4G, and 5G), Wireless local area networks, and/or WiFi (Wireless Fidelity) networks. In some embodiments, the radio frequency circuit 904 may also include NFC (Near Field Communication) related circuits, which are not limited in this application.
The display screen 905 is used to display a UI (User Interface). The UI may include graphics, text, icons, video, and any combination thereof. When the display screen 905 is a touch display screen, the display screen 905 also has the ability to capture touch signals on or over the surface of the display screen 905. The touch signal may be input to the processor 901 as a control signal for processing. At this point, the display 905 may also be used to provide virtual buttons and/or a virtual keyboard, also referred to as soft buttons and/or a soft keyboard. In some embodiments, the display 905 may be one, disposed on the front panel of the terminal 900; in other embodiments, the number of the display panels 905 may be at least two, and each of the display panels is disposed on a different surface of the terminal 900 or is in a foldable design; in other embodiments, the display 905 may be a flexible display disposed on a curved surface or a folded surface of the terminal 900. Even more, the display screen 905 may be arranged in a non-rectangular irregular figure, i.e. a shaped screen. The Display panel 905 can be made of LCD (Liquid Crystal Display), OLED (Organic Light-Emitting Diode), and other materials.
The camera assembly 906 is used to capture images or video. Optionally, camera assembly 906 includes a front camera and a rear camera. Generally, a front camera is disposed at a front panel of the terminal, and a rear camera is disposed at a rear surface of the terminal. In some embodiments, the number of the rear cameras is at least two, and each rear camera is any one of a main camera, a depth-of-field camera, a wide-angle camera and a telephoto camera, so that the main camera and the depth-of-field camera are fused to realize a background blurring function, and the main camera and the wide-angle camera are fused to realize panoramic shooting and VR (Virtual Reality) shooting functions or other fusion shooting functions. In some embodiments, camera assembly 906 may also include a flash. The flash lamp can be a monochrome temperature flash lamp or a bicolor temperature flash lamp. The double-color-temperature flash lamp is a combination of a warm-light flash lamp and a cold-light flash lamp, and can be used for light compensation at different color temperatures.
Audio circuit 907 may include a microphone and a speaker. The microphone is used for collecting sound waves of a user and the environment, converting the sound waves into electric signals, and inputting the electric signals to the processor 901 for processing, or inputting the electric signals to the radio frequency circuit 904 for realizing voice communication. For stereo sound acquisition or noise reduction purposes, the microphones may be multiple and disposed at different locations of the terminal 900. The microphone may also be an array microphone or an omni-directional pick-up microphone. The speaker is used to convert electrical signals from the processor 901 or the radio frequency circuit 904 into sound waves. The loudspeaker can be a traditional film loudspeaker or a piezoelectric ceramic loudspeaker. When the speaker is a piezoelectric ceramic speaker, the speaker can be used for purposes such as converting an electric signal into a sound wave audible to a human being, or converting an electric signal into a sound wave inaudible to a human being to measure a distance. In some embodiments, audio circuit 907 may also include a headphone jack.
Power supply 909 is used to provide power to the various components in terminal 900. The power source 909 may be alternating current, direct current, disposable or rechargeable. When the power source 909 includes a rechargeable battery, the rechargeable battery may be a wired rechargeable battery or a wireless rechargeable battery. The wired rechargeable battery is a battery charged through a wired line, and the wireless rechargeable battery is a battery charged through a wireless coil. The rechargeable battery may also be used to support fast charge technology.
In some embodiments, terminal 900 can also include one or more sensors 910. The one or more sensors 910 include, but are not limited to: acceleration sensor 911, gyro sensor 912, pressure sensor 913, optical sensor 915, and proximity sensor 916.
The acceleration sensor 911 can detect the magnitude of acceleration in three coordinate axes of the coordinate system established with the terminal 900. For example, the acceleration sensor 911 may be used to detect the components of the gravitational acceleration in three coordinate axes. The processor 901 can control the display screen 905 to display the user interface in a landscape view or a portrait view according to the gravitational acceleration signal collected by the acceleration sensor 911. The acceleration sensor 911 may also be used for acquisition of motion data of a game or a user.
The gyro sensor 912 may detect a body direction and a rotation angle of the terminal 900, and the gyro sensor 912 may cooperate with the acceleration sensor 911 to acquire a 3D motion of the user on the terminal 900. The processor 901 can implement the following functions according to the data collected by the gyro sensor 912: motion sensing (such as changing the UI according to a user's tilting operation), image stabilization at the time of photographing, game control, and inertial navigation.
The pressure sensor 913 may be disposed on a side bezel of the terminal 900 and/or underneath the display 905. When the pressure sensor 913 is disposed on the side frame of the terminal 900, the user's holding signal of the terminal 900 may be detected, and the processor 901 performs left-right hand recognition or shortcut operation according to the holding signal collected by the pressure sensor 913. When the pressure sensor 913 is disposed at a lower layer of the display screen 905, the processor 901 controls the operability control on the UI interface according to the pressure operation of the user on the display screen 905. The operability control comprises at least one of a button control, a scroll bar control, an icon control and a menu control.
The optical sensor 915 is used to collect ambient light intensity. In one embodiment, the processor 901 may control the display brightness of the display screen 905 based on the ambient light intensity collected by the optical sensor 915. Specifically, when the ambient light intensity is high, the display brightness of the display screen 905 is increased; when the ambient light intensity is low, the display brightness of the display screen 905 is reduced. In another embodiment, the processor 901 can also dynamically adjust the shooting parameters of the camera assembly 906 according to the ambient light intensity collected by the optical sensor 915.
Proximity sensor 916, also known as a distance sensor, is typically disposed on the front panel of terminal 900. The proximity sensor 916 is used to collect the distance between the user and the front face of the terminal 900. In one embodiment, when the proximity sensor 916 detects that the distance between the user and the front face of the terminal 900 gradually decreases, the processor 901 controls the display 905 to switch from the bright screen state to the dark screen state; when the proximity sensor 916 detects that the distance between the user and the front surface of the terminal 900 gradually becomes larger, the display 905 is controlled by the processor 901 to switch from the breath screen state to the bright screen state.
Those skilled in the art will appreciate that the configuration shown in fig. 9 does not constitute a limitation of terminal 900, and may include more or fewer components than those shown, or may combine certain components, or may employ a different arrangement of components.
The embodiment of the present application further provides a computer-readable storage medium, which is applied to a terminal, and at least one piece of computer program is stored in the computer-readable storage medium, and is loaded and executed by a processor to implement the operations executed by the terminal in the display method of a virtual object of the foregoing embodiment.
Embodiments of the present application also provide a computer program product or a computer program comprising computer program code stored in a computer readable storage medium. The processor of the terminal reads the computer program code from the computer-readable storage medium, and the processor executes the computer program code, so that the terminal performs the display method of the virtual object provided in the above-described various alternative implementations.
It will be understood by those skilled in the art that all or part of the steps for implementing the above embodiments may be implemented by hardware, or may be implemented by a program instructing relevant hardware, where the program may be stored in a computer-readable storage medium, and the above-mentioned storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
The above description is only exemplary of the present application and should not be taken as limiting, as any modification, equivalent replacement, or improvement made within the spirit and principle of the present application should be included in the protection scope of the present application.

Claims (15)

1. A method for displaying a virtual object, the method comprising:
displaying a first virtual scene image, wherein the first virtual scene image comprises a controlled virtual object and at least one team member virtual object in a virtual scene, the controlled virtual object is a virtual object controlled by a current terminal, and the team member virtual object and the controlled virtual object belong to the same team;
in response to detecting a triggering operation of a recovery class skill of the controlled virtual object, displaying a second virtual scene image, wherein a three-dimensional stereoscopic model of the at least one player virtual object in the second virtual scene image is rendered by color into a single color corresponding to a current virtual life value; wherein the shade of the rendered color of the team member virtual object is inversely related to the current virtual life value of the team member virtual object, i.e. the higher the current virtual life value is, the lighter the rendered color is, and the lower the current virtual life value is, the darker the rendered color is;
and in response to the selection operation of any member virtual object, controlling the controlled virtual object to release the recovery-class skills to the member virtual object.
2. The method of claim 1, wherein the displaying a second virtual scene image in response to detecting the triggering action of the recovery-class skill of the controlled virtual object comprises:
in response to detecting the triggering operation of the recovery class skill of the controlled virtual object, acquiring a current virtual life value of the at least one team member virtual object;
for any member virtual object, rendering the member virtual object into a color corresponding to the current virtual life value according to the current virtual life value of the member virtual object;
and displaying the second virtual scene image.
3. The method according to claim 2, wherein the rendering the player virtual object to a color corresponding to the current virtual life value according to the current virtual life value of the player virtual object comprises:
determining a target proportion according to the current virtual life value of the team member virtual object, wherein the target proportion is the proportion of the current virtual life value to the total life value;
determining target color depth according to a proportion interval in which the target proportion is located, wherein one proportion interval corresponds to one color depth, and the color depth is used for representing the depth of color;
and rendering the player virtual object to be a color corresponding to the target color depth.
4. The method according to claim 2, wherein the rendering the player virtual object to a color corresponding to the current virtual life value according to the current virtual life value of the player virtual object comprises:
determining a target proportion according to the current virtual life value of the team member virtual object, wherein the target proportion is the proportion of the current virtual life value to the total life value;
and rendering the target proportion of the player virtual object into a target color.
5. The method of claim 4, wherein after determining a target proportion based on the current virtual life value of the player virtual object, the method further comprises:
and determining the target color corresponding to the proportion interval in which the target proportion is positioned, wherein one proportion interval corresponds to one color.
6. The method of claim 1, wherein after displaying the first virtual scene image, the method further comprises:
in response to the fact that the pressing operation of any skill icon is detected, determining a skill and pressing duration corresponding to the skill icon;
and in response to the skill belonging to the recovery-class skill and the pressing time length not less than the target time length, determining that the trigger operation of the recovery-class skill is detected.
7. The method according to claim 1, wherein said controlling the controlled virtual object to release the recovery class skills to any player virtual object in response to a selection operation of said player virtual object comprises:
in response to detecting a click operation on any player virtual object, determining the player virtual object as a target object of the recovery-class skill;
and controlling the controlled virtual object to release the recovery class skills to the team member virtual object.
8. The method according to claim 1, wherein said controlling the controlled virtual object to release the recovery class skills to any player virtual object in response to a selection operation of said player virtual object comprises:
determining a selected player virtual object according to the dragging operation of the skill icon of the recovery class skill in a target range;
and in response to the pressing ending operation, controlling the controlled virtual object to release the recovery class skills to the team member virtual object.
9. The method according to any one of claims 1 to 8, further comprising:
rendering the virtual scene and the controlled virtual object in the second virtual scene image to be lighter than the color of the at least one team member virtual object in response to detecting a triggering operation of the recovery class skill of the controlled virtual object.
10. An apparatus for displaying a virtual object, the apparatus comprising:
the system comprises a first display module, a second display module and a third display module, wherein the first display module is used for displaying a first virtual scene image, the first virtual scene image comprises a controlled virtual object and at least one team member virtual object in a virtual scene, the controlled virtual object is a virtual object controlled by a current terminal, and the team member virtual object and the controlled virtual object belong to the same team;
a second display module, configured to display a second virtual scene image in response to detecting a trigger operation of a recovery class skill of the controlled virtual object, where a three-dimensional stereoscopic model of the at least one team member virtual object in the second virtual scene image is rendered by color into a single color corresponding to a current virtual life value, respectively; wherein the shade of the rendered color of the team member virtual object is inversely related to the current virtual life value of the team member virtual object, i.e. the higher the current virtual life value is, the lighter the rendered color is, and the lower the current virtual life value is, the darker the rendered color is;
and the control module is used for responding to the selection operation of any player virtual object and controlling the controlled virtual object to release the recovery class skills to the player virtual object.
11. The apparatus of claim 10, wherein the second display module comprises:
the acquisition unit is used for responding to the trigger operation of detecting the recovery class skill of the controlled virtual object and acquiring the current virtual life value of the at least one team member virtual object;
the rendering unit is used for rendering any player virtual object into a color corresponding to the current virtual life value according to the current virtual life value of the player virtual object;
and the display unit is used for displaying the second virtual scene image.
12. The apparatus according to claim 11, wherein the rendering unit is configured to determine a target proportion according to a current virtual life value of the team member virtual object, where the target proportion is a proportion of the current virtual life value to a total life value; determining target color depth according to a proportion interval where the target proportion is located, wherein one proportion interval corresponds to one color depth, and the color depth is used for representing the depth of color; and rendering the player virtual object to be a color corresponding to the target color depth.
13. The apparatus according to claim 11, wherein the rendering unit is configured to determine a target proportion according to a current virtual life value of the team member virtual object, the target proportion being a proportion of the current virtual life value to a total life value; and rendering the target proportion of the player virtual object into a target color.
14. A computer device, characterized in that the computer device comprises a processor and a memory for storing at least one piece of computer program, which is loaded by the processor and executes the method of displaying a virtual object according to any one of claims 1 to 9.
15. A storage medium for storing at least one computer program for executing the method for displaying a virtual object according to any one of claims 1 to 9.
CN202011497277.1A 2020-12-17 2020-12-17 Virtual object display method, device, terminal and storage medium Active CN112604274B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011497277.1A CN112604274B (en) 2020-12-17 2020-12-17 Virtual object display method, device, terminal and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011497277.1A CN112604274B (en) 2020-12-17 2020-12-17 Virtual object display method, device, terminal and storage medium

Publications (2)

Publication Number Publication Date
CN112604274A CN112604274A (en) 2021-04-06
CN112604274B true CN112604274B (en) 2022-08-09

Family

ID=75240309

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011497277.1A Active CN112604274B (en) 2020-12-17 2020-12-17 Virtual object display method, device, terminal and storage medium

Country Status (1)

Country Link
CN (1) CN112604274B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113628494A (en) * 2021-08-31 2021-11-09 重庆科华安全设备有限责任公司 Immersive virtual simulation emergency rescue drilling method for fire disaster in coal mine

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160335765A1 (en) * 2015-05-12 2016-11-17 Justus W. Harris Three dimensional modeling of health data
CN106528032A (en) * 2016-12-05 2017-03-22 上海逗屋网络科技有限公司 Object display method and equipment
CN108619718A (en) * 2018-05-09 2018-10-09 网易(杭州)网络有限公司 Processing method, device, storage medium and the electronic device of virtual role

Also Published As

Publication number Publication date
CN112604274A (en) 2021-04-06

Similar Documents

Publication Publication Date Title
CN111589128B (en) Operation control display method and device based on virtual scene
CN111013142B (en) Interactive effect display method and device, computer equipment and storage medium
CN111589140B (en) Virtual object control method, device, terminal and storage medium
CN111760278B (en) Skill control display method, device, equipment and medium
CN112083848B (en) Method, device and equipment for adjusting position of control in application program and storage medium
CN111589127B (en) Control method, device and equipment of virtual role and storage medium
CN111672104B (en) Virtual scene display method, device, terminal and storage medium
CN111603771B (en) Animation generation method, device, equipment and medium
CN111589136B (en) Virtual object control method and device, computer equipment and storage medium
CN111672106B (en) Virtual scene display method and device, computer equipment and storage medium
CN111596838B (en) Service processing method and device, computer equipment and computer readable storage medium
CN112704876B (en) Method, device and equipment for selecting virtual object interaction mode and storage medium
CN111603770A (en) Virtual environment picture display method, device, equipment and medium
CN111921194A (en) Virtual environment picture display method, device, equipment and storage medium
CN112843679A (en) Skill release method, device, equipment and medium for virtual object
CN113577765A (en) User interface display method, device, equipment and storage medium
CN112755517B (en) Virtual object control method, device, terminal and storage medium
CN113599819B (en) Prompt information display method, device, equipment and storage medium
CN114130020A (en) Virtual scene display method, device, terminal and storage medium
CN114404972A (en) Method, device and equipment for displaying visual field picture
CN112156471B (en) Skill selection method, device, equipment and storage medium of virtual object
CN113457173A (en) Remote teaching method, device, computer equipment and storage medium
CN113181647A (en) Information display method, device, terminal and storage medium
CN112604274B (en) Virtual object display method, device, terminal and storage medium
CN112156454A (en) Virtual object generation method and device, terminal and readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
REG Reference to a national code

Ref country code: HK

Ref legal event code: DE

Ref document number: 40043409

Country of ref document: HK

GR01 Patent grant
GR01 Patent grant