WO2020227983A1 - Hybrid-learning neural network architecture - Google Patents

Hybrid-learning neural network architecture Download PDF

Info

Publication number
WO2020227983A1
WO2020227983A1 PCT/CN2019/087083 CN2019087083W WO2020227983A1 WO 2020227983 A1 WO2020227983 A1 WO 2020227983A1 CN 2019087083 W CN2019087083 W CN 2019087083W WO 2020227983 A1 WO2020227983 A1 WO 2020227983A1
Authority
WO
WIPO (PCT)
Prior art keywords
neural network
input feature
feature parameters
output
components
Prior art date
Application number
PCT/CN2019/087083
Other languages
French (fr)
Inventor
Zhan Li
Zhixing Ren
Yun Zhang
Jialong WANG
Original Assignee
Alibaba Group Holding Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Alibaba Group Holding Limited filed Critical Alibaba Group Holding Limited
Priority to PCT/CN2019/087083 priority Critical patent/WO2020227983A1/en
Priority to CN201980093428.6A priority patent/CN113518962A/en
Publication of WO2020227983A1 publication Critical patent/WO2020227983A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/02Knowledge representation; Symbolic representation
    • G06N5/022Knowledge engineering; Knowledge acquisition
    • G06N5/025Extracting rules from data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/088Non-supervised learning, e.g. competitive learning

Definitions

  • an environment control system such as a heating, ventilation, and air conditioning (HVAC) system
  • HVAC heating, ventilation, and air conditioning
  • the HVAC system may include many duplicative and/or similar components, such as coolers, fans, secondary pumps, air conditioners, refrigeration units, water pumps, such as cooling water pumps (CWPs) , secondary chilling water pumps (SCWPs) , and the like.
  • CRAC computer room air conditioning units
  • CFD computational fluid dynamics
  • the deep learning neural network does not depend on any of the physical models and does not distinguish various input features.
  • Classic neural networks obtain knowledge/relations only from historical data and does not have any domain knowledge. It is more difficult for general deep learning model to apply to a system having a large number of duplicate and similar devices such as a computer room of a data center.
  • these HVAC components have complex nonlinear correlations, the inputs from the sensors are treated equally from the perspective of a neural network structure, and information behind data from the inputs may be biased by the duplicative and/or similar inputs, which may result overfitting and eventual inaccuracy, causing inefficiency.
  • a popular solution is to manually aggregate the inputs based on a human expert’s domain knowledge, and to set the aggregated input as the input to the neural network.
  • this solution is room-specific and introduces extra manual work.
  • this solution relies on the experience and analysis of an HVAC expert, it is difficult to fully understand the most reasonable correlations among various HVAC components to achieve an energy efficient computer room condition in different operating conditions, such as outdoor temperature, outdoor humidity, computing load, and the like.
  • FIG. 1 illustrates an example block diagram of an environment control system used with a hybrid-learning neural network (HLNN) which may be utilized to predict power usage effectiveness (PUE) of a computer room.
  • HLNN hybrid-learning neural network
  • FIG. 2 illustrates an example detailed block diagram of the environment control system of FIG. 1 with associated levels.
  • FIG. 3 illustrates an example block diagram of the HLNN architecture.
  • FIG. 4 illustrates an example flowchart describing a process of predicting the PUE by the HLNN.
  • the two-tower deep learning architecture may learn embeddings from data and an ontology structure automatically, and may include simultaneous training of two sub-networks, an unsupervised Auto-encoder Net (AE-Net) that captures the domain knowledge embedded in the ontology and a supervised Prediction Net (P-Net) that predicts the PUE from inputs.
  • AE-Net Auto-encoder Net
  • P-Net supervised Prediction Net
  • the simultaneous learning of the AE-Net (unsupervised) and the P-Net (supervised) may be referred to as hybrid-learning, and the two-tower deep learning architecture may also be referred to as a hybrid-learning neural network (HLNN) architecture.
  • HLNN hybrid-learning neural network
  • machine learning methods may be used to learn from historical data to obtain complex relationships among various HVAC components and the energy efficiency of the computer room in different operating conditions.
  • the first and second neural networks may share a shared structure comprising one input layer and two concept layers.
  • Each of the AE-Net and the P-Net may have its own hidden layers and an output layer.
  • the AE-Net, the P-Net, and the shared structure may form a hybrid-learning neural network (HLNN) architecture.
  • the AE-Net may be an unsupervised-learning network, which may be trained to make its output copy input with a lowest possible error, while P-Net may be a deep feedforward neural network to predict the PUE.
  • Domain knowledge of the components associated with the HVAC system and the computing equipment of the computer room may be embedded into the HLNN architecture.
  • the design of the two-tower deep learning architecture of the input layer and concept layers may be guided by the domain ontology containing multiple levels of nodes where a top level may contain a root concept and a bottom level may contain multiple instance.
  • the instances in the bottom level of the ontology may be represented by the nodes in the input layer, and the concepts in the middle levels may also have corresponding nodes in concept layers of the shared structure.
  • the relations and/or connections between levels may be copied in the input and the concept layers.
  • FIG. 1 illustrates an example block diagram of an environment control system 100 used with a hybrid-learning neural network (HLNN) which may be utilized to predict power usage effectiveness (PUE) of a computer room 102.
  • HLNN hybrid-learning neural network
  • the environment control system 100 may include a plurality of components such as an equipment and data collection module 104 communicatively coupled to an HVAC group 106 and an outside equipment and data group 108.
  • the equipment and data collection module 104 may be configured to maintain profiles of components managed by the HVAC group 106 and the outside equipment and data group 108, receive input data from various sensors associated with those components, and transmit data to those components to, in part, control the environment of, and calculate a predicted PUE of, the computer room 102.
  • Some of the environment control system components may be located in the computer room 102, and other components may be located outside of a building in which the computer room 102 is located.
  • the environment control system 100 may monitor energy consumption of components associated with the computer room 102, the equipment and data collection module 104, the HVAC group 106, and the outside equipment and data group 108.
  • the environment control system 100 may be communicatively coupled to a computer 110.
  • the computer 110 may comprise one or more processors 112 and memory 114 communicatively coupled to the one or more processors 112, which may store computer-readable instructions to be executed by the computer 110 to perform functions of the HLNN described below.
  • the computer 110 may be located within the computer room 102 or may be remotely located from the computer room 102.
  • the computer room 102 may house computing equipment 116 including servers, power supplies, displays, routers, network and communication modules, and the like (not shown) .
  • the computing equipment 116 may be coupled to the environment control system 100 and may provide information regarding energy usage by the computing equipment 116 based on historical, current, and expected energy usage and computing loads for calculating the predicted PUE of the computer room 102.
  • FIG. 2 illustrates an example detailed block diagram of the environment control system 100 of FIG. 1 with associated levels (levels 1-4 shown) .
  • the HVAC group 106 may comprise an HVAC control module 202 communicatively coupled to the equipment and data collection module 104, an air conditioning group 204, and a refrigeration group 206.
  • the HVAC control module 202 may be configured to receive operating information from various sensors and controllers of the air conditioning group 204 and from the refrigeration group 206.
  • the HVAC control module 202 may forward the operating information to the equipment and data collection module 104 for calculation by the HLNN.
  • the HVAC control module 202 may also be configured to transmit control information received from the equipment and data collection module 104 to the air conditioning group 204 and the refrigeration group 206 for adjusting various parameters of the air conditioning group 204 and the refrigeration group 206 to optimize a desired parameter for predicting the PUE.
  • the HVAC group 106 may further comprise a secondary pump group (not shown) and may similarly communicate associated operating information to and from the HVAC control module 202.
  • the air conditioning group 204 may comprise N air conditioners (two, AC-1 208 and AC-N 210, shown) .
  • each of N air conditioners may comprise several controls and sensors, such as a corresponding switch, a corresponding fan speed controller/sensor, a corresponding air conditioner output air temperature sensor, and a corresponding air conditioner return air temperature sensor.
  • Each of N air conditioners may be configured to receive AC operating information from the corresponding controls and sensors and forward the AC operating information to the air conditioning system 204, which, in turn, forwards the AC operating information to the HVAC control module 202.
  • Each of N air conditioners may also be configured to transmit AC control information received from the air conditioning system 204 to the corresponding controls to optimize a desired parameter for predicting the PUE.
  • the refrigeration group 206 may comprise a plurality of refrigeration systems including a plurality of coolers (cooler-1 212 shown) and a plurality of cooling towers (tower-1 214 shown) .
  • each of the plurality of coolers may comprise associated switch, cooling mode controller, outflow cooling water temperature controller/sensor
  • each of the plurality of cooling towers may comprise associated cooling tower fan speed controller/sensor, outflow cooling water temperature controller/sensor, and return cooling water temperature controller/sensor.
  • Each of the plurality of refrigeration systems may be configured to receive refrigeration operating information from the corresponding controls, switches, and sensors (not shown) and forward the refrigeration operating information to the HVAC control module 202 via the refrigeration group 206.
  • Each of the plurality of refrigeration systems may also be configured to transmit refrigeration control information received from the refrigeration group 206 to the corresponding controls, switches, and sensors to optimize the desired parameter for predicting the PUE.
  • the outside equipment and data group 108 may comprise an outside equipment monitoring module 216 communicatively coupled to the equipment and data collection module 104, an outside humidity module 218, an outside wet bulb temperature module 220, and other modules (not shown) .
  • the outside humidity module 218 may be communicatively coupled to M humidity sensors (two humidity sensors, humidity sensor-1 222 and humidity sensor-M 224, shown) .
  • the outside wet bulb temperature module 220 may be communicatively coupled to M wet bulb temperature sensors (two wet bulb temperature sensors, wet bulb temperature sensor-1 226 and wet bulb temperature sensor-M 228, shown) .
  • the outside equipment monitoring module 216 may receive humidity and wet bulb temperature information from the corresponding sensors and forward the information to the equipment and data collection module 104 for optimizing the desired parameter for predicting the PUE.
  • Each block illustrated in FIG. 2 may be associated with one of a plurality of levels of the domain ontology.
  • the domain ontology having four levels is illustrated herein as an example, however, the number of the levels of the domain ontology may not be limited to four and may be more or less than four levels.
  • Level 1 may include the equipment and data collection module 104, which may be referred to as D1.
  • Level 2 may include q modules including the HVAC control module 202 and the outside equipment monitoring module 216, which may be referred to as C_1, C_2, ...C_q, respectively.
  • Level 3 may include p modules including the air conditioning group 204, the refrigeration group 206, the outside humidity module 218, and the outside wet bulb temperature module 220, which may be referred to as B_1, B_2, ...B_p, respectively.
  • Level 4 may include k modules including the AC-1 208, the AC-N 210, the cooler-1 212, the tower-1 214, the humidity sensor-1 222, the humidity sensor-M 224, the wet bulb temp sensor-1 226, and the wet bulb temp sensor-M 228, which may be referred to as A_1, A_2, ...A_k, respectively.
  • FIG. 3 illustrates an example block diagram of the HLNN architecture 300.
  • the HLNN structure 300 may comprise a domain ontology 302, a shared structure 304, a first neural network, such as an AE-Net 306, and a second neural network, such as a P-Net 308.
  • There may be a plurality of levels in the ontology and four levels corresponding to the blocks illustrated in FIG. 2 are shown in the domain ontology 302 as examples.
  • the top level, Level 1 may contain the root concept, D_1 310, and the bottom level, Level 4, may contain a plurality of instances, of which four instances, A_1 312, A_2 314, A_n 316, and A_k 318, are shown.
  • These four instances in Level 4 of the domain ontology 302 may be represented as nodes, A_1 320, A_2 322, A_n 324, and A_k 326, respectively, in an input layer 328 of the shared structure 304.
  • the second level, Level 2, and the third level, Level 3, of the domain ontology 302 may represent a plurality of instances, of which two concepts, C_1 330 and C_q 332, in Level 2 and three concepts, B_1 334, B_2 336, and B_p 338, in Level 3 are shown. These instances in Level 2 and Level 3 of the domain ontology 302 may also have corresponding nodes, C_1 340, C_q 342, B_1 344, B_2 346, and B_p 348, respectively, in concept layers 350 of the shared structure 304. Additionally, relations/connections between levels may also be copied in the input layer 328 and the concept layers 350.
  • the concept B _1 334 is shown to be connected to a set of instances, A_1 312, A_2 314, and A_n 316, and in the concept layer 350, the corresponding node, B_1 344 is also shown to be connected to the corresponding nodes A_1 320, A_2 322, and A_n 324 in the input layer 328.
  • the P-Net 308 may be a deep feedforward neural network and may comprise hidden layers 352 and a one-node output layer 354 to output PUE parameters 356, plus the input layer 328 and the concept layers 350 of the shared structure 304.
  • An example feed-forward operation of the P-Net 308 is described below.
  • a neutron and a node may be interchangeable used.
  • the connections may be guided by the domain knowledge, which may not fully connect the nodes in the concept layers 350.
  • the concept relation weight between two concept nodes i.e., node j and node i
  • the weighted input of node i may then be expressed as:
  • Each layer of the concept layers 350 may be mapped from a corresponding level of the concepts in the domain ontology 302.
  • L PN (a, d p )
  • o P denote the calculated output of the neural network
  • d P denote the desired output
  • the AE-Net 306 may be an unsupervised learning model comprising hidden layers 358 and an output layer 360, plus the input layer 328 and the concept layers 350 of the shared structure 304.
  • the AE-Net 306 may be designed to minimize the difference between the input from the input layer 328 of the shared structure 304 and the output from the output layer 360.
  • an input vector a from the input layer 328, a representation vector c from the top concept layer of the concept layers 350, and an output vector r (R_1 362 and R_k 364 shown) from the output layer 360 a mapping that transforms a into c may be called an encoder, and a mapping that transforms c back to r may be called a decoder.
  • the encoder may be composed of the input layer 328 and the concept layers 350, while the decoder may be composed of the hidden layers 358 and the output layer 360.
  • the training process in the AE-Net 306 may help the encoder preserve the domain knowledge in the domain ontology 302.
  • the output vector r ⁇ r 1 , r 2 , ..., r k ⁇
  • the encoder function, f ⁇ , and the decoder function, g ⁇ may be expressed as:
  • the encoder function f ⁇ and decoder function g ⁇ may then be expressed, respectively, as:
  • s f and s g are the encoder and decoder activation functions.
  • r is not an exact reconstruction of a but the parameters of a distribution p (A
  • R r) that generates a with high probability.
  • the AE-Net 306 may be trained to find a parameter set to minimize reconstruction error in the equation below:
  • L AE is the loss function or the reconstruction error.
  • Both s f and s g may be sigmoid functions.
  • the HLNN 300 may be trained in a manner similar to standard neural networks. The only difference may be that the loss function L Model may be composed of two components: the loss of the AE-Net 306, L AE , and the prediction loss L PN of the P-Net 308:
  • L PN may be biased or weighted by another constant ⁇
  • the derivatives of the loss may be expressed as
  • the equation (15) may be transformed to:
  • Equation (15) may be transformed to:
  • the equation (15) may be transformed to:
  • Th equation (16) , (17) , and (18) show that the derivatives of the loss function L Model are back propagated for learning both the AE-Net 306 and the P-Net 308.
  • the solution for the PUE may be optimized by minimizing the loss calculated by the loss function L Model as expressed by the equation (11) , which may be accomplished by setting the derivative, such as the equations (16) , (17) , and (18) , of the loss function L Model to zero and solving for the variables. Because the solution may not always converge to zero, or may take longer than an acceptable time or a number of iterations, the value for the derivative may be set to a sufficiently small and acceptable threshold value.
  • FIG. 4 illustrates an example flowchart 400 describing a process of predicting the power utilization effectiveness (PUE) by the HLNN 300.
  • the HLNN 300 may create an ontology having a plurality of levels, such as the domain ontology 302, of the components associated with the environment control system 100 associated with the computer room 102 as illustrated in FIGs. 1-3.
  • the HLNN 300 may receive information of the components associated with the environment control system 100 automatically including corresponding associated historical data, locations and physical connections, and hierarchy among the components as illustrated in FIGs. 1-3.
  • the computing equipment 116 may include servers, power supplies, displays, routers, network and communication modules (telephone, internet, wireless devices, etc. ) , and the like.
  • the relationships among components of the environment control system 100 and the computing equipment 116 may be based on loading of the computing equipment 116, such as a workload, or computing load, of the servers and an electrical load of the servers as a function of the workload of the servers.
  • the HLNN 300 may receive input feature parameters of the components associated with the environment control system 100. More specifically, the input layer 328 of the shared structure 304 may receive k instances, A_1 312, A_2 314, A_n 316, and A_k 318, from the domain ontology 302, where k is an integer. Each of k instances may have a corresponding input feature parameter in the input layer 328 (A_1 320, A_2 322, A_n 324, and A_k 326 as illustrated in FIG. 3) may belong to one or more corresponding upper concepts of a plurality of upper concepts as illustrated hierarchically in the concept layers 350.
  • both the first neural network, such as the AE-Net 306, and the second neural network, such as the P-Net 308, may be trained simultaneously.
  • a mapping that transforms a into c may be called an encoder, and a mapping that transforms c back to r may be called a decoder.
  • the encoder may be composed of the input layer 328 and the concept layers 350, while the decoder may be composed of the hidden layers 358 and the output layer 360.
  • the training process in the AE-Net 306 may help the encoder preserve the domain knowledge in the domain ontology 302.
  • the HLNN 300 may minimize the loss based on the loss function L Model by utilizing the trained AE-Net 306 and the trained P-Net 308 and predict a power usage effectiveness (PUE) of the computer room 102 at block 410.
  • the derivative of the loss function L Model such as the equations (16) , (17) , and (18) , may be to zero for solving for the variables. Because the solution may not always converge to zero, or may take longer than an acceptable time, the value for the derivative may be set to a sufficiently small and acceptable threshold value.
  • the trained neural networks may be generated automatically, and the training of the trained neural networks may be performed by using a gradient descent algorithm to implement learning of the input feature parameters for corresponding concepts.
  • An architecture of the trained neural networks may reflect deep learning of the plurality of components and associated concepts based on the relationships among the plurality of components.
  • the trained neural networks may comprise hierarchical concept layers, such as the concept layers 350, coupled between the input layer, such as the input layer 328, and an output layer, such as the output layer 354 or 360.
  • the concept layers 350 may be added between the input layer 328 and the hidden layers 352 and 358 as illustrated in FIG. 3.
  • the concept layer 350 may be embedded with domain knowledge from the domain ontology 302.
  • the concept layer 350 may construct a concept structure based on relationships among the plurality of components.
  • the concept structure may be created manually or automatically with smart components capable of communicating with each other.
  • the training portion of the HLNN 300 and the prediction the PUE utilizing the HLNN 300 may be performed separately and/or by different parties.
  • a general deep learning network may not be capable of reasonably distinguishing all duplicative and/or similar input features, and may identify the importance of each feature based entirely on historical data.
  • a structure such as the computer room 102 with a large number of duplicative and similar devices, if these duplicate and/or similar input features parameters were not categorized, aggregated or abstracted, the complexity of the network and space for learning and searching would greatly increase, requiring higher quality and quantity of data. Although, it may be easy to obtain unreasonable overfitting, it would decrease prediction accuracy.
  • Computer-readable instructions include routines, applications, application modules, program modules, programs, components, data structures, algorithms, and the like.
  • Computer-readable instructions can be implemented on various system configurations, including single-processor or multiprocessor systems, minicomputers, mainframe computers, personal computers, hand-held computing devices, microprocessor-based, programmable consumer electronics, combinations thereof, and the like.
  • the computer-readable storage media may include volatile memory (such as random-access memory (RAM) ) and/or non-volatile memory (such as read-only memory (ROM) , flash memory, etc. ) .
  • volatile memory such as random-access memory (RAM)
  • non-volatile memory such as read-only memory (ROM) , flash memory, etc.
  • the computer-readable storage media may also include additional removable storage and/or non-removable storage including, but not limited to, flash memory, magnetic storage, optical storage, and/or tape storage that may provide non-volatile storage of computer-readable instructions, data structures, program modules, and the like.
  • a non-transient computer-readable storage medium is an example of computer-readable media.
  • Computer-readable media includes at least two types of computer-readable media, namely computer-readable storage media and communications media.
  • Computer-readable storage media includes volatile and non-volatile, removable and non-removable media implemented in any process or technology for storage of information such as computer-readable instructions, data structures, program modules, or other data.
  • Computer-readable storage media includes, but is not limited to, phase change memory (PRAM) , static random-access memory (SRAM) , dynamic random-access memory (DRAM) , other types of random-access memory (RAM) , read-only memory (ROM) , electrically erasable programmable read-only memory (EEPROM) , flash memory or other memory technology, compact disk read-only memory (CD-ROM) , digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information for access by a computing device.
  • communication media may embody computer-readable instructions, data structures, program modules, or other data in a modulated data signal, such as a carrier wave, or other transmission mechanism. As defined herein, computer-readable storage media do not include communication media.
  • the computer-readable instructions stored on one or more non-transitory computer-readable storage media that, when executed by one or more processors, may perform operations described above with reference to FIGs. 1-4.
  • computer-readable instructions include routines, programs, objects, components, data structures, and the like that perform particular functions or implement particular abstract data types.
  • the order in which the operations are described is not intended to be construed as a limitation, and any number of the described operations can be combined in any order and/or in parallel to implement the processes.
  • a method comprising: receiving input feature parameters of a plurality of components associated with at least one computer room; raining a first neural network and a second neural network based on the input feature parameters; and predicting the power usage effectiveness (PUE) of the at least one computer room based on an output of the first neural network and an output of the second neural network.
  • PUE power usage effectiveness
  • the method as paragraph A recites, wherein the first neural network is an unsupervised neural network and the second network is a supervised prediction neural network.
  • training the first neural network and the second neural network based on the input feature parameters includes simultaneously training the first neural network and the second neural network based on the input feature parameters.
  • receiving the input feature parameters of the plurality of components associated with the at least one computer room includes: creating an ontology having a plurality of levels associated with the plurality of components; and receiving information of the plurality of components based on the on the ontology including corresponding associated concept, historical data, locations, physical connections, and hierarchy among the plurality of components.
  • training the first neural network and the second neural network based on the input feature parameters includes using a gradient descent algorithm to implement learning of the input feature parameters for corresponding concepts.
  • predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes minimizing a total loss calculated based on the biased loss associated with the first neural network and an unbiased loss associated with the second neural network.
  • a system comprising: one or more processors; and memory communicatively coupled to the one or more processors, the memory storing computer-readable instructions executable by one or more processors, that when executed by the one or more processors, cause the one or more processors to perform operations comprising: receiving input feature parameters of a plurality of components associated with at least one computer room; training a first neural network and a second neural network based on the input feature parameters; and predicting the power usage effectiveness (PUE) of the at least one computer room based on an output of the first neural network and an output of the second neural network.
  • PUE power usage effectiveness
  • the system as paragraph M recites, wherein the first neural network is an unsupervised neural network and the second network is a supervised prediction neural network.
  • training the first neural network and the second neural network based on the input feature parameters includes simultaneously training the first neural network and the second neural network based on the input feature parameters.
  • receiving the input feature parameters of the plurality of components associated with the at least one computer room includes: creating an ontology having a plurality of levels associated with the plurality of components; and receiving information of the plurality of components based on the on the ontology including corresponding associated concept, historical data, locations, physical connections, and hierarchy among the plurality of components.
  • training the first neural network and the second neural network based on the input feature parameters includes using a gradient descent algorithm to implement learning of the input feature parameters for corresponding concepts.
  • predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes minimizing a total loss calculated based on the biased loss associated with the first neural network and an unbiased loss associated with the second neural network.
  • a non-transitory computer-readable storage medium storing computer-readable instructions executable by one or more processors, that when executed by the one or more processors, cause the one or more processors to perform operations comprising: receiving input feature parameters of a plurality of components associated with at least one computer room; training a first neural network and a second neural network based on the input feature parameters; and predicting the power usage effectiveness (PUE) of the at least one computer room based on an output of the first neural network and an output of the second neural network.
  • PUE power usage effectiveness
  • the non-transitory computer-readable storage medium as paragraph Y recites, wherein the first neural network is an unsupervised neural network and the second network is a supervised prediction neural network.
  • non-transitory computer-readable storage medium as paragraph Y recites, wherein training the first neural network and the second neural network based on the input feature parameters includes simultaneously training the first neural network and the second neural network based on the input feature parameters.
  • receiving the input feature parameters of the plurality of components associated with the at least one computer room includes: creating an ontology having a plurality of levels associated with the plurality of components; and receiving information of the plurality of components based on the on the ontology including corresponding associated concept, historical data, locations, physical connections, and hierarchy among the plurality of components.
  • non-transitory computer-readable storage medium as paragraph AB recites, wherein the relationships among the plurality of components are based, at least in part, on loading of computing equipment in the computer room.
  • AD The non-transitory computer-readable storage medium as paragraph AC recites, wherein the loading of the computing equipment includes a workload of the computing equipment and an electrical load used by the computing equipment.
  • the non-transitory computer-readable storage medium as paragraph AD recites, wherein the computing equipment includes a server and a power supply for the server.
  • non-transitory computer-readable storage medium as paragraph AB recites, wherein training the first neural network and the second neural network based on the input feature parameters includes using a gradient descent algorithm to implement learning of the input feature parameters for corresponding concepts.
  • non-transitory computer-readable storage medium as paragraph Y recites, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes biasing a loss associated with the first neural network with a constant value.
  • non-transitory computer-readable storage medium as paragraph AG recites, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network t includes minimizing a total loss calculated based on the biased loss associated with the first neural network and an unbiased loss associated with the second neural network.
  • AI The non-transitory computer-readable storage medium as paragraph AH recites, wherein minimizing the loss calculated based on the loss function by utilizing the first neural network and the second neural network includes solving for a derivative of the loss function equaling zero.
  • non-transitory computer-readable storage medium as paragraph AH recites, wherein minimizing the loss calculated based on the loss function by utilizing the trained first neural network and the trained second neural network includes solving for a derivative of the loss function being less than or equal to a threshold value.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Mathematical Physics (AREA)
  • Computing Systems (AREA)
  • Molecular Biology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

Systems and methods for predicting energy efficiency of a computer room in a data center, and more specifically to predicting power usage effectiveness (PUE) of a computer room with optimized parameters using a two-tower deep learning architecture are provided. The two-tower deep learning architecture may learn embeddings from data and an ontology structure automatically, and may include training of two sub-networks, such as a first neural network that captures the domain knowledge embedded in the ontology and a second neural network that predicts the PUE from inputs. The learning of the first neural network, which may be unsupervised, and the second neural network, which may be supervised, may be simultaneous and referred to as hybrid-learning, and the two-tower deep learning architecture may also be referred to as a hybrid-learning neural network (HLNN) architecture.

Description

HYBRID-LEARNING NEURAL NETWORK ARCHITECTURE BACKGROUND
In a computer room of a data center, an environment control system, such as a heating, ventilation, and air conditioning (HVAC) system, is provided to maintain an acceptable operating environment for computing equipment that includes components such as servers, power supplies, displays, routers, network and communication modules, and the like, in the computer room. Based on the total energy consumed by the computer room and the total energy consumed by the computing equipment, power usage effectiveness (PUE) , may be calculated and used to assess the energy efficiency of the computer room. The HVAC system may include many duplicative and/or similar components, such as coolers, fans, secondary pumps, air conditioners, refrigeration units, water pumps, such as cooling water pumps (CWPs) , secondary chilling water pumps (SCWPs) , and the like. For example, it is not unusual to equip more than fifty computer room air conditioning units (CRAC) in one computer room with tens of temperature and humidity sensors.
One method used to optimize the PUE is computational fluid dynamics (CFD) , which utilizes numerical analysis and data structure to analyze and solve problems involving fluid. The CFD contains partial differential equations to provide predictions of air flows and heat distributions in computer rooms, and it is widely used in the design phase. However, CFD-based methods  are computationally intensive and are not well-suited for real-time operations. Moreover, extensive verifications and validations are needed to guarantee accuracy of the simulation, especially when there are changes involved.
Another method that may be used to predict the PUE and/or control the HVAC system is a deep learning neural network. The deep learning neural network does not depend on any of the physical models and does not distinguish various input features. Classic neural networks obtain knowledge/relations only from historical data and does not have any domain knowledge. It is more difficult for general deep learning model to apply to a system having a large number of duplicate and similar devices such as a computer room of a data center. Although these HVAC components have complex nonlinear correlations, the inputs from the sensors are treated equally from the perspective of a neural network structure, and information behind data from the inputs may be biased by the duplicative and/or similar inputs, which may result overfitting and eventual inaccuracy, causing inefficiency.
To avoid the duplicative and/or similar inputs and to improve the PUE or the energy consumption of the computer room, a popular solution is to manually aggregate the inputs based on a human expert’s domain knowledge, and to set the aggregated input as the input to the neural network. However, this solution is room-specific and introduces extra manual work. Further, because this solution relies on the experience and analysis of an HVAC expert, it is difficult to fully understand the most reasonable correlations among various HVAC components to achieve an energy efficient computer room condition in  different operating conditions, such as outdoor temperature, outdoor humidity, computing load, and the like.
BRIEF DESCRIPTION OF THE DRAWINGS
The detailed description is set forth with reference to the accompanying figures. In the figures, the left-most digit (s) of a reference number identifies the figure in which the reference number first appears. The use of the same reference numbers in different figures indicates similar or identical items or features.
FIG. 1 illustrates an example block diagram of an environment control system used with a hybrid-learning neural network (HLNN) which may be utilized to predict power usage effectiveness (PUE) of a computer room.
FIG. 2 illustrates an example detailed block diagram of the environment control system of FIG. 1 with associated levels.
FIG. 3 illustrates an example block diagram of the HLNN architecture.
FIG. 4 illustrates an example flowchart describing a process of predicting the PUE by the HLNN.
DETAILED DESCRIPTION
Systems and methods discussed herein are directed to predicting energy efficiency of a computer room in a data center, and more specifically to predicting power usage effectiveness (PUE) of a computer room with optimized parameters using a two-tower deep learning architecture. The two-tower deep  learning architecture may learn embeddings from data and an ontology structure automatically, and may include simultaneous training of two sub-networks, an unsupervised Auto-encoder Net (AE-Net) that captures the domain knowledge embedded in the ontology and a supervised Prediction Net (P-Net) that predicts the PUE from inputs. The simultaneous learning of the AE-Net (unsupervised) and the P-Net (supervised) may be referred to as hybrid-learning, and the two-tower deep learning architecture may also be referred to as a hybrid-learning neural network (HLNN) architecture.
To achieve PUE optimization by ensuring a reasonable and appropriate operating environment, such as the environment of a computer room, and reducing waste in setting components of an environment control system, such as an HVAC system, machine learning methods may be used to learn from historical data to obtain complex relationships among various HVAC components and the energy efficiency of the computer room in different operating conditions.
In the HLNN architecture, the first and second neural networks, such as the AE-Net and the P-Net respectively, may share a shared structure comprising one input layer and two concept layers. Each of the AE-Net and the P-Net may have its own hidden layers and an output layer. The AE-Net, the P-Net, and the shared structure may form a hybrid-learning neural network (HLNN) architecture. The AE-Net may be an unsupervised-learning network, which may be trained to make its output copy input with a lowest possible error, while P-Net may be a deep feedforward neural network to predict the PUE.
Domain knowledge of the components associated with the HVAC system and the computing equipment of the computer room, for example, may be embedded into the HLNN architecture. By embedding the domain knowledge of the components, the number of inputs and the complexity of a search space may be reduced and the accuracy of the PUE prediction may be increased. The design of the two-tower deep learning architecture of the input layer and concept layers may be guided by the domain ontology containing multiple levels of nodes where a top level may contain a root concept and a bottom level may contain multiple instance. The instances in the bottom level of the ontology may be represented by the nodes in the input layer, and the concepts in the middle levels may also have corresponding nodes in concept layers of the shared structure. Moreover, the relations and/or connections between levels may be copied in the input and the concept layers.
FIG. 1 illustrates an example block diagram of an environment control system 100 used with a hybrid-learning neural network (HLNN) which may be utilized to predict power usage effectiveness (PUE) of a computer room 102.
The environment control system 100 may include a plurality of components such as an equipment and data collection module 104 communicatively coupled to an HVAC group 106 and an outside equipment and data group 108. The equipment and data collection module 104 may be configured to maintain profiles of components managed by the HVAC group 106 and the outside equipment and data group 108, receive input data from various sensors associated with those components, and transmit data to those  components to, in part, control the environment of, and calculate a predicted PUE of, the computer room 102. Some of the environment control system components may be located in the computer room 102, and other components may be located outside of a building in which the computer room 102 is located. The environment control system 100 may monitor energy consumption of components associated with the computer room 102, the equipment and data collection module 104, the HVAC group 106, and the outside equipment and data group 108. In addition, the environment control system 100 may be communicatively coupled to a computer 110. The computer 110 may comprise one or more processors 112 and memory 114 communicatively coupled to the one or more processors 112, which may store computer-readable instructions to be executed by the computer 110 to perform functions of the HLNN described below. The computer 110 may be located within the computer room 102 or may be remotely located from the computer room 102.
The computer room 102 may house computing equipment 116 including servers, power supplies, displays, routers, network and communication modules, and the like (not shown) . The computing equipment 116 may be coupled to the environment control system 100 and may provide information regarding energy usage by the computing equipment 116 based on historical, current, and expected energy usage and computing loads for calculating the predicted PUE of the computer room 102.
FIG. 2 illustrates an example detailed block diagram of the environment control system 100 of FIG. 1 with associated levels (levels 1-4 shown) .
The HVAC group 106 may comprise an HVAC control module 202 communicatively coupled to the equipment and data collection module 104, an air conditioning group 204, and a refrigeration group 206. The HVAC control module 202 may be configured to receive operating information from various sensors and controllers of the air conditioning group 204 and from the refrigeration group 206. The HVAC control module 202 may forward the operating information to the equipment and data collection module 104 for calculation by the HLNN. The HVAC control module 202 may also be configured to transmit control information received from the equipment and data collection module 104 to the air conditioning group 204 and the refrigeration group 206 for adjusting various parameters of the air conditioning group 204 and the refrigeration group 206 to optimize a desired parameter for predicting the PUE. The HVAC group 106 may further comprise a secondary pump group (not shown) and may similarly communicate associated operating information to and from the HVAC control module 202.
The air conditioning group 204 may comprise N air conditioners (two, AC-1 208 and AC-N 210, shown) . Although not shown, each of N air conditioners may comprise several controls and sensors, such as a corresponding switch, a corresponding fan speed controller/sensor, a corresponding air conditioner output air temperature sensor, and a corresponding air conditioner  return air temperature sensor. Each of N air conditioners may be configured to receive AC operating information from the corresponding controls and sensors and forward the AC operating information to the air conditioning system 204, which, in turn, forwards the AC operating information to the HVAC control module 202. Each of N air conditioners may also be configured to transmit AC control information received from the air conditioning system 204 to the corresponding controls to optimize a desired parameter for predicting the PUE.
The refrigeration group 206 may comprise a plurality of refrigeration systems including a plurality of coolers (cooler-1 212 shown) and a plurality of cooling towers (tower-1 214 shown) . Although not shown, each of the plurality of coolers may comprise associated switch, cooling mode controller, outflow cooling water temperature controller/sensor, and each of the plurality of cooling towers may comprise associated cooling tower fan speed controller/sensor, outflow cooling water temperature controller/sensor, and return cooling water temperature controller/sensor.
Each of the plurality of refrigeration systems may be configured to receive refrigeration operating information from the corresponding controls, switches, and sensors (not shown) and forward the refrigeration operating information to the HVAC control module 202 via the refrigeration group 206. Each of the plurality of refrigeration systems may also be configured to transmit refrigeration control information received from the refrigeration group 206 to the corresponding controls, switches, and sensors to optimize the desired parameter for predicting the PUE.
The outside equipment and data group 108 may comprise an outside equipment monitoring module 216 communicatively coupled to the equipment and data collection module 104, an outside humidity module 218, an outside wet bulb temperature module 220, and other modules (not shown) . The outside humidity module 218 may be communicatively coupled to M humidity sensors (two humidity sensors, humidity sensor-1 222 and humidity sensor-M 224, shown) . The outside wet bulb temperature module 220 may be communicatively coupled to M wet bulb temperature sensors (two wet bulb temperature sensors, wet bulb temperature sensor-1 226 and wet bulb temperature sensor-M 228, shown) . The outside equipment monitoring module 216 may receive humidity and wet bulb temperature information from the corresponding sensors and forward the information to the equipment and data collection module 104 for optimizing the desired parameter for predicting the PUE.
Each block illustrated in FIG. 2 may be associated with one of a plurality of levels of the domain ontology. The domain ontology having four levels is illustrated herein as an example, however, the number of the levels of the domain ontology may not be limited to four and may be more or less than four levels. Level 1 may include the equipment and data collection module 104, which may be referred to as D1. Level 2 may include q modules including the HVAC control module 202 and the outside equipment monitoring module 216, which may be referred to as C_1, C_2, …C_q, respectively. Level 3 may include p modules including the air conditioning group 204, the refrigeration group 206, the outside humidity module 218, and the outside wet bulb temperature module  220, which may be referred to as B_1, B_2, …B_p, respectively. Level 4 may include k modules including the AC-1 208, the AC-N 210, the cooler-1 212, the tower-1 214, the humidity sensor-1 222, the humidity sensor-M 224, the wet bulb temp sensor-1 226, and the wet bulb temp sensor-M 228, which may be referred to as A_1, A_2, …A_k, respectively.
FIG. 3 illustrates an example block diagram of the HLNN architecture 300.
The HLNN structure 300 may comprise a domain ontology 302, a shared structure 304, a first neural network, such as an AE-Net 306, and a second neural network, such as a P-Net 308. There may be a plurality of levels in the ontology, and four levels corresponding to the blocks illustrated in FIG. 2 are shown in the domain ontology 302 as examples. The top level, Level 1, may contain the root concept, D_1 310, and the bottom level, Level 4, may contain a plurality of instances, of which four instances, A_1 312, A_2 314, A_n 316, and A_k 318, are shown. These four instances in Level 4 of the domain ontology 302 may be represented as nodes, A_1 320, A_2 322, A_n 324, and A_k 326, respectively, in an input layer 328 of the shared structure 304.
The second level, Level 2, and the third level, Level 3, of the domain ontology 302 may represent a plurality of instances, of which two concepts, C_1 330 and C_q 332, in Level 2 and three concepts, B_1 334, B_2 336, and B_p 338, in Level 3 are shown. These instances in Level 2 and Level 3 of the domain ontology 302 may also have corresponding nodes, C_1 340, C_q 342, B_1 344, B_2 346, and B_p 348, respectively, in concept layers 350 of the shared structure  304. Additionally, relations/connections between levels may also be copied in the input layer 328 and the concept layers 350. For example, in the domain ontology 302, the concept B _1 334 is shown to be connected to a set of instances, A_1 312, A_2 314, and A_n 316, and in the concept layer 350, the corresponding node, B_1 344 is also shown to be connected to the corresponding nodes A_1 320, A_2 322, and A_n 324 in the input layer 328.
The P-Net 308 may be a deep feedforward neural network and may comprise hidden layers 352 and a one-node output layer 354 to output PUE parameters 356, plus the input layer 328 and the concept layers 350 of the shared structure 304. An example feed-forward operation of the P-Net 308 is described below. A neutron and a node may be interchangeable used.
Let
Figure PCTCN2019087083-appb-000001
denote the weight between the j th neuron, or node, in the (l-1)  th layer and the i th neuron in the l th layer, and
Figure PCTCN2019087083-appb-000002
be the bias of the i th neuron in the l th layer. With these notations, the feed-forward operation may be described as
Figure PCTCN2019087083-appb-000003
where
Figure PCTCN2019087083-appb-000004
is the weighted input of node i, 
Figure PCTCN2019087083-appb-000005
is the output of the j th node in the (l-1)  th layer, and R l-1 is the number of neurons in the (l-1)  th layer.
Given
Figure PCTCN2019087083-appb-000006
and
Figure PCTCN2019087083-appb-000007
the equation (1) may be simplified to:
Figure PCTCN2019087083-appb-000008
Using the notation above, the activation of node i is
Figure PCTCN2019087083-appb-000009
where f p is the activation function.
In the shared structure 304, the connections may be guided by the domain knowledge, which may not fully connect the nodes in the concept layers 350. Let
Figure PCTCN2019087083-appb-000010
denote the concept relation weight between two concept nodes, i.e., node j and node i, the weighted input of node i may then be expressed as:
Figure PCTCN2019087083-appb-000011
Each layer of the concept layers 350 may be mapped from a corresponding level of the concepts in the domain ontology 302. Let
Figure PCTCN2019087083-appb-000012
denote the number of nodes (instances and concepts) in the concept hierarchy that are connected to the i th node, i.e., node C_i in the domain ontology 302, in the l th layer, then the concept relation weight
Figure PCTCN2019087083-appb-000013
may be expressed as
Figure PCTCN2019087083-appb-000014
That is, if the corresponding sub-concept/instance node j in the (l-1)  th layer is not connected to the concept of node i, then
Figure PCTCN2019087083-appb-000015
is zero. If
Figure PCTCN2019087083-appb-000016
the concept relation weights are 1, which do not affect the learning process. The loss function, L PN (a, d p) , may then define the error between d P and o P for the input a by letting a denote the input vector, o P denote the calculated output of the neural network, and d P denote the desired output.
The AE-Net 306 may be an unsupervised learning model comprising hidden layers 358 and an output layer 360, plus the input layer 328 and the concept layers 350 of the shared structure 304. The AE-Net 306 may be designed to minimize the difference between the input from the input layer 328 of the shared structure 304 and the output from the output layer 360. Given that an input  vector a from the input layer 328, a representation vector c from the top concept layer of the concept layers 350, and an output vector r (R_1 362 and R_k 364 shown) from the output layer 360, a mapping that transforms a into c may be called an encoder, and a mapping that transforms c back to r may be called a decoder. The encoder may be composed of the input layer 328 and the concept layers 350, while the decoder may be composed of the hidden layers 358 and the output layer 360. The training process in the AE-Net 306 may help the encoder preserve the domain knowledge in the domain ontology 302.
Setting the input vector a as a = {a 1, a 2, …, a k} , the representation vector c may be expressed as c = {c 1, c 2, …, c q} , and the output vector r as r = {r 1, r 2, …, r k} , then the encoder function, f θ, and the decoder function, g θ, may be expressed as:
c = f θ (a)              (6)
r = g θ (c)            (7)
In both encoder function f θ and decoder function g θ, the parameter set is θ = {W, b, W’, d} , where W and W’ are the encoder and decoder weight matrices, and b and d are encoder and decoder bias vectors. The encoder function f θ and decoder function g θ may then be expressed, respectively, as:
f θ (a) = s f (b + Wa)          (8)
g θ (c) = s g (d + W’c)           (9)
where s f and s g are the encoder and decoder activation functions. In probabilistic terms r is not an exact reconstruction of a but the parameters of a distribution  p (A|R = r) that generates a with high probability. The AE-Net 306 may be trained to find a parameter set to minimize reconstruction error in the equation below:
E AE (θ) = ∑ a∈AL AE (a, r) = ∑ a∈AL AE (a, g θ (f θ (a) ) )         (10)
where A denotes the training set of examples, L AE is the loss function or the reconstruction error. The input vector may be of real-value and the loss function L AE may be squared error L AE (a, r) = ||a -r|| 2. Both s f and s g may be sigmoid functions.
The HLNN 300 may be trained in a manner similar to standard neural networks. The only difference may be that the loss function L Model may be composed of two components: the loss of the AE-Net 306, L AE, and the prediction loss L PN of the P-Net 308:
L Model = L PN + αL AE          (11)
where α is a constant providing a bias or weight to L AE. Alternatively, L PN may be biased or weighted by another constant β, and L Model may be expressed as L Model = β L PN + L AE.
In the training, the derivatives of the loss may be expressed as
Figure PCTCN2019087083-appb-000017
and the following substitutions may be made:
Figure PCTCN2019087083-appb-000018
Figure PCTCN2019087083-appb-000019
where R l+1 is the number of nodes in the (l+1)  th layer. Combining equations (12) , (13) , and (14) , yields
Figure PCTCN2019087083-appb-000020
If the layer l+1 is in the shared structure 304, i.e., the input layer 328 and the concept layers 350, the equation (15) may be transformed to:
Figure PCTCN2019087083-appb-000021
If the layer l+1 is in the hidden layers 352 or the output layer 354 of the P-Net 308, the equation (15) may be transformed to:
Figure PCTCN2019087083-appb-000022
where
Figure PCTCN2019087083-appb-000023
denotes the number of nodes in the (l+1)  th layer that are used to calculate the output of the P-Net 308.
If the layer l+1 is the hidden layers 358 or the output layer 360 of the AE-Net 306, the equation (15) may be transformed to:
Figure PCTCN2019087083-appb-000024
where
Figure PCTCN2019087083-appb-000025
denotes the number of nodes in the (l+1)  th layer that are used to calculate the output of the AE-Net 306.
Th equation (16) , (17) , and (18) show that the derivatives of the loss function L Model are back propagated for learning both the AE-Net 306 and the P-Net 308. The solution for the PUE may be optimized by minimizing the loss calculated by the loss function L Model as expressed by the equation (11) , which may be accomplished by setting the derivative, such as the equations (16) , (17) , and (18) , of the loss function L Model to zero and solving for the variables. Because the solution may not always converge to zero, or may take longer than an acceptable time or a number of iterations, the value for the derivative may be set to a sufficiently small and acceptable threshold value.
FIG. 4 illustrates an example flowchart 400 describing a process of predicting the power utilization effectiveness (PUE) by the HLNN 300.
At block 402, the HLNN 300 may create an ontology having a plurality of levels, such as the domain ontology 302, of the components associated with the environment control system 100 associated with the computer room 102 as illustrated in FIGs. 1-3. The HLNN 300 may receive information of the components associated with the environment control system 100 automatically including corresponding associated historical data, locations and physical connections, and hierarchy among the components as illustrated in FIGs. 1-3. The computing equipment 116 may include servers, power supplies, displays, routers, network and communication modules (telephone, internet, wireless devices, etc. ) , and the like. The relationships among components of the environment control system 100 and the computing equipment 116 may be based on loading of the computing equipment 116, such as a workload, or computing  load, of the servers and an electrical load of the servers as a function of the workload of the servers.
At block 404, the HLNN 300 may receive input feature parameters of the components associated with the environment control system 100. More specifically, the input layer 328 of the shared structure 304 may receive k instances, A_1 312, A_2 314, A_n 316, and A_k 318, from the domain ontology 302, where k is an integer. Each of k instances may have a corresponding input feature parameter in the input layer 328 (A_1 320, A_2 322, A_n 324, and A_k 326 as illustrated in FIG. 3) may belong to one or more corresponding upper concepts of a plurality of upper concepts as illustrated hierarchically in the concept layers 350.
At block 406, both the first neural network, such as the AE-Net 306, and the second neural network, such as the P-Net 308, may be trained simultaneously. As discussed above, the input vector a, or the input feature parameters, may be expressed as a = {a 1, a 2, …, a k} , the representation vector c, or the concepts, as c = {c 1, c 2, …, c q} , and the output vector r as r = {r 1, r 2, …, r k} . A mapping that transforms a into c may be called an encoder, and a mapping that transforms c back to r may be called a decoder. The encoder may be composed of the input layer 328 and the concept layers 350, while the decoder may be composed of the hidden layers 358 and the output layer 360. The training process in the AE-Net 306 may help the encoder preserve the domain knowledge in the domain ontology 302.
At block 408, the HLNN 300 may minimize the loss based on the loss function L Model by utilizing the trained AE-Net 306 and the trained P-Net 308 and predict a power usage effectiveness (PUE) of the computer room 102 at block 410. The derivative of the loss function L Model, such as the equations (16) , (17) , and (18) , may be to zero for solving for the variables. Because the solution may not always converge to zero, or may take longer than an acceptable time, the value for the derivative may be set to a sufficiently small and acceptable threshold value.
The trained neural networks may be generated automatically, and the training of the trained neural networks may be performed by using a gradient descent algorithm to implement learning of the input feature parameters for corresponding concepts. An architecture of the trained neural networks may reflect deep learning of the plurality of components and associated concepts based on the relationships among the plurality of components. The trained neural networks may comprise hierarchical concept layers, such as the concept layers 350, coupled between the input layer, such as the input layer 328, and an output layer, such as the  output layer  354 or 360. The concept layers 350 may be added between the input layer 328 and the hidden layers 352 and 358 as illustrated in FIG. 3. The concept layer 350 may be embedded with domain knowledge from the domain ontology 302. The concept layer 350 may construct a concept structure based on relationships among the plurality of components. The concept structure may be created manually or automatically with smart components capable of communicating with each other. The training portion of the HLNN  300 and the prediction the PUE utilizing the HLNN 300 may be performed separately and/or by different parties.
A general deep learning network may not be capable of reasonably distinguishing all duplicative and/or similar input features, and may identify the importance of each feature based entirely on historical data. In a structure, such as the computer room 102 with a large number of duplicative and similar devices, if these duplicate and/or similar input features parameters were not categorized, aggregated or abstracted, the complexity of the network and space for learning and searching would greatly increase, requiring higher quality and quantity of data. Although, it may be easy to obtain unreasonable overfitting, it would decrease prediction accuracy.
Some or all operations of the methods described above can be performed by execution of computer-readable instructions stored on a computer-readable storage medium, as defined below. The term “computer-readable instructions” as used in the description and claims, include routines, applications, application modules, program modules, programs, components, data structures, algorithms, and the like. Computer-readable instructions can be implemented on various system configurations, including single-processor or multiprocessor systems, minicomputers, mainframe computers, personal computers, hand-held computing devices, microprocessor-based, programmable consumer electronics, combinations thereof, and the like.
The computer-readable storage media may include volatile memory (such as random-access memory (RAM) ) and/or non-volatile memory (such as  read-only memory (ROM) , flash memory, etc. ) . The computer-readable storage media may also include additional removable storage and/or non-removable storage including, but not limited to, flash memory, magnetic storage, optical storage, and/or tape storage that may provide non-volatile storage of computer-readable instructions, data structures, program modules, and the like.
A non-transient computer-readable storage medium is an example of computer-readable media. Computer-readable media includes at least two types of computer-readable media, namely computer-readable storage media and communications media. Computer-readable storage media includes volatile and non-volatile, removable and non-removable media implemented in any process or technology for storage of information such as computer-readable instructions, data structures, program modules, or other data. Computer-readable storage media includes, but is not limited to, phase change memory (PRAM) , static random-access memory (SRAM) , dynamic random-access memory (DRAM) , other types of random-access memory (RAM) , read-only memory (ROM) , electrically erasable programmable read-only memory (EEPROM) , flash memory or other memory technology, compact disk read-only memory (CD-ROM) , digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information for access by a computing device. In contrast, communication media may embody computer-readable instructions, data structures, program modules, or other data in a modulated data signal, such as a carrier wave, or other transmission  mechanism. As defined herein, computer-readable storage media do not include communication media.
The computer-readable instructions stored on one or more non-transitory computer-readable storage media that, when executed by one or more processors, may perform operations described above with reference to FIGs. 1-4. Generally, computer-readable instructions include routines, programs, objects, components, data structures, and the like that perform particular functions or implement particular abstract data types. The order in which the operations are described is not intended to be construed as a limitation, and any number of the described operations can be combined in any order and/or in parallel to implement the processes.
EXAMPLE CLAUSES
A. A method comprising: receiving input feature parameters of a plurality of components associated with at least one computer room; raining a first neural network and a second neural network based on the input feature parameters; and predicting the power usage effectiveness (PUE) of the at least one computer room based on an output of the first neural network and an output of the second neural network.
B. The method as paragraph A recites, wherein the first neural network is an unsupervised neural network and the second network is a supervised prediction neural network.
C. The method as paragraph A recites, wherein training the first neural network and the second neural network based on the input feature parameters includes simultaneously training the first neural network and the second neural network based on the input feature parameters.
D. The method as paragraph A recites, wherein receiving the input feature parameters of the plurality of components associated with the at least one computer room includes: creating an ontology having a plurality of levels associated with the plurality of components; and receiving information of the plurality of components based on the on the ontology including corresponding associated concept, historical data, locations, physical connections, and hierarchy among the plurality of components.
E. The method as paragraph D recites, wherein the relationships among the plurality of components are based, at least in part, on loading of computing equipment in the computer room.
F. The method as paragraph E recites, wherein the loading of the computing equipment includes a workload of the computing equipment and an electrical load used by the computing equipment.
G. The method as paragraph F recites, wherein the computing equipment includes a server and a power supply for the server.
H. The method as paragraph D recites, wherein training the first neural network and the second neural network based on the input feature parameters includes using a gradient descent algorithm to implement learning of the input feature parameters for corresponding concepts.
I. The method as paragraph A recites, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes biasing a loss associated with the first neural network with a constant value.
J. The method as paragraph I recites, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes minimizing a total loss calculated based on the biased loss associated with the first neural network and an unbiased loss associated with the second neural network.
K. The method as paragraph J recites wherein minimizing the loss calculated based on the loss function by utilizing the first neural network and the second neural network includes solving for a derivative of the loss function equaling zero.
L. The method as paragraph J recites, wherein minimizing the loss calculated based on the loss function by utilizing the trained first neural network and the trained second neural network includes solving for a derivative of the loss function being less than or equal to a threshold value.
M. A system comprising: one or more processors; and memory communicatively coupled to the one or more processors, the memory storing computer-readable instructions executable by one or more processors, that when executed by the one or more processors, cause the one or more processors to perform operations comprising: receiving input feature parameters of a plurality of components associated with at least one computer room; training a first neural  network and a second neural network based on the input feature parameters; and predicting the power usage effectiveness (PUE) of the at least one computer room based on an output of the first neural network and an output of the second neural network.
N. The system as paragraph M recites, wherein the first neural network is an unsupervised neural network and the second network is a supervised prediction neural network.
O. The system as paragraph M recites, wherein training the first neural network and the second neural network based on the input feature parameters includes simultaneously training the first neural network and the second neural network based on the input feature parameters.
P. The system as paragraph M recites, wherein receiving the input feature parameters of the plurality of components associated with the at least one computer room includes: creating an ontology having a plurality of levels associated with the plurality of components; and receiving information of the plurality of components based on the on the ontology including corresponding associated concept, historical data, locations, physical connections, and hierarchy among the plurality of components.
Q. The system as paragraph P recites, wherein the relationships among the plurality of components are based, at least in part, on loading of the computing equipment.
R. The system as paragraph Q recites, wherein the loading of the computing equipment includes a workload of the computing equipment and an electrical load used by the computing equipment.
S. The system as paragraph R recites, wherein the computing equipment includes a server and a power supply for the server.
T. The system as paragraph P recites, wherein training the first neural network and the second neural network based on the input feature parameters includes using a gradient descent algorithm to implement learning of the input feature parameters for corresponding concepts.
U. The system as paragraph M recites, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes biasing a loss associated with the first neural network with a constant value.
V. The system as paragraph U recites, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes minimizing a total loss calculated based on the biased loss associated with the first neural network and an unbiased loss associated with the second neural network.
W. The system as paragraph V recites, wherein minimizing the loss calculated based on the loss function by utilizing the first neural network and the second neural network includes solving for a derivative of the loss function equaling zero.
X. The system as paragraph V recites, wherein minimizing the loss calculated based on the loss function by utilizing the trained first neural network and the trained second neural network includes solving for a derivative of the loss function being less than or equal to a threshold value.
Y. A non-transitory computer-readable storage medium storing computer-readable instructions executable by one or more processors, that when executed by the one or more processors, cause the one or more processors to perform operations comprising: receiving input feature parameters of a plurality of components associated with at least one computer room; training a first neural network and a second neural network based on the input feature parameters; and predicting the power usage effectiveness (PUE) of the at least one computer room based on an output of the first neural network and an output of the second neural network.
Z. The non-transitory computer-readable storage medium as paragraph Y recites, wherein the first neural network is an unsupervised neural network and the second network is a supervised prediction neural network.
AA. The non-transitory computer-readable storage medium as paragraph Y recites, wherein training the first neural network and the second neural network based on the input feature parameters includes simultaneously training the first neural network and the second neural network based on the input feature parameters.
AB. The non-transitory computer-readable storage medium as paragraph Y recites, wherein receiving the input feature parameters  of the plurality of components associated with the at least one computer room includes: creating an ontology having a plurality of levels associated with the plurality of components; and receiving information of the plurality of components based on the on the ontology including corresponding associated concept, historical data, locations, physical connections, and hierarchy among the plurality of components.
AC. The non-transitory computer-readable storage medium as paragraph AB recites, wherein the relationships among the plurality of components are based, at least in part, on loading of computing equipment in the computer room.
AD. The non-transitory computer-readable storage medium as paragraph AC recites, wherein the loading of the computing equipment includes a workload of the computing equipment and an electrical load used by the computing equipment.
AE. The non-transitory computer-readable storage medium as paragraph AD recites, wherein the computing equipment includes a server and a power supply for the server.
AF. The non-transitory computer-readable storage medium as paragraph AB recites, wherein training the first neural network and the second neural network based on the input feature parameters includes using a gradient descent algorithm to implement learning of the input feature parameters for corresponding concepts.
AG. The non-transitory computer-readable storage medium as paragraph Y recites, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes biasing a loss associated with the first neural network with a constant value.
AH. The non-transitory computer-readable storage medium as paragraph AG recites, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network t includes minimizing a total loss calculated based on the biased loss associated with the first neural network and an unbiased loss associated with the second neural network.
AI. The non-transitory computer-readable storage medium as paragraph AH recites, wherein minimizing the loss calculated based on the loss function by utilizing the first neural network and the second neural network includes solving for a derivative of the loss function equaling zero.
AJ. The non-transitory computer-readable storage medium as paragraph AH recites, wherein minimizing the loss calculated based on the loss function by utilizing the trained first neural network and the trained second neural network includes solving for a derivative of the loss function being less than or equal to a threshold value.
CONCLUSION
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described. Rather, the specific features and acts are disclosed as exemplary forms of implementing the claims.

Claims (36)

  1. A method comprising:
    receiving input feature parameters of a plurality of components associated with at least one computer room;
    training a first neural network and a second neural network based on the input feature parameters; and
    predicting a power usage effectiveness (PUE) of the at least one computer room based on an output of the first neural network and an output of the second neural network.
  2. The method of claim 1, wherein the first neural network is an unsupervised neural network and the second network is a supervised prediction neural network.
  3. The method of claim 1, wherein training the first neural network and the second neural network based on the input feature parameters includes simultaneously training the first neural network and the second neural network based on the input feature parameters.
  4. The method of claim 1, wherein receiving the input feature parameters of the plurality of components associated with the at least one computer room includes:
    creating an ontology having a plurality of levels associated with the plurality of components; and
    receiving information of the plurality of components based on the on the ontology including corresponding associated concept, historical data, locations, physical connections, and hierarchy among the plurality of components.
  5. The method of claim 4, wherein the relationships among the plurality of components are based, at least in part, on loading of computing equipment in the computer room.
  6. The method of claim 5, wherein the loading of the computing equipment includes a workload of the computing equipment and an electrical load used by the computing equipment.
  7. The method of claim 6, wherein the computing equipment includes a server and a power supply for the server.
  8. The method of claim 4, wherein training the first neural network and the P-Net based on the input feature parameters includes using a gradient descent algorithm to implement learning of the input feature parameters for corresponding concepts.
  9. The method of claim 1, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes biasing a loss associated with the first neural network with a constant value.
  10. The method of claim 9, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes minimizing a total loss calculated based on the biased loss associated with the first neural network and an unbiased loss associated with the second neural network.
  11. The method of claim 10, wherein minimizing the loss calculated based on the loss function by utilizing the first neural network and the second neural network includes solving for a derivative of the loss function equaling zero.
  12. The method of claim 10, wherein minimizing the loss calculated based on the loss function by utilizing the trained first neural network and the trained second neural network includes solving for a derivative of the loss function being less than or equal to a threshold value.
  13. A system comprising:
    one or more processors; and
    memory communicatively coupled to the one or more processors, the memory storing computer-readable instructions executable by one or more processors, that when executed by the one or more processors, cause the one or more processors to perform operations comprising:
    receiving input feature parameters of a plurality of components associated with at least one computer room;
    training a first neural network and a second neural network based on the input feature parameters; and
    predicting the power usage effectiveness (PUE) of the at least one computer room based on an output of the first neural network and an output of the second neural network.
  14. The system of claim 13, wherein the first neural network is an unsupervised neural network and the second network is a supervised prediction neural network.
  15. The system of claim 13, wherein training the first neural network and the second neural network based on the input feature parameters includes simultaneously training the first neural network and the second neural network based on the input feature parameters.
  16. The system of claim 13, wherein receiving the input feature parameters of the plurality of components associated with the at least one computer room includes:
    creating an ontology having a plurality of levels associated with the plurality of components; and
    receiving information of the plurality of components based on the on the ontology including corresponding associated concept, historical data, locations, physical connections, and hierarchy among the plurality of components.
  17. The system of claim 16, wherein the relationships among the plurality of components are based, at least in part, on loading of the computing equipment.
  18. The system of claim 17, wherein the loading of the computing equipment includes a workload of the computing equipment and an electrical load used by the computing equipment.
  19. The system of claim 18, wherein the computing equipment includes a server and a power supply for the server.
  20. The system of claim 16, wherein training the first neural network and the second neural network based on the input feature parameters includes  using a gradient descent algorithm to implement learning of the input feature parameters for corresponding concepts.
  21. The system of claim 13, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes biasing a loss associated with the first neural network with a constant value.
  22. The system of claim 21, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes minimizing a total loss calculated based on the biased loss associated with the first neural network and an unbiased loss associated with the second neural network.
  23. The system of claim 22, wherein minimizing the loss calculated based on the loss function by utilizing the first neural network and the second neural network includes solving for a derivative of the loss function equaling zero.
  24. The system of claim 22, wherein minimizing the loss calculated based on the loss function by utilizing the trained first neural network and the trained second neural network includes solving for a derivative of the loss function being less than or equal to a threshold value.
  25. A non-transitory computer-readable storage medium storing computer-readable instructions executable by one or more processors, that when executed by the one or more processors, cause the one or more processors to perform operations comprising:
    receiving input feature parameters of a plurality of components associated with at least one computer room;
    training a first neural network and a second neural network based on the input feature parameters; and
    predicting the power usage effectiveness (PUE) of the at least one computer room based on an output of the first neural network and an output of the second neural network.
  26. The non-transitory computer-readable storage medium of claim 25, wherein the first neural network is an unsupervised neural network and the second network is a supervised prediction neural network.
  27. The non-transitory computer-readable storage medium of claim 25, wherein training the first neural network and the second neural network based on the input feature parameters includes simultaneously training the first neural network and the second neural network based on the input feature parameters.
  28. The non-transitory computer-readable storage medium of claim 25, wherein receiving the input feature parameters of the plurality of components associated with the at least one computer room includes:
    creating an ontology having a plurality of levels associated with the plurality of components; and
    receiving information of the plurality of components based on the on the ontology including corresponding associated concept, historical data, locations, physical connections, and hierarchy among the plurality of components.
  29. The non-transitory computer-readable storage medium of claim 28, wherein the relationships among the plurality of components are based, at least in part, on loading of computing equipment in the computer room.
  30. The non-transitory computer-readable storage medium of claim 29, wherein the loading of the computing equipment includes a workload of the computing equipment and an electrical load used by the computing equipment.
  31. The non-transitory computer-readable storage medium of claim 30, wherein the computing equipment includes a server and a power supply for the server.
  32. The non-transitory computer-readable storage medium of claim 28, wherein training the first neural network and the second neural network based  on the input feature parameters includes using a gradient descent algorithm to implement learning of the input feature parameters for corresponding concepts.
  33. The non-transitory computer-readable storage medium of claim 25, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes biasing a loss associated with the first neural network with a constant value.
  34. The non-transitory computer-readable storage medium of claim 33, wherein predicting the PUE of the at least one computer room based on the output of the first neural network and the output of the second neural network includes minimizing a total loss calculated based on the biased loss associated with the first neural network and an unbiased loss associated with the second neural network.
  35. The non-transitory computer-readable storage medium of claim 34, wherein minimizing the loss calculated based on the loss function by utilizing the first neural network and the second neural network includes solving for a derivative of the loss function equaling zero.
  36. The non-transitory computer-readable storage medium of claim 34, wherein minimizing the loss calculated based on the loss function by utilizing  the trained first neural network and the trained second neural network includes solving for a derivative of the loss function being less than or equal to a threshold value.
PCT/CN2019/087083 2019-05-15 2019-05-15 Hybrid-learning neural network architecture WO2020227983A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2019/087083 WO2020227983A1 (en) 2019-05-15 2019-05-15 Hybrid-learning neural network architecture
CN201980093428.6A CN113518962A (en) 2019-05-15 2019-05-15 Hybrid learning neural network architecture

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2019/087083 WO2020227983A1 (en) 2019-05-15 2019-05-15 Hybrid-learning neural network architecture

Publications (1)

Publication Number Publication Date
WO2020227983A1 true WO2020227983A1 (en) 2020-11-19

Family

ID=73290100

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/087083 WO2020227983A1 (en) 2019-05-15 2019-05-15 Hybrid-learning neural network architecture

Country Status (2)

Country Link
CN (1) CN113518962A (en)
WO (1) WO2020227983A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113465139A (en) * 2021-05-28 2021-10-01 山东英信计算机技术有限公司 Refrigeration optimization method, system, storage medium and equipment
CN115907202A (en) * 2022-12-13 2023-04-04 中国通信建设集团设计院有限公司 Data center PUE calculation analysis method and system under double-carbon background

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113721151B (en) * 2021-11-03 2022-02-08 杭州宇谷科技有限公司 Battery capacity estimation model and method based on double-tower deep learning network

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090201293A1 (en) * 2008-02-12 2009-08-13 Accenture Global Services Gmbh System for providing strategies for increasing efficiency of data centers
US20100076607A1 (en) * 2008-08-08 2010-03-25 Osman Ahmed Data center thermal performance optimization using distributed cooling systems
CN103645795A (en) * 2013-12-13 2014-03-19 浪潮电子信息产业股份有限公司 Cloud computing data center energy saving method based on ANN (artificial neural network)

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109670623A (en) * 2017-10-16 2019-04-23 优酷网络技术(北京)有限公司 Neural net prediction method and device
CN109002942A (en) * 2018-09-28 2018-12-14 河南理工大学 A kind of short-term load forecasting method based on stochastic neural net

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090201293A1 (en) * 2008-02-12 2009-08-13 Accenture Global Services Gmbh System for providing strategies for increasing efficiency of data centers
US20100076607A1 (en) * 2008-08-08 2010-03-25 Osman Ahmed Data center thermal performance optimization using distributed cooling systems
CN103645795A (en) * 2013-12-13 2014-03-19 浪潮电子信息产业股份有限公司 Cloud computing data center energy saving method based on ANN (artificial neural network)

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
LIU, WEI ET AL.: "Research of Mutual Learning Neural Network Training Method", CHINESE JOURNAL OF COMPUTERS, vol. 40, no. 6, 30 June 2017 (2017-06-30), XP055752928, ISSN: 0254-4164 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113465139A (en) * 2021-05-28 2021-10-01 山东英信计算机技术有限公司 Refrigeration optimization method, system, storage medium and equipment
CN113465139B (en) * 2021-05-28 2022-11-08 山东英信计算机技术有限公司 Refrigeration optimization method, system, storage medium and equipment
CN115907202A (en) * 2022-12-13 2023-04-04 中国通信建设集团设计院有限公司 Data center PUE calculation analysis method and system under double-carbon background
CN115907202B (en) * 2022-12-13 2023-10-24 中国通信建设集团设计院有限公司 Data center PUE (physical distribution element) calculation analysis method and system under double-carbon background

Also Published As

Publication number Publication date
CN113518962A (en) 2021-10-19

Similar Documents

Publication Publication Date Title
Li et al. Intelligent multi-zone residential HVAC control strategy based on deep reinforcement learning
Wei et al. Multi-objective optimization of the HVAC (heating, ventilation, and air conditioning) system performance
Dash et al. Efficient stock price prediction using a self evolving recurrent neuro-fuzzy inference system optimized through a modified differential harmony search technique
Fallah-Mehdipour et al. Real-time operation of reservoir system by genetic programming
WO2020227983A1 (en) Hybrid-learning neural network architecture
CN110866592B (en) Model training method, device, energy efficiency prediction method, device and storage medium
EP3835895A1 (en) Multi-agent deep reinforcement learning for dynamically controlling electrical equipment in buildings
Wahid et al. An efficient approach for energy consumption optimization and management in residential building using artificial bee colony and fuzzy logic
CN110826784B (en) Method and device for predicting energy use efficiency, storage medium and terminal equipment
US11835928B2 (en) Adaptive mixed integer nonlinear programming for process management
Lissa et al. Transfer learning applied to reinforcement learning-based hvac control
Behrang et al. Assessment of electricity demand in Iran's industrial sector using different intelligent optimization techniques
Cheng et al. Forecasting monthly runoff time series by single-layer feedforward artificial neural network and grey wolf optimizer
Cao et al. PSO-Stacking improved ensemble model for campus building energy consumption forecasting based on priority feature selection
WO2019227273A1 (en) Hierarchical concept based neural network model for data center power usage effectiveness prediction
Gruber et al. Efficient Bayesian network learning for system optimization in reliability engineering
Cho et al. Rule reduction for control of a building cooling system using explainable AI
CN112330044A (en) Support vector regression model based on iterative aggregation grid search algorithm
CN116717839A (en) Heating control method, control device and heating system
CN116954329A (en) Method, device, equipment, medium and program product for regulating state of refrigeration system
CN112234599B (en) Advanced dynamic self-adaptive partitioning method and system for multi-element complex urban power grid
Guo et al. Fruit Fly Optimization Algorithm Based on Single‐Gene Mutation for High‐Dimensional Unconstrained Optimization Problems
Genkin et al. A transfer learning approach to minimize reinforcement learning risks in energy optimization for automated and smart buildings
Yu et al. A Combined Neural and Genetic Algorithm Model for Data Center Temperature Control.
Chen et al. Exploiting more robust and efficacious deep learning techniques for modeling wind power with speed

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19928522

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19928522

Country of ref document: EP

Kind code of ref document: A1