US20180241631A1 - Systems and methods for network topology validation - Google Patents

Systems and methods for network topology validation Download PDF

Info

Publication number
US20180241631A1
US20180241631A1 US15/440,932 US201715440932A US2018241631A1 US 20180241631 A1 US20180241631 A1 US 20180241631A1 US 201715440932 A US201715440932 A US 201715440932A US 2018241631 A1 US2018241631 A1 US 2018241631A1
Authority
US
United States
Prior art keywords
chassis
information handling
topology
switch
validation rules
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US15/440,932
Other versions
US10785118B2 (en
Inventor
Sudhir Vittal Shetty
Sanjay Sane
Balaji Rajagopalan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dell Products LP
Original Assignee
Dell Products LP
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Assigned to DELL PRODUCTS L.P. reassignment DELL PRODUCTS L.P. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: RAJAGOPALAN, BALAJI, SHETTY, SUDHIR VITTAL, SANE, SANJAY
Priority to US15/440,932 priority Critical patent/US10785118B2/en
Application filed by Dell Products LP filed Critical Dell Products LP
Assigned to THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., AS COLLATERAL AGENT reassignment THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., AS COLLATERAL AGENT PATENT SECURITY INTEREST (NOTES) Assignors: DELL PRODUCTS L.P., EMC CORPORATION, EMC IP Holding Company LLC, MOZY, INC., WYSE TECHNOLOGY L.L.C.
Assigned to CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENT reassignment CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENT PATENT SECURITY INTEREST (CREDIT) Assignors: DELL PRODUCTS L.P., EMC CORPORATION, EMC IP Holding Company LLC, MOZY, INC., WYSE TECHNOLOGY L.L.C.
Publication of US20180241631A1 publication Critical patent/US20180241631A1/en
Assigned to THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A. reassignment THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A. SECURITY AGREEMENT Assignors: CREDANT TECHNOLOGIES, INC., DELL INTERNATIONAL L.L.C., DELL MARKETING L.P., DELL PRODUCTS L.P., DELL USA L.P., EMC CORPORATION, EMC IP Holding Company LLC, FORCE10 NETWORKS, INC., WYSE TECHNOLOGY L.L.C.
Assigned to THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A. reassignment THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A. SECURITY AGREEMENT Assignors: CREDANT TECHNOLOGIES INC., DELL INTERNATIONAL L.L.C., DELL MARKETING L.P., DELL PRODUCTS L.P., DELL USA L.P., EMC CORPORATION, EMC IP Holding Company LLC, FORCE10 NETWORKS, INC., WYSE TECHNOLOGY L.L.C.
Publication of US10785118B2 publication Critical patent/US10785118B2/en
Application granted granted Critical
Assigned to DELL PRODUCTS L.P., EMC CORPORATION, WYSE TECHNOLOGY L.L.C., MOZY, INC., EMC IP Holding Company LLC reassignment DELL PRODUCTS L.P. RELEASE OF SECURITY INTEREST AT REEL 042768 FRAME 0585 Assignors: CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH
Assigned to EMC IP HOLDING COMPANY LLC (ON BEHALF OF ITSELF AND AS SUCCESSOR-IN-INTEREST TO MOZY, INC.), DELL PRODUCTS L.P., DELL MARKETING CORPORATION (SUCCESSOR-IN-INTEREST TO WYSE TECHNOLOGY L.L.C.), EMC CORPORATION reassignment EMC IP HOLDING COMPANY LLC (ON BEHALF OF ITSELF AND AS SUCCESSOR-IN-INTEREST TO MOZY, INC.) RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL/FRAME (042769/0001) Assignors: THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., AS NOTES COLLATERAL AGENT
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/12Discovery or management of network topologies
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/22Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks comprising specially adapted graphical user interfaces [GUI]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L49/00Packet switching elements
    • H04L49/25Routing or path finding in a switch fabric
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/34Signalling channels for network management communication
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/34Signalling channels for network management communication
    • H04L41/344Out-of-band transfers
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks

Definitions

  • the present disclosure relates in general to information handling systems, and more particularly to providing for network topology validation in a system comprising multiple information handling system chassis.
  • An information handling system generally processes, compiles, stores, and/or communicates information or data for business, personal, or other purposes thereby allowing users to take advantage of the value of the information.
  • information handling systems may also vary regarding what information is handled, how the information is handled, how much information is processed, stored, or communicated, and how quickly and efficiently the information may be processed, stored, or communicated.
  • the variations in information handling systems allow for information handling systems to be general or configured for a specific user or specific use such as financial transaction processing, airline reservations, enterprise data storage, or global communications.
  • information handling systems may include a variety of hardware and software components that may be configured to process, store, and communicate information and may include one or more computer systems, data storage systems, and networking systems.
  • the various chassis may be of different types and may be cabled and wired together in a particular manner.
  • a console for managing the system may need to determine the connectivity among the various chassis and components internal to the chassis in order to determine a topology of the system.
  • an information handling system may include a processor and a memory coupled to the processor, the memory having program instructions stored thereon that, upon execution by the processor, cause the processor to determine a topology of connectivity of various components of a system comprising multiple information handling system chassis and apply validation rules to the topology to validate the topology.
  • a method may include determining a topology of connectivity of various components of a system comprising multiple information handling system chassis and applying validation rules to the topology to validate the topology.
  • an article of manufacture may include a non-transitory computer readable medium and computer-executable instructions carried on the computer readable medium, the instructions readable by a processor, the instructions, when read and executed, for causing the processor to determine a topology of connectivity of various components of a system comprising multiple information handling system chassis apply validation rules to the topology to validate the topology.
  • FIG. 1 illustrates a block diagram of a system comprising multiple information handling system chassis, in accordance with embodiments of the present disclosure
  • FIG. 2 illustrates a flow chart of an example method for network topology discovery, in accordance with embodiments of the present disclosure
  • FIG. 3 illustrates a flow chart of an example method for network topology validation, in accordance with embodiments of the present disclosure.
  • FIG. 4 illustrates a block diagram of an information handling system, in accordance with embodiments of the present disclosure.
  • FIGS. 1 through 4 Preferred embodiments and their advantages are best understood by reference to FIGS. 1 through 4 , wherein like numbers are used to indicate like and corresponding parts.
  • an information handling system may include any instrumentality or aggregate of instrumentalities operable to compute, classify, process, transmit, receive, retrieve, originate, switch, store, display, manifest, detect, record, reproduce, handle, or utilize any form of information, intelligence, or data for business, scientific, control, entertainment, or other purposes.
  • an information handling system may be a personal computer, a personal digital assistant (PDA), a consumer electronic device, a network storage device, or any other suitable device and may vary in size, shape, performance, functionality, and price.
  • the information handling system may include memory, one or more processing resources such as a central processing unit (“CPU”) or hardware or software control logic.
  • Additional components of the information handling system may include one or more storage devices, one or more communications ports for communicating with external devices as well as various I/O devices, such as a keyboard, a mouse, and a video display.
  • the information handling system may also include one or more buses operable to transmit communication between the various hardware components.
  • information handling resources may broadly refer to any component system, device or apparatus of an information handling system, including without limitation processors, buses, memories, I/O devices and/or interfaces, storage resources, network interfaces, motherboards, electro-mechanical devices (e.g., fans), displays, and power supplies.
  • Computer-readable media may include any instrumentality or aggregation of instrumentalities that may retain data and/or instructions for a period of time.
  • Computer-readable media may include, without limitation, storage media such as a direct access storage device (e.g., a hard disk drive or floppy disk), a sequential access storage device (e.g., a tape disk drive), compact disk, CD-ROM, DVD, random access memory (“RAM”), read-only memory (“ROM”), electrically erasable programmable read-only memory (“EEPROM”), and/or flash memory; as well as communications media such as wires, optical fibers, microwaves, radio waves, and other electromagnetic and/or optical carriers; and/or any combination of the foregoing.
  • direct access storage device e.g., a hard disk drive or floppy disk
  • sequential access storage device e.g., a tape disk drive
  • compact disk CD-ROM, DVD, random access memory (“RAM”)
  • ROM read-only memory
  • EEPROM electrically erasable programmable
  • FIG. 1 illustrates a block diagram of a system 100 comprising multiple information handling system chassis 101 , in accordance with embodiments of the present disclosure.
  • system 100 may comprise a plurality of chassis 101 (e.g., chassis 101 a , 101 b , and 101 c ), each chassis 101 including information handling systems and/or information handling resources, a private inter-chassis network 150 , and a console 120 .
  • chassis 101 a may include a switch 106 a and a chassis management controller 112 .
  • chassis 101 b may include a switch 106 b and a chassis management controller 112 .
  • chassis 101 c may include switches 106 c and 106 d , one or more servers 102 , and a chassis management controller 112 .
  • the various servers 102 and switches 106 may each comprise ports 110 for interfaces with one another, with example connectivity among ports 110 depicted in FIG. 1 .
  • chassis 101 a and 101 b are shown without servers (e.g., servers 102 ).
  • one or both of chassis 101 a and 101 b may include one or more servers 102 .
  • a server 102 may generally be operable to receive data from and/or communicate data to one or more information handling resources of chassis 101 c .
  • a server 102 may comprise a blade server having modular physical design.
  • server 102 may comprise an M class server.
  • Each server 102 may include a host management controller 104 .
  • Host management controller 104 may be implemented by, for example, a microprocessor, microcontroller, DSP, ASIC, EEPROM, or any combination thereof.
  • Host management controller 104 may be configured to communicate with chassis management controller 112 . Such communication may be made, for example, via a private management network fabric integral to server 102 (not explicitly shown).
  • Host management controller 104 may be configured to provide out-of-band management facilities for management of system 100 . Such management may be made by information handling resources of chassis 101 even if system 100 is powered off or powered to a standby state.
  • Host management controller 104 may include a processor, memory, and network connection separate from the rest of system 100 .
  • host management controller 104 may include or may be an integral part of a baseboard management controller (BMC) or an Integrated Dell Remote Access Controller (iDRAC).
  • BMC baseboard management controller
  • iDRAC Integrated Dell Remote Access Controller
  • a switch 106 may comprise any suitable system, device, or apparatus for receiving, processing, and forwarding packets.
  • each of switch 106 a and 106 b may serve as an interface between chassis 101 c and a data network and may process and route packets between chassis 101 c and such data network.
  • switches 106 c and 106 d may each serve as an interface between servers 102 and other chassis 101 (e.g., 101 a and 101 b ) external to chassis 101 c.
  • Each port 110 may comprise a connector, slot, or another interface for receiving one end of a physical connection (e.g., wire, cable) coupled to a corresponding port 110 of another information handling resource.
  • a physical connection e.g., wire, cable
  • a chassis management controller 112 may comprise any system, device, or apparatus configured to facilitate management and/or control of system 100 embodied by chassis 101 , its information handling systems 102 , and/or one or more of its component information handling resources. Chassis management controller 112 may be configured to issue commands and/or other signals to manage and/or control an information handling system 102 and/or information handling resources of system 100 . Chassis management controller 112 may comprise a microprocessor, microcontroller, DSP, ASIC, field programmable gate array (“FPGA”), EEPROM, or any combination thereof. In some embodiments, chassis management controller 112 may provide a management console for user/administrator access to these functions.
  • chassis management controller 112 may implement Representational State Transfer (“REST”) or another suitable management protocol permitting a user to remotely access chassis management controller 112 to configure system 100 and its various information handling resources.
  • chassis management controller 112 may interface with a network interface, thus allowing for “out-of-band” control of chassis 101 , such that communications to and from chassis management controller 112 are communicated via a management channel physically isolated from an “in-band” communication channel of chassis 101 for which non-management communication may take place.
  • REST Representational State Transfer
  • chassis management controller 112 may interface with a network interface, thus allowing for “out-of-band” control of chassis 101 , such that communications to and from chassis management controller 112 are communicated via a management channel physically isolated from an “in-band” communication channel of chassis 101 for which non-management communication may take place.
  • chassis management controller 112 may allow an administrator to remotely manage one or more parameters associated with operation of chassis 101 and its various information handling resources (e.g., power usage, processor allocation, memory allocation, security privileges, etc.).
  • chassis management controller 112 may include a management services module.
  • one of the chassis management controllers 112 of the various chassis 101 may be selected as a “lead” chassis management controller 112 , such that an administrator may manage the entirety of system 100 by interfacing with a single chassis management controller 112 .
  • Private inter-chassis network 150 may comprise a network and/or fabric configured to couple information chassis management controllers 112 of the various chassis 101 in system 100 to each other.
  • inter-chassis network 150 may include a communication infrastructure, which provides physical connections, and a management layer, which organizes the physical connections and chassis management controllers 112 communicatively coupled to private inter-chassis network 150 .
  • Private inter-chassis network 150 may be implemented as, or may be a part of, an Ethernet local area network (LAN) or any other appropriate architecture or system that facilitates the communication of signals, data, and/or messages.
  • LAN Ethernet local area network
  • Console 120 may be communicatively coupled to private inter-chassis network 150 and may comprise an information handling system or a program executable on an information handling system for monitoring and management of the various chassis 101 of system 100 and their various components. For example, as described in greater detail below, console 120 may be able to, based on messages communicated among chassis 101 and management controllers 112 , determine the topology of connectivity among the various components of system 100 . In some embodiments, console 120 may also be capable of rendering a graphical representation of the topology of connectivity to a user via a user interface (not explicitly shown) of console 120 . Although FIG. 1 depicts console 120 as a stand-alone component of system 100 , in some embodiments, console 120 may be integral to or otherwise embodied in a chassis management controller 112 .
  • console 120 may leverage payload information of packets (e.g., Link Layer Discovery Protocol or “LLDP” packets) communicated among switches 106 and/or servers 102 in response to a switch 106 or server 102 being coupled to another device in system 100 and may also leverage advertisements (e.g., multicast Domain Name Service or mDNS) communicated via private inter-chassis network 150 in order to extract information from such payloads and advertisements and process such information to determine topology of connectivity of devices of system 100 .
  • packets e.g., Link Layer Discovery Protocol or “LLDP” packets
  • advertisements e.g., multicast Domain Name Service or mDNS
  • a host management controller 104 of such server may communicate LLDP packets which may be received not only by switches 106 c and 106 d of the chassis 101 c comprising the server 102 , but may be also passed-through from switches 106 c and 106 d to switches 106 a and 106 b of chassis 101 a and 101 b , respectively.
  • the payload of each such LLDP packet may include various identifying information, including without limitation identifying information (e.g., service tag or serial number) of the server 102 from which the LLDP packet originated, identifying information (e.g., service tag or serial number) of the chassis 101 comprising the server 102 , identifying information (e.g., fully-qualified device descriptor) of the port 110 of the server 102 originating the LLDP packet, and the switch fabric (e.g., 106 c or 106 d ) to which the packet-originating port 110 is coupled.
  • identifying information e.g., service tag or serial number
  • the switch fabric e.g., 106 c or 106 d
  • a console 120 may be able to determine the chassis 101 housing the server 102 , a switch fabric to which the server 102 is coupled, and connectivity between the server 102 and a switch 106 receiving the LLDP packet external to the chassis 101 .
  • chassis management controllers 112 may communicate mDNS advertisements on private inter-chassis network 150 .
  • Such advertisements may include various identifying information, including without limitation identifying information (e.g., service tag or serial number) of a chassis 101 including a switch fabric, identifying information (e.g., slot number) of a switch 106 of the switch fabric, and a switch type for the switch 106 .
  • console 120 may correlate such information from mDNS advertisements to the information from LLDP payloads to determine the switch type of each switch fabric.
  • a switch 106 may include metadata information (e.g., an I/O module or “IOM” descriptor file) that is specific to a switch type and uplink (e.g., connectivity between a server and switch 106 c / 106 d ) and downlink connectivity (connectivity between a switch 106 c / 106 d and a switch 106 a / 106 b ) of a switch 106 .
  • console 120 may correlate such IOM metadata information to information from mDNS advertisements and/or information from LLDP payloads to determine the complete topology of each switch fabric.
  • Inter-switch connectivity e.g., between switches 106 a and 106 b
  • external network connectivity e.g., between an external data network and switches 106 a / 106 b
  • Inter-switch connectivity may be determined based on LLDP payloads communicated between switches (e.g., between switches 106 a and 106 b ) and between individual switches (e.g., between switches 106 a and 106 b ) and the external data network.
  • FIG. 2 illustrates a flow chart of an example method 200 for network topology discovery, in accordance with embodiments of the present disclosure.
  • method 200 may begin at step 202 .
  • teachings of the present disclosure may be implemented in a variety of configurations of system 100 . As such, the preferred initialization point for method 200 and the order of the steps comprising method 200 may depend on the implementation chosen.
  • console 120 may receive an LLDP packet, mDNS advertisement, or IOM metadata in response to a physical connection being made in system 100 .
  • console 120 may process information present in the LLDP packet, mDNS advertisement, or IOM metadata and along with other information previously received from other LLDP packets, mDNS advertisements, or IOM metadata.
  • console 120 may update a table, list, map, database, or other data structure defining the topology of connectivity of the various components of system 100 .
  • console 120 may generate and output to a user interface a graphical display representative of the topology.
  • console 120 receives various traffic
  • console 120 may not receive all of such traffic (e.g., console 120 may not receive LLDP packets). Instead, a switch 106 may receive LLDP packets and determine topology based on such packets. Thus, a complete topology of a system may be determined by the aggregate of traffic received by switches 106 and console 120 .
  • FIG. 2 discloses a particular number of steps to be taken with respect to method 200
  • method 200 may be executed with greater or lesser steps than those depicted in FIG. 2 .
  • FIG. 2 discloses a certain order of steps to be taken with respect to method 200
  • the steps comprising method 200 may be completed in any suitable order.
  • Method 200 may be implemented using system 100 , and/or any other system operable to implement method 200 .
  • method 200 may be implemented partially or fully in software and/or firmware embodied in computer-readable media.
  • console 120 may also be configured to validate the topology against a set of validation rules.
  • a multi-chassis group such as system 100
  • the various I/O modules must typically be wired in prescriptive ways.
  • console 120 provide guidance to a user regarding any incorrect wiring in system 100 .
  • a topology could be incorrectly wired for numerous reasons. For example, a cable may be missing or not properly connected between its source and destination.
  • validation rules may be complex rules that go beyond simply matching particular source ports and destination ports. Validation rules may be expressed on characteristics of I/O modules/switches 106 (e.g., slot location, mismatch of switch types in a chassis 101 ) or aggregate characteristics of chassis 101 in a group (e.g., group membership of chassis 101 ).
  • Inter-chassis switch wiring such a rule may be expressed in terms of a matching of a slot location for switches 106 disposed in different chassis. For example, if a first switch 106 is in a slot “A” of a first chassis 101 and is coupled to a second switch 106 of a second chassis 101 , the rule may ensure that second switch 106 is in the slot “A” of the second chassis.
  • Group membership such a rule may be expressed in terms of a match in chassis group membership between a device (e.g., switch 106 ) having a source port 110 and another device (e.g., switch 106 ) having a destination port 110 of a connection.
  • a device e.g., switch 106
  • another device e.g., switch 106
  • Switch types such a rule may be expressed in terms of the switch type in an appropriate fabric location. For example, if two slots of a chassis 101 are populated with switches 106 , this rule may ensure that both switches 106 are of the same switch type.
  • FIG. 3 illustrates a flow chart of an example method 300 for network topology validation, in accordance with embodiments of the present disclosure.
  • method 300 may begin at step 302 .
  • teachings of the present disclosure may be implemented in a variety of configurations of system 100 . As such, the preferred initialization point for method 300 and the order of the steps comprising method 300 may depend on the implementation chosen.
  • console 120 may determine the topology of system 100 .
  • such topology may be constructed in accordance with method 200 described above.
  • console 120 may apply a set of validation rules (e.g., one or more of the various rules discussed above) to the topology to determine if the topology satisfies the validation rules. If the topology satisfies the topology rules, method 300 may proceed again to step 302 , and the topology may be continuously validated as the topology is changed. If the topology violates one or more of the topology rules, method 300 may proceed to step 306 .
  • a set of validation rules e.g., one or more of the various rules discussed above
  • console 120 may generate and output to a user interface an indication of which of the topology rules have been violated.
  • indication may be in terms of a list or a graphical representation of the topology with indications within the graphical representation of the portion of the topology that are in violation of the topology rules.
  • FIG. 3 discloses a particular number of steps to be taken with respect to method 300
  • method 300 may be executed with greater or lesser steps than those depicted in FIG. 3 .
  • FIG. 3 discloses a certain order of steps to be taken with respect to method 300
  • the steps comprising method 300 may be completed in any suitable order.
  • Method 300 may be implemented using system 100 , and/or any other system operable to implement method 300 .
  • method 300 may be implemented partially or fully in software and/or firmware embodied in computer-readable media.
  • FIG. 4 illustrates a block diagram of an information handling system 400 , in accordance with embodiments of the present disclosure.
  • Information handling system 400 is an instance of console 120 and/or a server 102 illustrated in FIG. 1 .
  • information handling system 400 includes one or more CPUs 402 .
  • information handling system 400 may be a single-processor system including one CPU 402 , or a multi-processor system including two or more CPUs 402 (e.g., two, four, eight, or any other suitable number).
  • CPU(s) 402 may include any processor capable of executing program instructions.
  • CPU(s) 402 may be general-purpose or embedded processors implementing any of a variety of instruction set architectures (ISAs), such as the x86, POWERPC, ARM, SPARC, or MIPS ISAs, or any other suitable ISA. In multi-processor systems, each of CPU(s) 402 may commonly, but not necessarily, implement the same ISA.
  • a motherboard may be configured to provide structural support, power, and electrical connectivity between the various aforementioned components. Such a motherboard may include multiple connector sockets in various configurations, adapted to receive pluggable circuit cards, component chip packages, etc.
  • Northbridge controller 404 may be configured to coordinate I/O traffic between CPU(s) 402 and other components.
  • northbridge controller 404 is coupled to graphics device(s) 408 (e.g., one or more video cards or adaptors, etc.) via graphics bus 410 (e.g., an Accelerated Graphics Port or AGP bus, a Peripheral Component Interconnect or PCI bus, etc.).
  • Northbridge controller 404 is also coupled to system memory 412 via memory bus 414 .
  • Memory 412 may be configured to store program instructions and/or data accessible by CPU(s) 402 .
  • memory 412 may be implemented using any suitable memory technology, such as static RAM (SRAM), synchronous dynamic RAM (SDRAM), nonvolatile/Flash-type memory, or any other type of memory.
  • Northbridge controller 404 is coupled to southbridge controller or chipset 416 via internal bus 418 .
  • southbridge controller 416 may be configured to handle various of computing device information handling system 400 's I/O operations, and it may provide interfaces such as, for instance, Universal Serial Bus (USB), audio, serial, parallel, Ethernet, etc., via port(s), pin(s), and/or adapter(s) 432 over bus 434 .
  • southbridge controller 416 may be configured to allow data to be exchanged between information handling system 400 and other devices, such as other information handling systems attached to a network.
  • southbridge controller 416 may support communication via wired or wireless general data networks, such as any suitable type of Ethernet network, for example; via telecommunications/telephony networks such as analog voice networks or digital fiber communications networks; via storage area networks such as Fiber Channel SANs; or via any other suitable type of network and/or protocol.
  • general data networks such as any suitable type of Ethernet network, for example; via telecommunications/telephony networks such as analog voice networks or digital fiber communications networks; via storage area networks such as Fiber Channel SANs; or via any other suitable type of network and/or protocol.
  • Southbridge controller 416 may also enable connection to one or more keyboards, keypads, touch screens, scanning devices, voice or optical recognition devices, or any other devices suitable for entering or retrieving data. Multiple I/O devices may be present in information handling system 400 . In some embodiments, I/O devices may be separate from information handling system 400 and may interact with information handling system 400 through a wired or wireless connection. As shown, southbridge controller 416 is further coupled to one or more PCI devices 420 (e.g., modems, network cards, sound cards, video cards, etc.) via PCI bus 422 . Southbridge controller 416 is also coupled to Basic I/O System (BIOS) 424 , Super I/O Controller 426 , and Baseboard Management Controller (BMC) 428 via Low Pin Count (LPC) bus 430 .
  • BIOS Basic I/O System
  • BMC Baseboard Management Controller
  • BIOS 424 includes non-volatile memory having program instructions stored thereon. Those instructions may be usable for CPU(s) 402 to initialize and test other hardware components and/or to load an Operating System (OS) onto information handling system 400 . As such, BIOS 424 may include a firmware interface that allows CPU(s) 402 to load and execute certain firmware, as described in more detail below. In some cases, such firmware may include program code that is compatible with the Unified Extensible Firmware Interface (UEFI) specification, although other types of firmware may be used.
  • UEFI Unified Extensible Firmware Interface
  • BMC controller 428 may include non-volatile memory having program instructions stored thereon that are usable by CPU(s) 402 to enable remote management of information handling system 400 .
  • BMC controller 428 may enable a user to discover, configure, and manage BMC controller 428 , setup configuration options, resolve and administer hardware or software problems, etc.
  • BMC controller 428 may include one or more firmware volumes, each volume having one or more firmware files used by the BIOS' firmware interface to initialize and test components of information handling system 400 .
  • Super I/O Controller 426 combines interfaces for a variety of lower bandwidth or low data rate devices. Those devices may include, for example, floppy disks, parallel ports, keyboard and mouse, temperature sensor and fan speed monitoring, etc.
  • the super I/O controller 426 may be coupled to the one or more upstream sensors and to the one or more downstream sensors.
  • information handling system 400 may be configured to access different types of computer-accessible media separate from memory 412 .
  • a computer-accessible medium may include any tangible, non-transitory storage media or memory media such as electronic, magnetic, or optical media—e.g., magnetic disk, a hard drive, a CD/DVD-ROM, a Flash memory, etc. coupled to information handling system 400 via northbridge controller 404 and/or southbridge controller 416 .
  • references in the appended claims to an apparatus or system or a component of an apparatus or system being adapted to, arranged to, capable of, configured to, enabled to, operable to, or operative to perform a particular function encompasses that apparatus, system, or component, whether or not it or that particular function is activated, turned on, or unlocked, as long as that apparatus, system, or component is so adapted, arranged, capable, configured, enabled, operable, or operative.

Abstract

An information handling system may include a processor and a memory coupled to the processor, the memory having program instructions stored thereon that, upon execution by the processor, cause the processor to determine a topology of connectivity of various components of a system comprising multiple information handling system chassis and apply validation rules to the topology to validate the topology.

Description

    TECHNICAL FIELD
  • The present disclosure relates in general to information handling systems, and more particularly to providing for network topology validation in a system comprising multiple information handling system chassis.
  • BACKGROUND
  • As the value and use of information continues to increase, individuals and businesses seek additional ways to process and store information. One option available to users is information handling systems. An information handling system generally processes, compiles, stores, and/or communicates information or data for business, personal, or other purposes thereby allowing users to take advantage of the value of the information. Because technology and information handling needs and requirements vary between different users or applications, information handling systems may also vary regarding what information is handled, how the information is handled, how much information is processed, stored, or communicated, and how quickly and efficiently the information may be processed, stored, or communicated. The variations in information handling systems allow for information handling systems to be general or configured for a specific user or specific use such as financial transaction processing, airline reservations, enterprise data storage, or global communications. In addition, information handling systems may include a variety of hardware and software components that may be configured to process, store, and communicate information and may include one or more computer systems, data storage systems, and networking systems.
  • In a system comprising multiple information handling system chassis, the various chassis may be of different types and may be cabled and wired together in a particular manner. A console for managing the system may need to determine the connectivity among the various chassis and components internal to the chassis in order to determine a topology of the system.
  • SUMMARY
  • In accordance with the teachings of the present disclosure, certain disadvantages and problems associated with validating a topology in a multi-chassis environment have been reduced or eliminated.
  • In accordance with embodiments of the present disclosure, an information handling system may include a processor and a memory coupled to the processor, the memory having program instructions stored thereon that, upon execution by the processor, cause the processor to determine a topology of connectivity of various components of a system comprising multiple information handling system chassis and apply validation rules to the topology to validate the topology.
  • In accordance with these and other embodiments of the present disclosure, a method may include determining a topology of connectivity of various components of a system comprising multiple information handling system chassis and applying validation rules to the topology to validate the topology.
  • In accordance with these and other embodiments of the present disclosure, an article of manufacture may include a non-transitory computer readable medium and computer-executable instructions carried on the computer readable medium, the instructions readable by a processor, the instructions, when read and executed, for causing the processor to determine a topology of connectivity of various components of a system comprising multiple information handling system chassis apply validation rules to the topology to validate the topology.
  • Technical advantages of the present disclosure may be readily apparent to one skilled in the art from the figures, description and claims included herein. The objects and advantages of the embodiments will be realized and achieved at least by the elements, features, and combinations particularly pointed out in the claims.
  • It is to be understood that both the foregoing general description and the following detailed description are examples and explanatory and are not restrictive of the claims set forth in this disclosure.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • A more complete understanding of the present embodiments and advantages thereof may be acquired by referring to the following description taken in conjunction with the accompanying drawings, in which like reference numbers indicate like features, and wherein:
  • FIG. 1 illustrates a block diagram of a system comprising multiple information handling system chassis, in accordance with embodiments of the present disclosure;
  • FIG. 2 illustrates a flow chart of an example method for network topology discovery, in accordance with embodiments of the present disclosure;
  • FIG. 3 illustrates a flow chart of an example method for network topology validation, in accordance with embodiments of the present disclosure; and
  • FIG. 4 illustrates a block diagram of an information handling system, in accordance with embodiments of the present disclosure.
  • DETAILED DESCRIPTION
  • Preferred embodiments and their advantages are best understood by reference to FIGS. 1 through 4, wherein like numbers are used to indicate like and corresponding parts.
  • For the purposes of this disclosure, an information handling system may include any instrumentality or aggregate of instrumentalities operable to compute, classify, process, transmit, receive, retrieve, originate, switch, store, display, manifest, detect, record, reproduce, handle, or utilize any form of information, intelligence, or data for business, scientific, control, entertainment, or other purposes. For example, an information handling system may be a personal computer, a personal digital assistant (PDA), a consumer electronic device, a network storage device, or any other suitable device and may vary in size, shape, performance, functionality, and price. The information handling system may include memory, one or more processing resources such as a central processing unit (“CPU”) or hardware or software control logic. Additional components of the information handling system may include one or more storage devices, one or more communications ports for communicating with external devices as well as various I/O devices, such as a keyboard, a mouse, and a video display. The information handling system may also include one or more buses operable to transmit communication between the various hardware components.
  • For the purposes of this disclosure, information handling resources may broadly refer to any component system, device or apparatus of an information handling system, including without limitation processors, buses, memories, I/O devices and/or interfaces, storage resources, network interfaces, motherboards, electro-mechanical devices (e.g., fans), displays, and power supplies.
  • For the purposes of this disclosure, computer-readable media may include any instrumentality or aggregation of instrumentalities that may retain data and/or instructions for a period of time. Computer-readable media may include, without limitation, storage media such as a direct access storage device (e.g., a hard disk drive or floppy disk), a sequential access storage device (e.g., a tape disk drive), compact disk, CD-ROM, DVD, random access memory (“RAM”), read-only memory (“ROM”), electrically erasable programmable read-only memory (“EEPROM”), and/or flash memory; as well as communications media such as wires, optical fibers, microwaves, radio waves, and other electromagnetic and/or optical carriers; and/or any combination of the foregoing.
  • FIG. 1 illustrates a block diagram of a system 100 comprising multiple information handling system chassis 101, in accordance with embodiments of the present disclosure. As depicted in FIG. 1, system 100 may comprise a plurality of chassis 101 (e.g., chassis 101 a, 101 b, and 101 c), each chassis 101 including information handling systems and/or information handling resources, a private inter-chassis network 150, and a console 120. For example, chassis 101 a may include a switch 106 a and a chassis management controller 112. As another example, chassis 101 b may include a switch 106 b and a chassis management controller 112. As a further example, chassis 101 c may include switches 106 c and 106 d, one or more servers 102, and a chassis management controller 112. The various servers 102 and switches 106 may each comprise ports 110 for interfaces with one another, with example connectivity among ports 110 depicted in FIG. 1. For the purposes of clarity and exposition, chassis 101 a and 101 b are shown without servers (e.g., servers 102). However, in some embodiments, one or both of chassis 101 a and 101 b may include one or more servers 102.
  • A server 102 may generally be operable to receive data from and/or communicate data to one or more information handling resources of chassis 101 c. In certain embodiments, a server 102 may comprise a blade server having modular physical design. In these and other embodiments, server 102 may comprise an M class server.
  • Each server 102 may include a host management controller 104. Host management controller 104 may be implemented by, for example, a microprocessor, microcontroller, DSP, ASIC, EEPROM, or any combination thereof. Host management controller 104 may be configured to communicate with chassis management controller 112. Such communication may be made, for example, via a private management network fabric integral to server 102 (not explicitly shown). Host management controller 104 may be configured to provide out-of-band management facilities for management of system 100. Such management may be made by information handling resources of chassis 101 even if system 100 is powered off or powered to a standby state. Host management controller 104 may include a processor, memory, and network connection separate from the rest of system 100. In certain embodiments, host management controller 104 may include or may be an integral part of a baseboard management controller (BMC) or an Integrated Dell Remote Access Controller (iDRAC).
  • A switch 106 may comprise any suitable system, device, or apparatus for receiving, processing, and forwarding packets. For example, each of switch 106 a and 106 b may serve as an interface between chassis 101 c and a data network and may process and route packets between chassis 101 c and such data network. As another example, switches 106 c and 106 d may each serve as an interface between servers 102 and other chassis 101 (e.g., 101 a and 101 b) external to chassis 101 c.
  • Each port 110 may comprise a connector, slot, or another interface for receiving one end of a physical connection (e.g., wire, cable) coupled to a corresponding port 110 of another information handling resource.
  • A chassis management controller 112 may comprise any system, device, or apparatus configured to facilitate management and/or control of system 100 embodied by chassis 101, its information handling systems 102, and/or one or more of its component information handling resources. Chassis management controller 112 may be configured to issue commands and/or other signals to manage and/or control an information handling system 102 and/or information handling resources of system 100. Chassis management controller 112 may comprise a microprocessor, microcontroller, DSP, ASIC, field programmable gate array (“FPGA”), EEPROM, or any combination thereof. In some embodiments, chassis management controller 112 may provide a management console for user/administrator access to these functions. For example, chassis management controller 112 may implement Representational State Transfer (“REST”) or another suitable management protocol permitting a user to remotely access chassis management controller 112 to configure system 100 and its various information handling resources. In such embodiments, chassis management controller 112 may interface with a network interface, thus allowing for “out-of-band” control of chassis 101, such that communications to and from chassis management controller 112 are communicated via a management channel physically isolated from an “in-band” communication channel of chassis 101 for which non-management communication may take place. Thus, for example, if a failure occurs in chassis 101 that prevents an administrator from interfacing with chassis 101 via the in-band communication channel or a user interface associated with chassis 101 (e.g., power failure, etc.), the administrator may still be able to monitor and/or manage chassis 101 (e.g., to diagnose problems that may have caused failure) via chassis management controller 112. In the same or alternative embodiments, chassis management controller 112 may allow an administrator to remotely manage one or more parameters associated with operation of chassis 101 and its various information handling resources (e.g., power usage, processor allocation, memory allocation, security privileges, etc.). In some embodiments, chassis management controller 112 may include a management services module. Although beyond the scope of this disclosure, in some embodiments, one of the chassis management controllers 112 of the various chassis 101 may be selected as a “lead” chassis management controller 112, such that an administrator may manage the entirety of system 100 by interfacing with a single chassis management controller 112.
  • Private inter-chassis network 150 may comprise a network and/or fabric configured to couple information chassis management controllers 112 of the various chassis 101 in system 100 to each other. In these and other embodiments, inter-chassis network 150 may include a communication infrastructure, which provides physical connections, and a management layer, which organizes the physical connections and chassis management controllers 112 communicatively coupled to private inter-chassis network 150. Private inter-chassis network 150 may be implemented as, or may be a part of, an Ethernet local area network (LAN) or any other appropriate architecture or system that facilitates the communication of signals, data, and/or messages.
  • Console 120 may be communicatively coupled to private inter-chassis network 150 and may comprise an information handling system or a program executable on an information handling system for monitoring and management of the various chassis 101 of system 100 and their various components. For example, as described in greater detail below, console 120 may be able to, based on messages communicated among chassis 101 and management controllers 112, determine the topology of connectivity among the various components of system 100. In some embodiments, console 120 may also be capable of rendering a graphical representation of the topology of connectivity to a user via a user interface (not explicitly shown) of console 120. Although FIG. 1 depicts console 120 as a stand-alone component of system 100, in some embodiments, console 120 may be integral to or otherwise embodied in a chassis management controller 112.
  • In operation, console 120 may leverage payload information of packets (e.g., Link Layer Discovery Protocol or “LLDP” packets) communicated among switches 106 and/or servers 102 in response to a switch 106 or server 102 being coupled to another device in system 100 and may also leverage advertisements (e.g., multicast Domain Name Service or mDNS) communicated via private inter-chassis network 150 in order to extract information from such payloads and advertisements and process such information to determine topology of connectivity of devices of system 100.
  • For example, in response to a server 102 being added to system 100, a host management controller 104 of such server may communicate LLDP packets which may be received not only by switches 106 c and 106 d of the chassis 101 c comprising the server 102, but may be also passed-through from switches 106 c and 106 d to switches 106 a and 106 b of chassis 101 a and 101 b, respectively. The payload of each such LLDP packet may include various identifying information, including without limitation identifying information (e.g., service tag or serial number) of the server 102 from which the LLDP packet originated, identifying information (e.g., service tag or serial number) of the chassis 101 comprising the server 102, identifying information (e.g., fully-qualified device descriptor) of the port 110 of the server 102 originating the LLDP packet, and the switch fabric (e.g., 106 c or 106 d) to which the packet-originating port 110 is coupled. Thus, based on such payload, a console 120 may be able to determine the chassis 101 housing the server 102, a switch fabric to which the server 102 is coupled, and connectivity between the server 102 and a switch 106 receiving the LLDP packet external to the chassis 101.
  • As another example, chassis management controllers 112 may communicate mDNS advertisements on private inter-chassis network 150. Such advertisements may include various identifying information, including without limitation identifying information (e.g., service tag or serial number) of a chassis 101 including a switch fabric, identifying information (e.g., slot number) of a switch 106 of the switch fabric, and a switch type for the switch 106. Accordingly, console 120 may correlate such information from mDNS advertisements to the information from LLDP payloads to determine the switch type of each switch fabric.
  • As a further example, a switch 106 may include metadata information (e.g., an I/O module or “IOM” descriptor file) that is specific to a switch type and uplink (e.g., connectivity between a server and switch 106 c/106 d) and downlink connectivity (connectivity between a switch 106 c/106 d and a switch 106 a/106 b) of a switch 106. Accordingly, console 120 may correlate such IOM metadata information to information from mDNS advertisements and/or information from LLDP payloads to determine the complete topology of each switch fabric.
  • Inter-switch connectivity (e.g., between switches 106 a and 106 b) and external network connectivity (e.g., between an external data network and switches 106 a/106 b) may be determined based on LLDP payloads communicated between switches (e.g., between switches 106 a and 106 b) and between individual switches (e.g., between switches 106 a and 106 b) and the external data network.
  • FIG. 2 illustrates a flow chart of an example method 200 for network topology discovery, in accordance with embodiments of the present disclosure. According to some embodiments, method 200 may begin at step 202. As noted above, teachings of the present disclosure may be implemented in a variety of configurations of system 100. As such, the preferred initialization point for method 200 and the order of the steps comprising method 200 may depend on the implementation chosen.
  • At step 202, console 120 may receive an LLDP packet, mDNS advertisement, or IOM metadata in response to a physical connection being made in system 100. At step 204, console 120 may process information present in the LLDP packet, mDNS advertisement, or IOM metadata and along with other information previously received from other LLDP packets, mDNS advertisements, or IOM metadata. At step 206, based on such processing, console 120 may update a table, list, map, database, or other data structure defining the topology of connectivity of the various components of system 100. At step 208, in some embodiments, console 120 may generate and output to a user interface a graphical display representative of the topology.
  • Although the foregoing contemplates that console 120 receives various traffic, it is noted that console 120 may not receive all of such traffic (e.g., console 120 may not receive LLDP packets). Instead, a switch 106 may receive LLDP packets and determine topology based on such packets. Thus, a complete topology of a system may be determined by the aggregate of traffic received by switches 106 and console 120.
  • Although FIG. 2 discloses a particular number of steps to be taken with respect to method 200, method 200 may be executed with greater or lesser steps than those depicted in FIG. 2. In addition, although FIG. 2 discloses a certain order of steps to be taken with respect to method 200, the steps comprising method 200 may be completed in any suitable order.
  • Method 200 may be implemented using system 100, and/or any other system operable to implement method 200. In certain embodiments, method 200 may be implemented partially or fully in software and/or firmware embodied in computer-readable media.
  • After building the topology as described above, console 120 may also be configured to validate the topology against a set of validation rules. To illustrate, in a multi-chassis group, such as system 100, that supports fabric-mode operations, the various I/O modules must typically be wired in prescriptive ways. After a topology is detected, it may be desirable that console 120 provide guidance to a user regarding any incorrect wiring in system 100. A topology could be incorrectly wired for numerous reasons. For example, a cable may be missing or not properly connected between its source and destination. As another example, cabling may violate prescriptive rules, such as a rule that dictates that a switch in one chassis 101 cannot be connected to a module in a different slot in a different chassis 101, a rule that dictates that wiring across different multi-chassis groups is not permitted, or a rule against mixing switch types in a chassis 101. Thus, validation rules may be complex rules that go beyond simply matching particular source ports and destination ports. Validation rules may be expressed on characteristics of I/O modules/switches 106 (e.g., slot location, mismatch of switch types in a chassis 101) or aggregate characteristics of chassis 101 in a group (e.g., group membership of chassis 101).
  • For further illustration, below are non-limiting examples of rules that may be applied by console 120 in validating a topology:
  • 1) Specific source/destination port wiring—such a rule may be expressed in terms of a match of source and destination ports with specific port numbers, which may ensure that switches 106 are connected in a redundant manner.
  • 2) Inter-chassis switch wiring: such a rule may be expressed in terms of a matching of a slot location for switches 106 disposed in different chassis. For example, if a first switch 106 is in a slot “A” of a first chassis 101 and is coupled to a second switch 106 of a second chassis 101, the rule may ensure that second switch 106 is in the slot “A” of the second chassis.
  • 3) Group membership: such a rule may be expressed in terms of a match in chassis group membership between a device (e.g., switch 106) having a source port 110 and another device (e.g., switch 106) having a destination port 110 of a connection.
  • 4) Switch types: such a rule may be expressed in terms of the switch type in an appropriate fabric location. For example, if two slots of a chassis 101 are populated with switches 106, this rule may ensure that both switches 106 are of the same switch type.
  • FIG. 3 illustrates a flow chart of an example method 300 for network topology validation, in accordance with embodiments of the present disclosure. According to some embodiments, method 300 may begin at step 302. As noted above, teachings of the present disclosure may be implemented in a variety of configurations of system 100. As such, the preferred initialization point for method 300 and the order of the steps comprising method 300 may depend on the implementation chosen.
  • At step 302, console 120 may determine the topology of system 100. In some embodiments, such topology may be constructed in accordance with method 200 described above.
  • At step 304, console 120 may apply a set of validation rules (e.g., one or more of the various rules discussed above) to the topology to determine if the topology satisfies the validation rules. If the topology satisfies the topology rules, method 300 may proceed again to step 302, and the topology may be continuously validated as the topology is changed. If the topology violates one or more of the topology rules, method 300 may proceed to step 306.
  • At step 306, in response to the topology violating one or more of the topology rules, console 120 may generate and output to a user interface an indication of which of the topology rules have been violated. Such indication may be in terms of a list or a graphical representation of the topology with indications within the graphical representation of the portion of the topology that are in violation of the topology rules. After completion of step 306, method 300 may proceed again to step 302, and the topology may be continuously validated as the topology is changed.
  • Although FIG. 3 discloses a particular number of steps to be taken with respect to method 300, method 300 may be executed with greater or lesser steps than those depicted in FIG. 3. In addition, although FIG. 3 discloses a certain order of steps to be taken with respect to method 300, the steps comprising method 300 may be completed in any suitable order.
  • Method 300 may be implemented using system 100, and/or any other system operable to implement method 300. In certain embodiments, method 300 may be implemented partially or fully in software and/or firmware embodied in computer-readable media.
  • FIG. 4 illustrates a block diagram of an information handling system 400, in accordance with embodiments of the present disclosure. Information handling system 400 is an instance of console 120 and/or a server 102 illustrated in FIG. 1. As shown, information handling system 400 includes one or more CPUs 402. In various embodiments, information handling system 400 may be a single-processor system including one CPU 402, or a multi-processor system including two or more CPUs 402 (e.g., two, four, eight, or any other suitable number). CPU(s) 402 may include any processor capable of executing program instructions. For example, in various embodiments, CPU(s) 402 may be general-purpose or embedded processors implementing any of a variety of instruction set architectures (ISAs), such as the x86, POWERPC, ARM, SPARC, or MIPS ISAs, or any other suitable ISA. In multi-processor systems, each of CPU(s) 402 may commonly, but not necessarily, implement the same ISA. In such an embodiment, a motherboard may be configured to provide structural support, power, and electrical connectivity between the various aforementioned components. Such a motherboard may include multiple connector sockets in various configurations, adapted to receive pluggable circuit cards, component chip packages, etc.
  • CPU(s) 402 are coupled to northbridge controller or chipset 404 via front-side bus 406. Northbridge controller 404 may be configured to coordinate I/O traffic between CPU(s) 402 and other components. For example, in this particular implementation, northbridge controller 404 is coupled to graphics device(s) 408 (e.g., one or more video cards or adaptors, etc.) via graphics bus 410 (e.g., an Accelerated Graphics Port or AGP bus, a Peripheral Component Interconnect or PCI bus, etc.). Northbridge controller 404 is also coupled to system memory 412 via memory bus 414. Memory 412 may be configured to store program instructions and/or data accessible by CPU(s) 402. In various embodiments, memory 412 may be implemented using any suitable memory technology, such as static RAM (SRAM), synchronous dynamic RAM (SDRAM), nonvolatile/Flash-type memory, or any other type of memory.
  • Northbridge controller 404 is coupled to southbridge controller or chipset 416 via internal bus 418. Generally, southbridge controller 416 may be configured to handle various of computing device information handling system 400's I/O operations, and it may provide interfaces such as, for instance, Universal Serial Bus (USB), audio, serial, parallel, Ethernet, etc., via port(s), pin(s), and/or adapter(s) 432 over bus 434. For example, southbridge controller 416 may be configured to allow data to be exchanged between information handling system 400 and other devices, such as other information handling systems attached to a network. In various embodiments, southbridge controller 416 may support communication via wired or wireless general data networks, such as any suitable type of Ethernet network, for example; via telecommunications/telephony networks such as analog voice networks or digital fiber communications networks; via storage area networks such as Fiber Channel SANs; or via any other suitable type of network and/or protocol.
  • Southbridge controller 416 may also enable connection to one or more keyboards, keypads, touch screens, scanning devices, voice or optical recognition devices, or any other devices suitable for entering or retrieving data. Multiple I/O devices may be present in information handling system 400. In some embodiments, I/O devices may be separate from information handling system 400 and may interact with information handling system 400 through a wired or wireless connection. As shown, southbridge controller 416 is further coupled to one or more PCI devices 420 (e.g., modems, network cards, sound cards, video cards, etc.) via PCI bus 422. Southbridge controller 416 is also coupled to Basic I/O System (BIOS) 424, Super I/O Controller 426, and Baseboard Management Controller (BMC) 428 via Low Pin Count (LPC) bus 430.
  • BIOS 424 includes non-volatile memory having program instructions stored thereon. Those instructions may be usable for CPU(s) 402 to initialize and test other hardware components and/or to load an Operating System (OS) onto information handling system 400. As such, BIOS 424 may include a firmware interface that allows CPU(s) 402 to load and execute certain firmware, as described in more detail below. In some cases, such firmware may include program code that is compatible with the Unified Extensible Firmware Interface (UEFI) specification, although other types of firmware may be used.
  • BMC controller 428 may include non-volatile memory having program instructions stored thereon that are usable by CPU(s) 402 to enable remote management of information handling system 400. For example, BMC controller 428 may enable a user to discover, configure, and manage BMC controller 428, setup configuration options, resolve and administer hardware or software problems, etc. Additionally or alternatively, BMC controller 428 may include one or more firmware volumes, each volume having one or more firmware files used by the BIOS' firmware interface to initialize and test components of information handling system 400.
  • Super I/O Controller 426 combines interfaces for a variety of lower bandwidth or low data rate devices. Those devices may include, for example, floppy disks, parallel ports, keyboard and mouse, temperature sensor and fan speed monitoring, etc. For example, the super I/O controller 426 may be coupled to the one or more upstream sensors and to the one or more downstream sensors.
  • In some cases, information handling system 400 may be configured to access different types of computer-accessible media separate from memory 412. Generally speaking, a computer-accessible medium may include any tangible, non-transitory storage media or memory media such as electronic, magnetic, or optical media—e.g., magnetic disk, a hard drive, a CD/DVD-ROM, a Flash memory, etc. coupled to information handling system 400 via northbridge controller 404 and/or southbridge controller 416.
  • As used herein, when two or more elements are referred to as “coupled” to one another, such term indicates that such two or more elements are in electronic communication or mechanical communication, as applicable, whether connected indirectly or directly, with or without intervening elements.
  • This disclosure encompasses all changes, substitutions, variations, alterations, and modifications to the example embodiments herein that a person having ordinary skill in the art would comprehend. Similarly, where appropriate, the appended claims encompass all changes, substitutions, variations, alterations, and modifications to the example embodiments herein that a person having ordinary skill in the art would comprehend. Moreover, reference in the appended claims to an apparatus or system or a component of an apparatus or system being adapted to, arranged to, capable of, configured to, enabled to, operable to, or operative to perform a particular function encompasses that apparatus, system, or component, whether or not it or that particular function is activated, turned on, or unlocked, as long as that apparatus, system, or component is so adapted, arranged, capable, configured, enabled, operable, or operative.
  • All examples and conditional language recited herein are intended for pedagogical objects to aid the reader in understanding the disclosure and the concepts contributed by the inventor to furthering the art, and are construed as being without limitation to such specifically recited examples and conditions. Although embodiments of the present disclosure have been described in detail, it should be understood that various changes, substitutions, and alterations could be made hereto without departing from the spirit and scope of the disclosure.

Claims (20)

What is claimed is:
1. An information handling system comprising:
a processor; and
a memory coupled to the processor, the memory having program instructions stored thereon that, upon execution by the processor, cause the processor to:
determine a topology of connectivity of various components of a system comprising multiple information handling system chassis; and
apply validation rules to the topology to validate the topology.
2. The information handling system of claim 1, the instructions for further causing the processor to generate an indication to a graphical display indicative of violations of the validation rules.
3. The information handling system of claim 1, wherein the validation rules comprise at least one rule based on characteristics of one or more switches integral to the system.
4. The information handling system of claim 3, wherein the at least one rule ensures that a plurality of switches integral to a chassis of the system are of the same switch type.
5. The information handling system of claim 3, wherein the at least one rule ensures that for a first switch of a first chassis of the system coupled to a second switch of a second chassis of the system, that the first switch and second switch are disposed in a particular respective slot of each of the first chassis and the second chassis.
6. The information handling system of claim 1, wherein the validation rules comprise at least one rule based on characteristics of the system.
7. The information handling system of claim 6, wherein the at least one rule ensures that for a first component of the system coupled to a second component of the system that a chassis group membership of the first component matches that of the second component.
8. The information handling system of claim 1, wherein the validation rules comprise at least one rule based on characteristics other than a match of a source port of a first component to a destination port of another component.
9. A method comprising:
determining a topology of connectivity of various components of a system comprising multiple information handling system chassis; and
applying validation rules to the topology to validate the topology.
10. The method of claim 9, further comprising generating an indication to a graphical display indicative of violations of the validation rules.
11. The method of claim 9, wherein the validation rules comprise at least one rule based on characteristics of one or more switches integral to the system.
12. The method of claim 11, wherein the at least one rule ensures that a plurality of switches integral to a chassis of the system are of the same switch type.
13. The method of claim 11, wherein the at least one rule ensures that for a first switch of a first chassis of the system coupled to a second switch of a second chassis of the system, that the first switch and second switch are disposed in a particular respective slot of each of the first chassis and the second chassis.
14. The method of claim 9, wherein the validation rules comprise at least one rule based on characteristics of the system.
15. The method of claim 14, wherein the at least one rule ensures that for a first component of the system coupled to a second component of the system that a chassis group membership of the first component matches that of the second component.
16. The method of claim 9, wherein the validation rules comprise at least one rule based on characteristics other than a match of a source port of a first component to a destination port of another component.
17. An article of manufacture comprising:
a non-transitory computer readable medium; and
computer-executable instructions carried on the computer readable medium, the instructions readable by a processor, the instructions, when read and executed, for causing the processor to:
determine a topology of connectivity of various components of a system comprising multiple information handling system chassis; and
apply validation rules to the topology to validate the topology.
18. The article of manufacture of claim 17, the instructions for further causing the processor to generate an indication to a graphical display indicative of violations of the validation rules.
19. The article of manufacture of claim 17, wherein the validation rules comprise at least one rule based on characteristics of one or more switches integral to the system.
20. The article of manufacture of claim 17, wherein the validation rules comprise at least one rule based on characteristics of the system.
US15/440,932 2017-02-23 2017-02-23 Systems and methods for network topology validation Active 2037-09-02 US10785118B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/440,932 US10785118B2 (en) 2017-02-23 2017-02-23 Systems and methods for network topology validation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US15/440,932 US10785118B2 (en) 2017-02-23 2017-02-23 Systems and methods for network topology validation

Publications (2)

Publication Number Publication Date
US20180241631A1 true US20180241631A1 (en) 2018-08-23
US10785118B2 US10785118B2 (en) 2020-09-22

Family

ID=63167539

Family Applications (1)

Application Number Title Priority Date Filing Date
US15/440,932 Active 2037-09-02 US10785118B2 (en) 2017-02-23 2017-02-23 Systems and methods for network topology validation

Country Status (1)

Country Link
US (1) US10785118B2 (en)

Citations (85)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5926463A (en) * 1997-10-06 1999-07-20 3Com Corporation Method and apparatus for viewing and managing a configuration of a computer network
US20020085551A1 (en) * 2000-11-14 2002-07-04 Altima Communications, Inc. Linked network switch configuration
US20020156888A1 (en) * 2001-04-23 2002-10-24 Lee Man-Ho L. Method and apparatus for detecting and reporting configuration errors in a multi-component switching fabric
US6480955B1 (en) * 1999-07-09 2002-11-12 Lsi Logic Corporation Methods and apparatus for committing configuration changes to managed devices prior to completion of the configuration change
US20030091042A1 (en) * 2001-10-05 2003-05-15 Broadcom Corporation Method and apparatus for enabling access on a network switch
US20030130833A1 (en) * 2001-04-20 2003-07-10 Vern Brownell Reconfigurable, virtual processing system, cluster, network and method
US6731631B1 (en) * 2000-08-11 2004-05-04 Paion Company, Limited System, method and article of manufacture for updating a switching table in a switch fabric chipset system
US7020145B1 (en) * 2001-09-24 2006-03-28 Hewlett-Packard Development Company, L.P. Network topology manager
US20070115962A1 (en) * 2005-11-18 2007-05-24 Cisco Technology, Inc. Techniques configuring customer equipment for network operations from provider edge
US20080219184A1 (en) * 2007-03-05 2008-09-11 Fowler Jeffery L Discovery of network devices
US20080310421A1 (en) * 2007-06-12 2008-12-18 Teisberg Robert R Managing Connectivity in a Virtual Network
US20090063666A1 (en) * 2007-09-05 2009-03-05 International Business Machines Corporation Network and Method for the Configuration Thereof
US20090189774A1 (en) * 2008-01-28 2009-07-30 Dell Products L.P. Power Topology Determination
US7640325B1 (en) * 1999-07-09 2009-12-29 Lsi Corporation Methods and apparatus for issuing updates to multiple management entities
US7656822B1 (en) * 2003-12-22 2010-02-02 Sun Microsystems, Inc. Method and apparatus for decentralized device and service description and discovery
US20100122175A1 (en) * 2008-11-12 2010-05-13 Sanjay Gupta Tool for visualizing configuration and status of a network appliance
US20100157516A1 (en) * 2008-12-22 2010-06-24 Panduit Corp. Physical infrastructure management system
US20100211664A1 (en) * 2009-02-13 2010-08-19 Adc Telecommunications, Inc. Aggregation of physical layer information related to a network
US20110279266A1 (en) * 2010-05-12 2011-11-17 International Business Machines Corporation Automatic verification of network cable connections
US20110299413A1 (en) * 2010-06-02 2011-12-08 Brocade Communications Systems, Inc. Port profile management for virtual cluster switching
US20120002670A1 (en) * 2010-07-01 2012-01-05 Force 10 Networks, Inc. Multiple chassis stacking using front end ports
US20120079101A1 (en) * 2010-06-02 2012-03-29 Blue Coat Systems, Inc. Behavioral Classification of Network Data Flows
US20120079139A1 (en) * 2010-09-28 2012-03-29 Hitachi, Ltd. Computer system, apparatus, and method for checking for cable misconnections
US20120207039A1 (en) * 2011-02-16 2012-08-16 Oracle International Corporation Method and system for validating network traffic classification in a blade server
US8443065B1 (en) * 2010-11-08 2013-05-14 Adtran, Inc. System and method for locating, identifying and provisioning newly deployed network devices
US20130148654A1 (en) * 2011-12-09 2013-06-13 Brocade Communications Systems, Inc. Port profile analytics
US20130148511A1 (en) * 2011-12-09 2013-06-13 Brocade Communications Systems, Inc. Ampp active profile presentation
US20130250810A1 (en) * 2011-08-30 2013-09-26 Qualcomm Incorporated Topology discovery in a hybrid network
US20140006597A1 (en) * 2012-06-29 2014-01-02 Mrittika Ganguli Method, system, and device for managing server hardware resources in a cloud scheduling environment
US20140012967A1 (en) * 2012-07-05 2014-01-09 Gopal Agarwal System and method for supporting multicast domain name system device and service classification
US20140010150A1 (en) * 2012-07-05 2014-01-09 Gopal Agarwal Location based service discovery
US20140036924A1 (en) * 2012-08-06 2014-02-06 International Business Machines Corporation Multi-chassis link aggregation in a distributed virtual bridge
US20140101467A1 (en) * 2012-01-30 2014-04-10 Microsoft Corporation Discovering, validating, and configuring hardware-inventory components
US20140105029A1 (en) * 2012-10-16 2014-04-17 Cisco Technology, Inc. Detection of cabling error in communication network
US20140126424A1 (en) * 2012-11-07 2014-05-08 Dell Products L.P. Event driven network system
US20140215027A1 (en) * 2013-01-25 2014-07-31 Dell Products L.P. System and method for determining the configuration of switches in virtual link trunking environments
US8805982B1 (en) * 2007-06-29 2014-08-12 Ciena Corporation Progressively determining a network topology and using neighbor information to determine network topology
US20140226461A1 (en) * 2013-02-13 2014-08-14 Hitachi Metals, Ltd. Communication System and Network Relay Device
US8934495B1 (en) * 2009-07-31 2015-01-13 Anue Systems, Inc. Filtering path view graphical user interfaces and related systems and methods
US20150081829A1 (en) * 2013-09-13 2015-03-19 American Megatrends, Inc. Out-of-band replicating bios setting data across computers
US20150095530A1 (en) * 2013-09-30 2015-04-02 International Business Machines Coporation Dynamic port naming in a chassis
US20150100700A1 (en) * 2013-10-09 2015-04-09 Cisco Technology, Inc. Communicating Service Denials Back to Client During MDNS Service Discovery
US20150120951A1 (en) * 2013-10-31 2015-04-30 Aruba Networks, Inc. Method and system for controlling access to shared devices
US20150124837A1 (en) * 2013-11-05 2015-05-07 Telefonaktiebolaget L M Ericsson (Publ) Method and system of supporting operator commands in link aggregation group
US20150188760A1 (en) * 2013-12-27 2015-07-02 Dell Products L.P. N-node virtual link trunking (vlt) systems management plane
US20150188808A1 (en) * 2013-12-27 2015-07-02 Dell Products L.P. Routing in spine-leaf networking systems
US20150200803A1 (en) * 2014-01-15 2015-07-16 Dell Products L.P. System and method for network topology management
US9130870B1 (en) * 2011-04-15 2015-09-08 Big Switch Networks, Inc. Methods for determining network topologies
US20150256413A1 (en) * 2014-03-06 2015-09-10 Sideband Networks Inc. Network system with live topology mechanism and method of operation thereof
US20150271022A1 (en) * 2014-03-23 2015-09-24 Avaya Inc. Configuration of networks using client device access of remote server
US20150326502A1 (en) * 2013-02-20 2015-11-12 Fujitsu Limited Switch and setting method
US20150334001A1 (en) * 2013-01-21 2015-11-19 Nec Corporation Communication system, control apparatus, and network topology management method
US20150365457A1 (en) * 2014-06-16 2015-12-17 Asurion, Llc Service discovery across different networks
US20160006620A1 (en) * 2014-07-01 2016-01-07 American Megatrends, Inc. Hardware management and control of computer components through physical layout diagrams
US20160006619A1 (en) * 2014-07-01 2016-01-07 American Megatrends, Inc. Generating graphical diagram of physical layout of computer platforms
US20160014073A1 (en) * 2014-07-11 2016-01-14 Wmware, Inc. Methods and apparatus to configure hardware management systems for use in virtual server rack deployments for virtual computing environments
US9246702B1 (en) * 2012-01-31 2016-01-26 Cisco Technology, Inc. System and method for configuring service appliances as virtual line cards in a network environment
US9286047B1 (en) * 2013-02-13 2016-03-15 Cisco Technology, Inc. Deployment and upgrade of network devices in a network environment
US9294347B2 (en) * 2014-03-20 2016-03-22 Dell Products Lp Systems and methods for automatic access layer configuration
US20160094461A1 (en) * 2014-09-25 2016-03-31 Dell Products L.P. Data traffic policy management system
US9319335B1 (en) * 2010-12-07 2016-04-19 Pluribus Networks, Inc. Distributed operating system for a layer 2 fabric
US20160119186A1 (en) * 2013-06-09 2016-04-28 Hangzhou H3C Technologies Co., Ltd. Zero-configuration networking protocol
US20160124477A1 (en) * 2014-10-31 2016-05-05 Cisco Technology, Inc. Device communication with power distribution units
US9338733B1 (en) * 2012-10-15 2016-05-10 Marvell International Ltd. Method and apparatus for determining the availability of a first device in a wireless network to provide a service to a second device within the wireless network
US9374285B1 (en) * 2013-02-07 2016-06-21 Big Switch Networks, Inc. Systems and methods for determining network topologies
US20160182284A1 (en) * 2014-12-17 2016-06-23 American Megatrends, Inc. System and method of performing high availability configuration and validation of virtual desktop infrastructure (vdi)
US20160191314A1 (en) * 2014-12-31 2016-06-30 Dell Products L.P. Multi-port selection and configuration
US20160212041A1 (en) * 2015-01-16 2016-07-21 Dell Products L.P. Switching layer 2 (l2) flows to a correct node in a vlt domain
US20160226704A1 (en) * 2015-02-04 2016-08-04 International Business Machines Corporation Port configuration for interconnected communications devices
US9497080B1 (en) * 2015-09-23 2016-11-15 International Business Machines Corporation Election and use of configuration manager
US20170026232A1 (en) * 2013-01-25 2017-01-26 Dell Products L.P. System and method for determining the configuration of switches in virtual link trunking environments
US20170034000A1 (en) * 2015-07-28 2017-02-02 Dell Products L.P. Inter-networking device link provisioning system
US9571411B1 (en) * 2014-12-29 2017-02-14 Juniper Networks, Inc. Using a firewall filter to select a member link of a link aggregation group
US20170054603A1 (en) * 2015-08-17 2017-02-23 Vmware, Inc. Hardware management systems for disaggregated rack architectures in virtual server rack deployments
US9590854B1 (en) * 2014-06-25 2017-03-07 Amazon Technologies, Inc. Automated network security
US20170078150A1 (en) * 2015-09-16 2017-03-16 Brocade Communications Systems, Inc. Ip-based interconnection of switches with a logical chassis
US20170118041A1 (en) * 2015-10-21 2017-04-27 Brocade Communications Systems, Inc. Distributed rule provisioning in an extended bridge
US20170126490A1 (en) * 2015-10-29 2017-05-04 Arista Networks, Inc. Method and system for configuring network devices
US9661022B2 (en) * 2015-04-24 2017-05-23 Dell Products L.P. System and method for authorizing devices joining a network fabric
US20170214600A1 (en) * 2016-01-27 2017-07-27 Oracle International Corporation System and method for supporting scalable representation of switch port status in a high performance computing environment
US20170230065A1 (en) * 2014-09-26 2017-08-10 Hewlett Packard Enterprise Development Lp Link aggregation configuration for a node in a software-defined network
US9906605B2 (en) * 2014-05-23 2018-02-27 Qualcomm Connected Experiences, Inc. Enhanced DNS-based service discovery in an internet of things (IoT) environment
US9935831B1 (en) * 2014-06-03 2018-04-03 Big Switch Networks, Inc. Systems and methods for controlling network switches using a switch modeling interface at a controller
US9942144B1 (en) * 2014-09-30 2018-04-10 Juniper Networks, Inc. Fibre channel over ethernet (FCoE) link aggregation group (LAG) support in data center networks
US10097457B1 (en) * 2015-12-28 2018-10-09 Juniper Networks, Inc. Resolving a mismatch among control plane parameter values received from multiple routing control devices

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7751405B1 (en) 2007-09-26 2010-07-06 Juniper Networks, Inc. Automatic configuration of label switched path tunnels using BGP attributes
CN102223278B (en) 2011-05-17 2015-06-24 中兴通讯股份有限公司 Realization method and system for enabling LLDP function on non-Ethernet link
US20140269299A1 (en) 2013-03-14 2014-09-18 Hewlett-Packard Development Company, L.P. Network controller normalization of network traffic
US9112790B2 (en) 2013-06-25 2015-08-18 Google Inc. Fabric network
US10102089B2 (en) 2014-12-17 2018-10-16 Intel Corporation Input/output (I/O) device configuration signature
WO2016099528A1 (en) 2014-12-19 2016-06-23 Hewlett Packard Enterprise Development Lp Multicast advertisement message for a network switch in a storage area network
US20170317899A1 (en) 2016-04-29 2017-11-02 Ca, Inc. Using traffic data to determine network topology

Patent Citations (87)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5926463A (en) * 1997-10-06 1999-07-20 3Com Corporation Method and apparatus for viewing and managing a configuration of a computer network
US6480955B1 (en) * 1999-07-09 2002-11-12 Lsi Logic Corporation Methods and apparatus for committing configuration changes to managed devices prior to completion of the configuration change
US7640325B1 (en) * 1999-07-09 2009-12-29 Lsi Corporation Methods and apparatus for issuing updates to multiple management entities
US6731631B1 (en) * 2000-08-11 2004-05-04 Paion Company, Limited System, method and article of manufacture for updating a switching table in a switch fabric chipset system
US20020085551A1 (en) * 2000-11-14 2002-07-04 Altima Communications, Inc. Linked network switch configuration
US20030130833A1 (en) * 2001-04-20 2003-07-10 Vern Brownell Reconfigurable, virtual processing system, cluster, network and method
US20020156888A1 (en) * 2001-04-23 2002-10-24 Lee Man-Ho L. Method and apparatus for detecting and reporting configuration errors in a multi-component switching fabric
US7020145B1 (en) * 2001-09-24 2006-03-28 Hewlett-Packard Development Company, L.P. Network topology manager
US20030091042A1 (en) * 2001-10-05 2003-05-15 Broadcom Corporation Method and apparatus for enabling access on a network switch
US7656822B1 (en) * 2003-12-22 2010-02-02 Sun Microsystems, Inc. Method and apparatus for decentralized device and service description and discovery
US20070115962A1 (en) * 2005-11-18 2007-05-24 Cisco Technology, Inc. Techniques configuring customer equipment for network operations from provider edge
US20080219184A1 (en) * 2007-03-05 2008-09-11 Fowler Jeffery L Discovery of network devices
US20080310421A1 (en) * 2007-06-12 2008-12-18 Teisberg Robert R Managing Connectivity in a Virtual Network
US8805982B1 (en) * 2007-06-29 2014-08-12 Ciena Corporation Progressively determining a network topology and using neighbor information to determine network topology
US20090063666A1 (en) * 2007-09-05 2009-03-05 International Business Machines Corporation Network and Method for the Configuration Thereof
US20090189774A1 (en) * 2008-01-28 2009-07-30 Dell Products L.P. Power Topology Determination
US20100122175A1 (en) * 2008-11-12 2010-05-13 Sanjay Gupta Tool for visualizing configuration and status of a network appliance
US20150236921A1 (en) * 2008-12-22 2015-08-20 Panduit Corp. Physical Infrastructure Management System
US20100157516A1 (en) * 2008-12-22 2010-06-24 Panduit Corp. Physical infrastructure management system
US20100211664A1 (en) * 2009-02-13 2010-08-19 Adc Telecommunications, Inc. Aggregation of physical layer information related to a network
US8934495B1 (en) * 2009-07-31 2015-01-13 Anue Systems, Inc. Filtering path view graphical user interfaces and related systems and methods
US20110279266A1 (en) * 2010-05-12 2011-11-17 International Business Machines Corporation Automatic verification of network cable connections
US20110299413A1 (en) * 2010-06-02 2011-12-08 Brocade Communications Systems, Inc. Port profile management for virtual cluster switching
US20120079101A1 (en) * 2010-06-02 2012-03-29 Blue Coat Systems, Inc. Behavioral Classification of Network Data Flows
US20120002670A1 (en) * 2010-07-01 2012-01-05 Force 10 Networks, Inc. Multiple chassis stacking using front end ports
US20120079139A1 (en) * 2010-09-28 2012-03-29 Hitachi, Ltd. Computer system, apparatus, and method for checking for cable misconnections
US8443065B1 (en) * 2010-11-08 2013-05-14 Adtran, Inc. System and method for locating, identifying and provisioning newly deployed network devices
US9319335B1 (en) * 2010-12-07 2016-04-19 Pluribus Networks, Inc. Distributed operating system for a layer 2 fabric
US20120207039A1 (en) * 2011-02-16 2012-08-16 Oracle International Corporation Method and system for validating network traffic classification in a blade server
US9130870B1 (en) * 2011-04-15 2015-09-08 Big Switch Networks, Inc. Methods for determining network topologies
US20130250810A1 (en) * 2011-08-30 2013-09-26 Qualcomm Incorporated Topology discovery in a hybrid network
US20130148511A1 (en) * 2011-12-09 2013-06-13 Brocade Communications Systems, Inc. Ampp active profile presentation
US20130148654A1 (en) * 2011-12-09 2013-06-13 Brocade Communications Systems, Inc. Port profile analytics
US20140101467A1 (en) * 2012-01-30 2014-04-10 Microsoft Corporation Discovering, validating, and configuring hardware-inventory components
US9246702B1 (en) * 2012-01-31 2016-01-26 Cisco Technology, Inc. System and method for configuring service appliances as virtual line cards in a network environment
US20140006597A1 (en) * 2012-06-29 2014-01-02 Mrittika Ganguli Method, system, and device for managing server hardware resources in a cloud scheduling environment
US20140012967A1 (en) * 2012-07-05 2014-01-09 Gopal Agarwal System and method for supporting multicast domain name system device and service classification
US20140010150A1 (en) * 2012-07-05 2014-01-09 Gopal Agarwal Location based service discovery
US20140036924A1 (en) * 2012-08-06 2014-02-06 International Business Machines Corporation Multi-chassis link aggregation in a distributed virtual bridge
US9338733B1 (en) * 2012-10-15 2016-05-10 Marvell International Ltd. Method and apparatus for determining the availability of a first device in a wireless network to provide a service to a second device within the wireless network
US20140105029A1 (en) * 2012-10-16 2014-04-17 Cisco Technology, Inc. Detection of cabling error in communication network
US20140126424A1 (en) * 2012-11-07 2014-05-08 Dell Products L.P. Event driven network system
US20150334001A1 (en) * 2013-01-21 2015-11-19 Nec Corporation Communication system, control apparatus, and network topology management method
US20140215027A1 (en) * 2013-01-25 2014-07-31 Dell Products L.P. System and method for determining the configuration of switches in virtual link trunking environments
US20170026232A1 (en) * 2013-01-25 2017-01-26 Dell Products L.P. System and method for determining the configuration of switches in virtual link trunking environments
US9374285B1 (en) * 2013-02-07 2016-06-21 Big Switch Networks, Inc. Systems and methods for determining network topologies
US9286047B1 (en) * 2013-02-13 2016-03-15 Cisco Technology, Inc. Deployment and upgrade of network devices in a network environment
US20140226461A1 (en) * 2013-02-13 2014-08-14 Hitachi Metals, Ltd. Communication System and Network Relay Device
US20150326502A1 (en) * 2013-02-20 2015-11-12 Fujitsu Limited Switch and setting method
US20160119186A1 (en) * 2013-06-09 2016-04-28 Hangzhou H3C Technologies Co., Ltd. Zero-configuration networking protocol
US20150081829A1 (en) * 2013-09-13 2015-03-19 American Megatrends, Inc. Out-of-band replicating bios setting data across computers
US20150095530A1 (en) * 2013-09-30 2015-04-02 International Business Machines Coporation Dynamic port naming in a chassis
US20150100700A1 (en) * 2013-10-09 2015-04-09 Cisco Technology, Inc. Communicating Service Denials Back to Client During MDNS Service Discovery
US20150120951A1 (en) * 2013-10-31 2015-04-30 Aruba Networks, Inc. Method and system for controlling access to shared devices
US20150124837A1 (en) * 2013-11-05 2015-05-07 Telefonaktiebolaget L M Ericsson (Publ) Method and system of supporting operator commands in link aggregation group
US20150188808A1 (en) * 2013-12-27 2015-07-02 Dell Products L.P. Routing in spine-leaf networking systems
US20150188760A1 (en) * 2013-12-27 2015-07-02 Dell Products L.P. N-node virtual link trunking (vlt) systems management plane
US20150200803A1 (en) * 2014-01-15 2015-07-16 Dell Products L.P. System and method for network topology management
US20150256413A1 (en) * 2014-03-06 2015-09-10 Sideband Networks Inc. Network system with live topology mechanism and method of operation thereof
US9294347B2 (en) * 2014-03-20 2016-03-22 Dell Products Lp Systems and methods for automatic access layer configuration
US20150271022A1 (en) * 2014-03-23 2015-09-24 Avaya Inc. Configuration of networks using client device access of remote server
US9906605B2 (en) * 2014-05-23 2018-02-27 Qualcomm Connected Experiences, Inc. Enhanced DNS-based service discovery in an internet of things (IoT) environment
US9935831B1 (en) * 2014-06-03 2018-04-03 Big Switch Networks, Inc. Systems and methods for controlling network switches using a switch modeling interface at a controller
US20150365457A1 (en) * 2014-06-16 2015-12-17 Asurion, Llc Service discovery across different networks
US9590854B1 (en) * 2014-06-25 2017-03-07 Amazon Technologies, Inc. Automated network security
US20160006619A1 (en) * 2014-07-01 2016-01-07 American Megatrends, Inc. Generating graphical diagram of physical layout of computer platforms
US20160006620A1 (en) * 2014-07-01 2016-01-07 American Megatrends, Inc. Hardware management and control of computer components through physical layout diagrams
US20160014073A1 (en) * 2014-07-11 2016-01-14 Wmware, Inc. Methods and apparatus to configure hardware management systems for use in virtual server rack deployments for virtual computing environments
US20160094461A1 (en) * 2014-09-25 2016-03-31 Dell Products L.P. Data traffic policy management system
US20170230065A1 (en) * 2014-09-26 2017-08-10 Hewlett Packard Enterprise Development Lp Link aggregation configuration for a node in a software-defined network
US9942144B1 (en) * 2014-09-30 2018-04-10 Juniper Networks, Inc. Fibre channel over ethernet (FCoE) link aggregation group (LAG) support in data center networks
US20160124477A1 (en) * 2014-10-31 2016-05-05 Cisco Technology, Inc. Device communication with power distribution units
US20160182284A1 (en) * 2014-12-17 2016-06-23 American Megatrends, Inc. System and method of performing high availability configuration and validation of virtual desktop infrastructure (vdi)
US9571411B1 (en) * 2014-12-29 2017-02-14 Juniper Networks, Inc. Using a firewall filter to select a member link of a link aggregation group
US20160191314A1 (en) * 2014-12-31 2016-06-30 Dell Products L.P. Multi-port selection and configuration
US20160212041A1 (en) * 2015-01-16 2016-07-21 Dell Products L.P. Switching layer 2 (l2) flows to a correct node in a vlt domain
US20160226704A1 (en) * 2015-02-04 2016-08-04 International Business Machines Corporation Port configuration for interconnected communications devices
US9661022B2 (en) * 2015-04-24 2017-05-23 Dell Products L.P. System and method for authorizing devices joining a network fabric
US20170034000A1 (en) * 2015-07-28 2017-02-02 Dell Products L.P. Inter-networking device link provisioning system
US20170054603A1 (en) * 2015-08-17 2017-02-23 Vmware, Inc. Hardware management systems for disaggregated rack architectures in virtual server rack deployments
US20170078150A1 (en) * 2015-09-16 2017-03-16 Brocade Communications Systems, Inc. Ip-based interconnection of switches with a logical chassis
US9497080B1 (en) * 2015-09-23 2016-11-15 International Business Machines Corporation Election and use of configuration manager
US20170118041A1 (en) * 2015-10-21 2017-04-27 Brocade Communications Systems, Inc. Distributed rule provisioning in an extended bridge
US20170126490A1 (en) * 2015-10-29 2017-05-04 Arista Networks, Inc. Method and system for configuring network devices
US10097457B1 (en) * 2015-12-28 2018-10-09 Juniper Networks, Inc. Resolving a mismatch among control plane parameter values received from multiple routing control devices
US20170214600A1 (en) * 2016-01-27 2017-07-27 Oracle International Corporation System and method for supporting scalable representation of switch port status in a high performance computing environment
US20170214595A1 (en) * 2016-01-27 2017-07-27 Oracle International Corporation System and method for supporting a scalable representation of link stability and availability in a high performance computing environment

Also Published As

Publication number Publication date
US10785118B2 (en) 2020-09-22

Similar Documents

Publication Publication Date Title
US10425287B2 (en) Systems and methods for network topology discovery
US10296328B2 (en) Infrastructure configuration and inventory manager
US9424148B2 (en) Automatic failover in modular chassis systems
US10949539B2 (en) Systems and methods for secure boot and runtime tamper detection
US9836309B2 (en) Systems and methods for in-situ fabric link optimization in a modular information handling system chassis
US20140208133A1 (en) Systems and methods for out-of-band management of an information handling system
US20170031694A1 (en) System and method for remote system configuration managment
US9891678B2 (en) Systems and methods for remotely resetting management controller via power over ethernet switch
US10467439B2 (en) Detecting tampering of memory contents in an information handling system
US20150358213A1 (en) Systems and methods for sharing a single firmware image in a chassis configured to receive a plurality of modular information handling systems
US10564218B2 (en) Systems and methods for debugging access
US11416446B2 (en) Systems and methods for remote management of non-standard devices
US10827005B2 (en) Systems and methods of group automation for multi-chassis management
US10855777B2 (en) Declarative security management plugins
US10579575B2 (en) Systems and methods of management console user interface pluggability
US10785118B2 (en) Systems and methods for network topology validation
US10397241B2 (en) Systems and methods for integration of directory service with management controllers
US10303568B2 (en) Systems and methods for high availability of management controllers
US10003463B2 (en) Systems and methods for revoking and replacing signing keys
US10348592B2 (en) Systems and methods for dynamic availability of executable endpoints
US20200021501A1 (en) Systems and methods for configuring network interface affinity to system management features supported by a management controller
US11088910B2 (en) Efficient method for managing and adding systems within a solution
US20230208718A1 (en) Network topology planning and validation
US11294840B2 (en) Dual-tree backplane
US20230216862A1 (en) Detection of on-premises systems

Legal Events

Date Code Title Description
AS Assignment

Owner name: DELL PRODUCTS L.P., TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SHETTY, SUDHIR VITTAL;SANE, SANJAY;RAJAGOPALAN, BALAJI;SIGNING DATES FROM 20170220 TO 20170223;REEL/FRAME:041362/0143

AS Assignment

Owner name: CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLATERAL AGENT, NORTH CAROLINA

Free format text: PATENT SECURITY INTEREST (CREDIT);ASSIGNORS:DELL PRODUCTS L.P.;EMC CORPORATION;EMC IP HOLDING COMPANY LLC;AND OTHERS;REEL/FRAME:042768/0585

Effective date: 20170526

Owner name: THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., AS COLLATERAL AGENT, TEXAS

Free format text: PATENT SECURITY INTEREST (NOTES);ASSIGNORS:DELL PRODUCTS L.P.;EMC CORPORATION;EMC IP HOLDING COMPANY LLC;AND OTHERS;REEL/FRAME:042769/0001

Effective date: 20170605

Owner name: CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH, AS COLLAT

Free format text: PATENT SECURITY INTEREST (CREDIT);ASSIGNORS:DELL PRODUCTS L.P.;EMC CORPORATION;EMC IP HOLDING COMPANY LLC;AND OTHERS;REEL/FRAME:042768/0585

Effective date: 20170526

Owner name: THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., A

Free format text: PATENT SECURITY INTEREST (NOTES);ASSIGNORS:DELL PRODUCTS L.P.;EMC CORPORATION;EMC IP HOLDING COMPANY LLC;AND OTHERS;REEL/FRAME:042769/0001

Effective date: 20170605

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

AS Assignment

Owner name: THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., T

Free format text: SECURITY AGREEMENT;ASSIGNORS:CREDANT TECHNOLOGIES, INC.;DELL INTERNATIONAL L.L.C.;DELL MARKETING L.P.;AND OTHERS;REEL/FRAME:049452/0223

Effective date: 20190320

Owner name: THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., TEXAS

Free format text: SECURITY AGREEMENT;ASSIGNORS:CREDANT TECHNOLOGIES, INC.;DELL INTERNATIONAL L.L.C.;DELL MARKETING L.P.;AND OTHERS;REEL/FRAME:049452/0223

Effective date: 20190320

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

AS Assignment

Owner name: THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., TEXAS

Free format text: SECURITY AGREEMENT;ASSIGNORS:CREDANT TECHNOLOGIES INC.;DELL INTERNATIONAL L.L.C.;DELL MARKETING L.P.;AND OTHERS;REEL/FRAME:053546/0001

Effective date: 20200409

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: WYSE TECHNOLOGY L.L.C., CALIFORNIA

Free format text: RELEASE OF SECURITY INTEREST AT REEL 042768 FRAME 0585;ASSIGNOR:CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH;REEL/FRAME:058297/0536

Effective date: 20211101

Owner name: MOZY, INC., WASHINGTON

Free format text: RELEASE OF SECURITY INTEREST AT REEL 042768 FRAME 0585;ASSIGNOR:CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH;REEL/FRAME:058297/0536

Effective date: 20211101

Owner name: EMC IP HOLDING COMPANY LLC, TEXAS

Free format text: RELEASE OF SECURITY INTEREST AT REEL 042768 FRAME 0585;ASSIGNOR:CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH;REEL/FRAME:058297/0536

Effective date: 20211101

Owner name: EMC CORPORATION, MASSACHUSETTS

Free format text: RELEASE OF SECURITY INTEREST AT REEL 042768 FRAME 0585;ASSIGNOR:CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH;REEL/FRAME:058297/0536

Effective date: 20211101

Owner name: DELL PRODUCTS L.P., TEXAS

Free format text: RELEASE OF SECURITY INTEREST AT REEL 042768 FRAME 0585;ASSIGNOR:CREDIT SUISSE AG, CAYMAN ISLANDS BRANCH;REEL/FRAME:058297/0536

Effective date: 20211101

AS Assignment

Owner name: DELL MARKETING CORPORATION (SUCCESSOR-IN-INTEREST TO WYSE TECHNOLOGY L.L.C.), TEXAS

Free format text: RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL/FRAME (042769/0001);ASSIGNOR:THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., AS NOTES COLLATERAL AGENT;REEL/FRAME:059803/0802

Effective date: 20220329

Owner name: EMC IP HOLDING COMPANY LLC (ON BEHALF OF ITSELF AND AS SUCCESSOR-IN-INTEREST TO MOZY, INC.), TEXAS

Free format text: RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL/FRAME (042769/0001);ASSIGNOR:THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., AS NOTES COLLATERAL AGENT;REEL/FRAME:059803/0802

Effective date: 20220329

Owner name: EMC CORPORATION, MASSACHUSETTS

Free format text: RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL/FRAME (042769/0001);ASSIGNOR:THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., AS NOTES COLLATERAL AGENT;REEL/FRAME:059803/0802

Effective date: 20220329

Owner name: DELL PRODUCTS L.P., TEXAS

Free format text: RELEASE OF SECURITY INTEREST IN PATENTS PREVIOUSLY RECORDED AT REEL/FRAME (042769/0001);ASSIGNOR:THE BANK OF NEW YORK MELLON TRUST COMPANY, N.A., AS NOTES COLLATERAL AGENT;REEL/FRAME:059803/0802

Effective date: 20220329

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4