The Setup—>Hardware configuration page allows you to view and edit the physical interfaces. It also allows you to create a bonded interface. Note that bonding will only work for physical appliances. Bonding should be done on the host Hyper Visor for virtual appliances.
This section is organised in two sections:
The settings on this screen control the network access. The defaults are to fix speed at 100 Mbps and full duplex. This avoids any issue with certain networking devices that have auto-negotiation which re-negotiates too frequently.
The appliance can support speeds from 10 to 1000; for 1000 this should set to auto/auto. If this does not work, set the exact network hardware values.
The speed and duplex setting should only be changed for hardware appliance. Virtual appliances will take their configuration from the underlying hypervisor.
Bonding allows you to aggregate multiple ports into a single group, effectively combining the bandwidth into a single connection. Bonding also allows you to create multi-gigabit pipes to transport traffic through the highest traffic areas of your network. Note: this is only relevant for your hardware version of ALB-X. Do not use bonding for the Virtual Appliance.
Transmits packets in sequential order from the first available slave to the last.
Has one interface which will be active and the second interface will be in standby. This secondary interface only becomes active if the active connection on the first interface fails.
Transmits based on source MAC address XOR’d with destination MAC address. This selects the same slave for each destination Mac address.
Transmits everything on all slave interfaces.
Creates aggregation groups that share the same speed and duplex settings. Utilizes all slaves in the active aggregator according to the 802.3ad specification.
The Adaptive transmit load balancing bonding mode: Provides channel bonding that does not require any special switch support. The outgoing traffic is distributed according to the current load (computed relative to the speed) on each slave. Incoming traffic is received by the current slave. If the receiving slave fails, another slave takes over the MAC address of the failed receiving slave.
The Adaptive load balancing bonding mode: also includes balance-tlb plus receive load balancing (rlb) for IPV4 traffic, and does not require any special switch support. The receive load balancing is achieved by ARP negotiation. The bonding driver intercepts the ARP Replies sent by the local system on their way out and overwrites the source hardware address with the unique hardware address of one of the slaves in the bond, such that different peers use different hardware addresses for the server.