Establish Call Admissions Control Status
A final consideration in developing a candidate list for handover is the ability of the new AP to support the phone’s traffic. This touches on Call Admissions Control, a way of restricting the number of voice calls on a particular AP before the capacity limit is reached. If the new AP is already at capacity, it will not be able to accept the voice call on handover.
While CAC will be enforced by the new AP towards the end of the handover, latency would be adversely affected were the phone to discover late in the proves that the new AP cannot carry its load. Thus it is important to indicate to the phone whether it is likely to be admitted by the new AP’s CAC, so it can eliminate overloaded APs from its candidate list.
Solutions to this problem include having APs publish their current load in the beacon, allowing the phone to make a decision, and an alternative method using the probe response (invented by Aruba). Aruba’s CAC implementation offers another way to avoid the problem: a reserve of capacity is allocated on every AP explicitly to allow handover of voice calls in progress to that AP.
CAC is even more complicated in a WMM environment, as capacity must be considered for each of the four traffic priorities. A solution for this is provided in the QBSS load element specified in the beacon for IEEE 802.11e, and in Aruba’s probe response: these give clients a view of current loads on the AP, providing an indication of whether the AP is likely to reject a handover attempt due to Call Admissions Control.
Decide it’s time to handover
In the 802.11model, it’s up to the client to decide when to move to a new AP. This is a surprisingly difficult decision: the primary indicators for the client are the signal strength and signal-to-noise ratio (SNR) of the current AP, but RF signal levels fluctuate, particularly in the case of a moving phone where it can be carried behind thick walls, metal objects such as bookcases, or even a human head as the caller turns around.
The difficulty for the phone is in distinguishing whether a sudden reduction in the signal power is a temporary event that will soon be reversed, or a sign that the phone is moving out of the coverage area of the current AP. In the former case, the correct behaviour would be to stay on the current AP, but in the latter such ‘sticky’ behaviour would result in longer handover latency, as the original AP’s signal could be lost completely before handover to a new one was complete.
Another factor is the data rate of the connection: in Aruba networks we normally set the minimum allowed data rate above the minimum possible in order to force a handover before the signal is impossibly weak: in an Enterprise Wi-Fi network there should be good, high-quality coverage throughout the building so there will always be a suitable AP to handover to.
Many phones today use a combination of factors to make the handover decision, based on the signal strength and SNR of the current and prospective APs but with various extensions and degrees of sophistication. They must also avoid rapid ‘flapping’ where clients shuttle rapidly between APs. While these algorithms are mostly the domain of the phone designer, Aruba has developed expertise in this area, and we continue to invent ways in which the WLAN infrastructure can provide the phone with relevant information to make a better decision.
Other work in the IEEE includes are submission to 802.11v for ‘directed handover’ which would allow the network to explicitly command a phone to handover to a designated new AP; Aruba supports this proposal.
Handover to the new AP
When a phone makes a decision to handover, the situation may already be grim. Phone designers have learned through experience that the penalty for moving prematurely is greater than that for making a decision late: therefore, if RF coverage is not uniform and good, the voice quality on the connection may already have deteriorated when the decision is made to handover.
Using the cached candidate list, the phone will choose what it considers the most likely candidate, and start the handover process. Since phones can only maintain association with one AP at a time, this is when the voice connection is interrupted, and the handover latency timer starts.
Authentication
The complexity and latency associated with authentication depends on the protocol used. Phones have historically lagged PC clients in their support of newer authentication regimes, but for this paper we will assume that technology has moved beyond WEP, and most phones are now (1st quarter of 2007) capable of 802.11i authentication, WPA or WPA2 using at least pre-shared keys. Successful authentication allows the phone to establish Pairwise Master Keys (PMKs) with the infrastructure.
The simplest (and worst, delay-wise) model assumes full re-authentication. In this case, an 802.1x authentication must begin from scratch, involving the exchange of many frames between the client, the authenticator (mobility controller in Aruba’s architecture) and the RADIUS server. This can take considerable time – several hundred msec was not unknown in early implementations – and can be further increased by WAN latency when exchanging frames with a RADIUS server at a remote site.
Even with full re-authentication, centralized architectures and centralized encryption (Aruba is the only vendor to use centralized encryption, but many use centralized traffic and control architectures) bring this time down to less than 100 msec in most cases. Recent features such as EAP offload can improve these numbers further.
Another improvement is to implement a part of the 802.11i/WPA2 specification: Opportunistic Key Caching. OKC offers a way for the phone to pre-associated with an AP before it needs to handover, establishing a PMK that can be cached for a period of time, usually some hours. Now, when the phone wishes to handover it can bypass the whole 802.1x re-authentication sequence by presenting its cached PMK. The authenticator will recognize the PMK and move straight to the next step, deriving session keys.
The ‘neighbour report’ defined in 802.11k draft can assist the phone again here. It includes a field called ‘key scope’ that identifies the switch (mobility controller in Aruba’s architecture) managing each AP. The phone needs to pre-authenticate only once per switch, greatly simplifying its work and reducing the load on the infrastructure.
Although OKC is defined only for WPA2, one of Aruba’s customers acquired a phone that did not support WPA2, but wished to see the benefits of lower handover latency by applying the OKC technique to the WPA protocol. Aruba complied, and we now support a (non-standard) version of OKC over WPA.
Derive session keys for the new AP
Once PMKs have been established, the phone and the infrastructure must exchange frames in order to derive session keys for unicast and multicast traffic. In WPA2 this is called the ‘four-way handshake’. This exchange is relatively short compared to 802.1x authentication, so it is not usually a significant factor in handover latency in centralized WLAN architectures.
However, the IEEE 802.11 standards group is working to improve performance everywhere and the 802.11r draft addresses this area. While centralized architectures will not benefit significantly, 802.11r will be an important improvement for ‘fat AP’ networks. 802.11r also deals explicitly with CAC and 802.11e (WMM) as part of handover, allowing the phone to present its existing traffic specification (TSpec) as part of the protocol.
After session keys have been exchanged, the phone is ready to communicate with the new AP: the L2 part of handover is complete.
Obtain a new IP address
The next stage would be for the phone to obtain a new IP address when the handover involves a move to another IP subnet (L3 roaming from the phone’s viewpoint). But there is a simple rule of network design: don’t do this. It takes a long time for a client to acquire a new IP address through DHCP, so Aruba and others go to great lengths to ensure it is not necessary. The following is a brief explanation of Aruba’s solutions that enable handover while avoiding the need for a new IP address for the client.
For the simplest handover case, the network designer builds a single IP subnet covering all APs where a phone may wish to handover (this is usually confined to an area of contiguous coverage such as a building or campus). Aruba terms this a ‘mobility domain’. In this case, once the phone has got its original IP address there is no reason to change it. Aruba adds an elegant twist to this, since all traffic is tunneled from the client through the AP to the mobility controller, so IP addresses can be maintained by the mobility controller rather than depending on the Ethernet connection to the AP.
There are well-known problems as subnets expand to hundreds of clients, so vendors (including Aruba) have developed features such as filtering multicast traffic and adding ARP proxy functions. Eventually, however, a layer 3 solution is needed. Centralized architectures with WLAN switches all use some form of the ‘mobile IP’ protocol for this, where incoming traffic is directed to a ‘home agent’ with the original IP address of the client. The HA redirects traffic to the ‘foreign agent’ which then forwards it to the client. All such architectures today use a ‘proxy MIP’ feature because phones do not support MIP clients.
Note that there is a penalty for setting up the HA-FA connection as a client hands over across a mobility controller L3 boundary: this can be of the order of 25 msec in an Aruba network.
Re-establish the media stream
While the WLAN infrastructure’s work is essentially complete at this point, the phone must still re-establish the media stream. In the transmit direction, this involves resuming the transmission of frames over the air. The codec has continued to generate frames independently of the handover event, so it may be helpful to cull the buffer in order to keep the delay on the media stream low.
In the receive direction, the incoming frames must be directed by the mobility controller to the new AP, and the phone usually holds the first 20-60 msec of the stream to establish its dejitter buffer.
Conclusion
Handover latency is a key performance measurement in Wi-Fi networks. This paper identified the various stages of a successful handover, and discussed the techniques used at each stage to minimize handover time.
It is clear that there are many aspects of handover where the Wi-Fi infrastructure can assist in reducing latency, but the particular client implementation is still a significant factor.