network overlays

WAN Virtualization

WAN Virtualization

In today's fast-paced digital world, seamless connectivity is the key to success for businesses of all sizes. WAN (Wide Area Network) virtualization has emerged as a game-changing technology, revolutionizing the way organizations connect their geographically dispersed branches and remote employees. In this blog post, we will explore the concept of WAN virtualization, its benefits, implementation considerations, and its potential impact on businesses.

WAN virtualization is a technology that abstracts the physical network infrastructure, allowing multiple logical networks to operate independently over a shared physical infrastructure. It enables organizations to combine various types of connectivity, such as MPLS, broadband, and cellular, into a single virtual network. By doing so, WAN virtualization enhances network performance, scalability, and flexibility.

Increased Flexibility and Scalability: WAN virtualization allows businesses to scale their network resources on-demand, facilitating seamless expansion or contraction based on their requirements. It provides flexibility to dynamically allocate bandwidth, prioritize critical applications, and adapt to changing network conditions.

Improved Performance and Reliability:By leveraging intelligent traffic management techniques and load balancing algorithms, WAN virtualization optimizes network performance. It intelligently routes traffic across multiple network paths, avoiding congestion and reducing latency. Additionally, it enables automatic failover and redundancy, ensuring high network availability.

Simplified Network Management:Traditional WAN architectures often involve complex configurations and manual provisioning. WAN virtualization simplifies network management by centralizing control and automating tasks. Administrators can easily set policies, monitor network performance, and make changes from a single management interface, saving time and reducing human errors.

Multi-Site Connectivity: For organizations with multiple remote sites, WAN virtualization offers a cost-effective solution. It enables seamless connectivity between sites, allowing efficient data transfer, collaboration, and resource sharing. With centralized management, network administrators can ensure consistent policies and security across all sites. Cloud Connectivity:

As more businesses adopt cloud-based applications and services, WAN virtualization becomes an essential component. It provides reliable and secure connectivity between on-premises infrastructure and public or private cloud environments. By prioritizing critical cloud traffic and optimizing routing, WAN virtualization ensures optimal performance for cloud-based applications.

Highlights: WAN Virtualization

Understanding Virtualization

Virtualization is a technology that allows for the creation of virtual versions of various IT resources, such as servers, networks, and storage devices. These virtual resources operate independently from physical hardware, enabling multiple operating systems and applications to run simultaneously on a single physical machine. Virtualization opens up a world of possibilities by breaking the traditional one-to-one relationship between hardware and software. Now, virtualization has moved to the WAN.

WAN Virtualization and SD-WAN

Organizations constantly seek innovative solutions in modern networking to enhance their network infrastructure and optimize connectivity. One such solution that has gained significant attention is WAN virtualization. In this blog post, we will delve into the concept of WAN virtualization, its benefits, and how it revolutionizes how businesses connect and communicate.

WAN virtualization, also known as Software-Defined WAN (SD-WAN), is a technology that enables organizations to abstract their wide area network (WAN) connections from the underlying physical infrastructure. It leverages software-defined networking (SDN) principles to decouple network control and data forwarding, providing a more flexible, scalable, and efficient network solution.

SD-WAN Highlights

SD-WAN, in essence, is a virtualized approach to wide-area networking. It leverages software-defined networking principles to simplify managing and operating a wide area network, connecting geographically dispersed locations. Unlike traditional networks, SD-WAN offers centralized control, automated traffic management, and enhanced security.

sd-wan technology

VPN and SDN Components

So, what is WAN virtualization? WAN virtualization is an essential technology in the modern business world. It creates virtualized versions of wide area networks (WANs) – networks spanning a wide geographic area. The virtualized WANs can then manage and secure a company’s data, applications, and services.

Regarding implementation, WAN virtualization requires using a virtual private network (VPN), a secure private network accessible only by authorized personnel. This ensures that only those with proper credentials can access the data. WAN virtualization also requires software-defined networking (SDN) to manage the network and its components.

Related: Before you proceed, you may find the following posts helpful:

  1. SD WAN Overlay
  2. Generic Routing Encapsulation
  3. WAN Monitoring
  4. SD WAN Security 
  5. Container Based Virtualization
  6. SD WAN and Nuage Networks



WAN Virtualization

Key WAN Virtualization Discussion Points:


  • Introduction to WAN Virtualization and what is involved.

  • Highlighting the issues around internet traffic left to its defaults.

  • Critical points on WAN utilization problems.

  • Technical details on routing protocol convergence.

  • Technical details on SD WAN Overlay and how this changes the WAN.

Back to Basics: WAN virtualization.

WAN Challenges

Deploying and managing the Wide Area Network (WAN) has become more challenging. Engineers face several design challenges, such as traffic flow decentralizing, inefficient WAN link utilization, routing protocol convergence, and application performance issues with active-active WAN edge designs. Active-active WAN designs that spray and pray over multiple active links present technical and business challenges.

To do this efficiently, you have to understand application flows. There may also be performance problems. When packets get to the other end, there may be out-of-order packets as each link propagates at different speeds. The remote end has to be reassembled and put back together, causing jitter and delay. Both high jitter and delay are bad for network performance. To recap on WAN virtualization, including the drivers for SD-WAN, you may follow this SD WAN tutorial.

What is WAN Virtualization
Diagram: What is WAN virtualization? Source Linkedin.

Knowledge Check: Cisco PfR

Cisco PfR is an intelligent routing solution that dynamically optimizes traffic flow within a network. Unlike traditional routing protocols, PfR makes real-time decisions based on network conditions, application requirements, and business policies. By monitoring various metrics such as delay, packet loss, and link utilization, PfR intelligently determines the best path for traffic.

Key Features and Functionalities

PfR offers many features and functionalities that significantly enhance network performance. Some notable features include:

1. Intelligent Path Control: PfR selects the optimal traffic path based on performance metrics, ensuring efficient utilization of network resources.

2. Application-Aware Routing: PfR considers the specific requirements of different applications and dynamically adjusts routing to provide the best user experience.

3. Load Balancing: By distributing traffic across multiple paths, PfR improves network efficiency and avoids bottlenecks.

Performance based routing

Knowledge Check: Control and Data Plane

Understanding the Control Plane

The control plane can be likened to a network’s brain. It is responsible for making high-level decisions and managing network-wide operations. From routing protocols to network management systems, the control plane ensures data is directed along the most optimal paths. By analyzing network topology, the control plane determines the best routes to reach a destination and establishes the necessary rules for data transmission.

Unveiling the Data Plane

In contrast to the control plane, the data plane focuses on the actual movement of data packets within the network. It can be thought of as the hands and feet executing the control plane’s instructions. The data plane handles packet forwarding, traffic classification, and Quality of Service (QoS) enforcement tasks. It ensures that data packets are correctly encapsulated, forwarded to their intended destinations, and delivered with the necessary priority and reliability.

Use Cases and Deployment Scenarios

Distributed Enterprises

For organizations with multiple branch locations, WAN virtualization offers a cost-effective solution for connecting remote sites to the central network. It allows for secure and efficient data transfer between branches, enabling seamless collaboration and resource sharing.

Cloud Connectivity

WAN virtualization is ideal for enterprises adopting cloud-based services. It provides a secure and optimized connection to public and private cloud environments, ensuring reliable access to critical applications and data hosted in the cloud.

Disaster Recovery and Business Continuity

WAN virtualization plays a vital role in disaster recovery strategies. Organizations can ensure business continuity during a natural disaster or system failure by replicating data and applications across geographically dispersed sites.

Challenges and Considerations

Implementing WAN virtualization requires careful planning and consideration. Factors such as network security, bandwidth requirements, and compatibility with existing infrastructure need to be evaluated. It is essential to choose a solution that aligns with the specific needs and goals of the organization.

SD-WAN vs. DMVPN

Two popular WAN solutions are DMVPN and SD-WAN.

DMVPN (Dynamic Multipoint Virtual Private Network) and SD-WAN (Software-Defined Wide Area Network) are popular solutions to improve connectivity between distributed branch offices. DMVPN is a Cisco-specific solution, and SD-WAN is a software-based solution that can be used with any router. Both solutions provide several advantages, but there are some differences between them.

DMVPN is a secure, cost-effective, and scalable network solution that combines underlying technologies and DMVVPN phases (for example, the traditional DMVPN phase 1 ) to connect multiple sites. It allows the customer to use existing infrastructure and provides easy deployment and management. This solution is an excellent choice for businesses with many branch offices because it allows for secure communication and the ability to deploy new sites quickly.

DMVPN and WAN Virtualization

SD-WAN is a software-based solution that is gaining popularity in the enterprise market. It provides improved application performance, security, and network reliability. SD-WAN is an excellent choice for businesses that require high-performance applications across multiple sites. It provides an easy-to-use centralized management console that allows companies to deploy new sites and manage the network quickly.

Dynamic Multipoint VPN
Diagram: Example with DMVPN. Source is Cisco

1st Lab Guide: DMVPN operating over the WAN

The following shows DMVPN operating over the WAN. The SP node represents the WAN network. Then we have R11 as the hub and R2, R3 as the spokes.  Several protocols make the DMVPM network over the WAN possible. We have GRE; in this case, the tunnel destination is specified as a point-to-point GRE tunnel instead of a mGRE tunnel.

Then we have NHRP, which is used to help create a mapping as this is a nonbroadcast network; we can not use ARP. So, we need to manually set this up on the spokes with the command: ip nhrp NHS 192.168.100.11

DMVPN configuration
Diagram: DMVPN Configuration.

Shift from network-centric to business intent.

The core of WAN virtualization involves shifting focus from a network-centric model to a business intent-based WAN network. So, instead of designing the WAN for the network, we can create the WAN for the application. This way, the WAN architecture can simplify application deployment and management.

First, however, the mindset must shift from a network topology focus to an application services topology. A new application style consumes vast bandwidth and is very susceptible to variations in bandwidth quality. Things such as jitter, loss, and delay impact most applications, which makes it essential to improve the WAN environment for these applications.

wan virtualization
Diagram: WAN virtualization.

The spray-and-pray method over two links increases bandwidth but decreases “goodput.” It also affects firewalls, as they will see asymmetric routes. When you want an active-active model, you need application session awareness and a design that eliminates asymmetric routing. It would help if you could slice the WAN properly so application flows can work efficiently over either link.

What is WAN Virtualization: Decentralizing Traffic

Decentralizing traffic from the data center to the branch requires more bandwidth to the network’s edges. As a result, we see many high-bandwidth applications running on remote sites. This is what businesses are now trying to accomplish. Traditional branch sites usually rely on hub sites for most services and do not host bandwidth-intensive applications. Today, remote locations require extra bandwidth, which is not cheaper yearly.

Inefficient WAN utilization

Redundant WAN links usually require a dynamic routing protocol for traffic engineering and failover. Routing protocols require complex tuning to load balance traffic between border devices. Border Gateway Protocol (BGP) is the primary protocol for connecting sites to external networks.

It relies on path attributes to choose the best path based on availability and distance. Although these attributes allow granular policy control, they do not cover aspects relating to path performance, such as Round Trip Time (RTT), delay, and jitter.

Port 179
Furthermore, BGP does not always choose the “best” path, which may have different meanings for customers. For example, customer A might consider the path via provider A as the best due to the price of links. Default routing does not take this into account. Packet-level routing protocols are not designed to handle the complexities of running over multiple transport-agnostic links. Therefore, a solution that eliminates the need for packet-level routing protocols must arise.
BGP Path Attributes
Diagram: BGP Path Attributes Source is Cisco.

Routing protocol convergence

WAN designs can also be active standby, which requires routing protocol convergence in the event of primary link failure. However, routing convergence is slow, and to speed up, additional features, such as Bidirectional Forwarding Detection (BFD), are implemented that may stress the network’s control plane. Although mechanisms exist to speed up convergence and failure detection, there are still several convergence steps, such as:

Rouitng Convergence

Convergence


Detect


Describe


Switch 


Find

 Branch office security

With traditional network solutions, branches connect back to the data center, which typically provides Internet access. However, the application world has evolved, and branches directly consume applications such as Office 365 in the cloud. This drives a need for branches to access these services over the Internet without going to the data center for Internet access or security scrubbing.

Extending the security diameter into the branches should be possible without requiring onsite firewalls / IPS and other security paradigm changes. A solution must exist that allows you to extend your security domain to the branch sites without costly security appliances at each branch—essentially, building a dynamic security fabric.

WAN Virtualization

The solution to all these problems is SD-WAN ( software-defined WAN ). SD-WAN is a transport-independent overlay software-based networking deployment. It uses software and cloud-based technologies to simplify the delivery of WAN services to branch offices. Similar to Software Defined Networking (SDN), SD-WAN works by abstraction. It abstracts network hardware into a control plane with multiple data planes to make up one large WAN fabric.

 SD-WAN in a nutshell 

When we consider the Wide Area Network (WAN) environment at a basic level, we connect data centers to several branch offices to deliver packets between those sites, supporting the transport of application transactions and services. The SD-WAN platform allows you to pull Internet connectivity into those sites, becoming part of one large transport-independent WAN fabric.

SD-WAN monitors the paths and the application performance on each link (Internet, MPLS, LTE ) and chooses the best path based on performance.

There are many forms of Internet connectivity (cable, DSL, broadband, and Ethernet). They are quick to deploy at a fraction of the cost of private MPLS circuits. SD-WAN provides the benefit of using all these links and monitoring which applications are best for them.

Application performance is continuously monitored across all eligible paths-direct internet, internet VPN, and private WAN. It creates an active-active network and eliminates the need to use and maintain traditional routing protocols for active-standby setups—no reliance on the active-standby model and associated problems.

WAN virtualization
Diagram: WAN virtualization. Source is Juniper

SD-WAN simplifies WAN management

SD-WAN simplifies managing a wide area network by providing a centralized platform for managing and monitoring traffic across the network. This helps reduce the complexity of managing multiple networks, eliminating the need for manual configuration of each site. Instead, all of the sites are configured from a single management console.

SD-WAN also provides advanced security features such as encryption and firewalling, which can be configured to ensure that only authorized traffic is allowed access to the network. Additionally, SD-WAN can optimize network performance by automatically routing traffic over the most efficient paths.

what is wan virtualization

SD-WAN Packet Steering

SD-WAN packet steering is a technology that efficiently routes packets across a wide area network (WAN). It is based on the concept of steering packets so that they can be delivered more quickly and reliably than traditional routing protocols. Packet steering is crucial to SD-WAN technology, allowing organizations to maximize their WAN connections.

SD-WAN packet steering works by analyzing packets sent across the WAN and looking for patterns or trends. Based on these patterns, the SD-WAN can dynamically route the packets to deliver them more quickly and reliably. This can be done in various ways, such as considering latency and packet loss or ensuring the packets are routed over the most reliable connections.

Spraying packets down both links can result in 20% drops or packet reordering. SD-WAN makes packets better utilized, no reorder, and better “goodput.” SD-WAN increases your buying power and results in buying lower bandwidth links and running them more efficiently. Over-provision is unnecessary as you are using the existing WAN bandwidth better.

Knowledge Check: Application-Aware Routing (AAR)

Understanding Application-Aware Routing (AAR)

Application-Aware Routing is a sophisticated networking technique that goes beyond traditional packet-based routing. It considers the unique requirements of different applications, such as video streaming, cloud-based services, or real-time communication, and optimizes the network path accordingly. By prioritizing and steering traffic based on application characteristics, it ensures smooth and efficient data transmission.

Benefits of Application-Aware Routing

2.1 Enhanced Performance: Application-aware routing significantly improves overall performance by dynamically allocating network resources to applications with high bandwidth or low latency requirements. This translates into faster downloads, seamless video streaming, and reduced response times for critical applications.

2.2 Increased Reliability: Traditional routing methods treat all traffic equally, often resulting in congestion and potential bottlenecks. Application Aware Routing intelligently distributes network traffic, avoiding congested paths and ensuring a reliable and consistent user experience. In network failure or congestion, it can dynamically reroute traffic to alternative paths, minimizing downtime and disruptions.

Implementation Strategies

Deep Packet Inspection: A key component of Application-Aware Routing is deep packet inspection (DPI), which analyzes the content of network packets to identify specific applications. DPI enables routers and switches to make informed decisions about handling each packet based on its application, ensuring optimal routing and resource allocation.

3.2 Quality of Service (QoS) Configuration: Implementing QoS parameters alongside Application Aware Routing allows network administrators to allocate bandwidth, prioritize specific applications over others, and enforce policies to ensure the best possible user experience. QoS configurations can be customized based on organizational needs and application requirements.

Future Possibilities

As the digital landscape continues to evolve, the potential for Application-Aware Routing is boundless. With emerging technologies like the Internet of Things (IoT) and 5G networks, the ability to intelligently route traffic based on specific application needs will become even more critical. Application-Aware Routing has the potential to optimize resource utilization, enhance security, and support the seamless integration of diverse applications and services.

Benefits of WAN Virtualization:

1. Enhanced Network Performance: WAN virtualization allows organizations to optimize network performance by intelligently routing traffic across multiple WAN links. Organizations can achieve improved application performance and reduced latency by dynamically selecting the most efficient path based on real-time network conditions.

2. Cost Savings: Traditional WAN solutions often require expensive dedicated circuits for each branch office. With WAN virtualization, organizations can leverage cost-effective internet connections, such as broadband or LTE, while ensuring secure and reliable connectivity. This flexibility in choosing connectivity options can significantly reduce operational costs.

3. Simplified Network Management: WAN virtualization provides centralized management and control of the entire network infrastructure. This simplifies network provisioning, configuration, and monitoring, reducing traditional WAN deployments’ complexity and administrative overhead.

4. Increased Scalability: WAN virtualization offers the scalability to accommodate evolving network requirements as organizations grow and expand their operations. It allows for the seamless integration of new branch offices and additional bandwidth without significant infrastructure changes.

5. Enhanced Security: With the rise in cybersecurity threats, network security is paramount. WAN virtualization enables organizations to implement robust security measures, such as encryption and firewall policies, across the entire network. This helps protect sensitive data and ensures compliance with industry regulations.

  • A final note on what is WAN virtualization

Server virtualization and automation in the data center are prevalent, but WANs are stalling in this space. It is the last bastion of hardware models that has complexity. Like hypervisors have transformed data centers, SD-WAN aims to change how WAN networks are built and managed. When server virtualization and hypervisor came along, we did not have to worry about the underlying hardware. Instead, a virtual machine (VM) can be provided and run as an application. Today’s WAN environment requires you to manage details of carrier infrastructure, routing protocols, and encryption. 

  • SD-WAN pulls all WAN resources together and slices up the WAN to match the applications on them.

The Role of WAN Virtualization in Digital Transformation:

In today’s digital era, where cloud-based applications and remote workforces are becoming the norm, WAN virtualization is critical in enabling digital transformation. It empowers organizations to embrace new technologies, such as cloud computing and unified communications, by providing secure and reliable connectivity to distributed resources.

Summary: WAN Virtualization

In our ever-connected world, seamless network connectivity is necessary for businesses of all sizes. However, traditional Wide Area Networks (WANs) often fall short of meeting the demands of modern data transmission and application performance. This is where the concept of WAN virtualization comes into play, promising to revolutionize network connectivity like never before.

Understanding WAN Virtualization

WAN virtualization, also known as Software-Defined WAN (SD-WAN), is a technology that abstracts the physical infrastructure of traditional WANs and allows for centralized control, management, and optimization of network resources. By decoupling the control plane from the underlying hardware, WAN virtualization enables organizations to dynamically allocate bandwidth, prioritize critical applications, and ensure optimal performance across geographically dispersed locations.

The Benefits of WAN Virtualization

Enhanced Flexibility and Scalability

With WAN virtualization, organizations can effortlessly scale their network infrastructure to accommodate growing business needs. The virtualized nature of the WAN allows for easy addition or removal of network resources, enabling businesses to adapt to changing requirements without costly hardware upgrades.

Improved Application Performance

WAN virtualization empowers businesses to optimize application performance by intelligently routing network traffic based on application type, quality of service requirements, and network conditions. By dynamically selecting the most efficient path for data transmission, WAN virtualization minimizes latency, improves response times, and enhances overall user experience.

Cost Savings and Efficiency

By leveraging WAN virtualization, organizations can reduce their reliance on expensive Multiprotocol Label Switching (MPLS) connections and embrace more cost-effective broadband links. The ability to intelligently distribute traffic across diverse network paths enhances network redundancy and maximizes bandwidth utilization, providing significant cost savings and improved efficiency.

Implementation Considerations

Network Security

When adopting WAN virtualization, it is crucial to implement robust security measures to protect sensitive data and ensure network integrity. Encryption protocols, threat detection systems, and secure access controls should be implemented to safeguard against potential security breaches.

Quality of Service (QoS)

Organizations should prioritize critical applications and allocate appropriate bandwidth resources through Quality of Service (QoS) policies to ensure optimal application performance. By adequately configuring QoS settings, businesses can guarantee mission-critical applications receive the necessary network resources, minimizing latency and providing a seamless user experience.

Real-World Use Cases

Global Enterprise Networks

Large multinational corporations with a widespread presence can significantly benefit from WAN virtualization. These organizations can achieve consistent performance across geographically dispersed locations by centralizing network management and leveraging intelligent traffic routing, improving collaboration and productivity.

Branch Office Connectivity

WAN virtualization simplifies connectivity and network management for businesses with multiple branch offices. It enables organizations to establish secure and efficient connections between headquarters and remote locations, ensuring seamless access to critical resources and applications.

Conclusion

In conclusion, WAN virtualization represents a paradigm shift in network connectivity, offering enhanced flexibility, improved application performance, and cost savings for businesses. By embracing this transformative technology, organizations can unlock the true potential of their networks, enabling them to thrive in the digital age.

IPsec Fault Tolerance

IPsec Fault Tolerance

IPSec Fault Tolerance

In today's interconnected world, network security is of utmost importance. One widely used protocol for securing network communications is IPsec (Internet Protocol Security). However, even the most robust security measures can encounter failures, potentially compromising the integrity of your network. In this blog post, we will explore the concept of fault tolerance in IPsec and how you can ensure the utmost security and reliability for your network.

IPsec is a suite of protocols used to establish secure connections over IP networks. It provides authentication, encryption, and integrity verification of data packets, ensuring secure communication between network devices. However, despite its strong security features, IPsec can still encounter faults that may disrupt the secure connections. Understanding these faults is crucial in implementing fault tolerance measures.

To ensure fault tolerance, it's important to be aware of potential vulnerabilities and common faults that can occur in an IPsec implementation. This section will discuss common faults such as key management issues, misconfigurations, and compatibility problems with different IPsec implementations. By identifying these faults, you can take proactive steps to mitigate them and enhance the fault tolerance of your IPsec setup.

To ensure fault tolerance, redundancy and load balancing techniques can be employed. Redundancy involves having multiple IPsec gateways or VPN concentrators that can take over in case of a failure. Load balancing distributes traffic across multiple gateways to optimize performance and prevent overload. This section will delve into the implementation of redundancy and load balancing strategies, including failover mechanisms and dynamic routing protocols.

To maintain fault tolerance, it is crucial to have effective monitoring and alerting systems in place. These systems can detect anomalies, failures, or potential security breaches in real-time, allowing for immediate response and remediation. This section will explore various monitoring tools and techniques that can help you proactively identify and address faults, ensuring the continuous secure operation of your IPsec infrastructure.

In conclusion, IPsec fault tolerance plays a vital role in ensuring the security and reliability of your network. By understanding common faults, implementing redundancy and load balancing, and employing robust monitoring and alerting systems, you can enhance the fault tolerance of your IPsec setup. Safeguarding your network with confidence becomes a reality when you take proactive steps to mitigate potential faults and continuously monitor your IPsec infrastructure.

Highlights: IPSec Fault Tolerance

Highlighting IPsec

IPsec is a secure network protocol used to encrypt and authenticate data over the internet. It is a critical part of any organization’s secure network infrastructure, and it is essential to ensure fault tolerance. Optimum end-to-end IPsec networks require IPsec fault tolerance in several areas for ingress and egress traffic flows. Key considerations must include asymmetric routing, where a packet traverses from a source to a destination in one path and takes a different path when it returns to the source.

Reverse Route Injection

Potential options include Reverse Route Injection (RRI), which can inject static routes automatically into the routing process for networks and hosts protected by a remote tunnel endpoint. Design each element with redundancy in mind. Failure components include the Backbone network, Access links, and IPsec gateway.

Site to Site VPN

Link Fault Tolerance

To ensure that VPN data networks provide a reliable service to users and their applications, they must meet a number of requirements. In this section, we will discuss how to design fault-tolerant networks. Fault-tolerant VPNs are resilient to changes in routing paths caused by hardware, software, or path failures between VPN ingress and egress points, including VPN access.

One of the primary rules of fault-tolerant network design is that there is no such thing as a cookie-cutter solution for all networks. However, the network’s goals and objectives dictate VPN fault-tolerant design principles. There are many cases in which economic factors influence the design more than technical considerations. Fault-tolerant IPSec VPN networks are also designed according to what faults they must be able to withstand

Backbone Network Fault Tolerance

In an IPSec VPN, the backbone network can be the public Internet, a private Layer 2 network, or an IP network of a single service provider. An organization other than the owner of the IPSec VPN may own and operate this network. A fault-tolerant network is usually built to withstand link and IP routing failures. The IP packet-routing functions the backbone provides are inherently used by IPSec protocols for transport. Often, IPsec VPN designers are unable to control IP fault tolerance on the backbone.

For additional pre-information, you may find the following helpful

  1. SD WAN SASE
  2. VPNOverview
  3. Dead Peer Detection
  4. What Is Generic Routing Encapsulation
  5. Routing Convergence

Back to basics with IPsec Fault tolerance

Concept of IPsec

Internet Protocol Security (IPsec) is a set of protocols to secure communications over an IP network. It provides authentication, integrity, and confidentiality of data transmitted over an IP network. IPsec establishes a secure tunnel between two endpoints, allowing data to be transmitted securely over the Internet. In addition, IPsec provides security by authenticating and encrypting each packet of data that is sent over the tunnel.

IPsec is typically used in Virtual Private Network (VPN) connections to ensure secure data sent over the Internet. It can also be used for tunneling to connect two remote networks securely. IPsec is an integral part of ensuring the security of data sent over the Internet and is often used in conjunction with other security measures such as firewalls and encryption.

IPsec VPN
Diagram: IPsec VPN. Source Wikimedia.

IPsec session

Several components exist that are used to create and maintain an IPsec session. By integrating these components, we get the required security services that protect the traffic for unauthorized observers. IPsec establishes tunnels between endpoints; these can also be described as peers. The tunnel can be protected by various means, such as integrity and confidentiality.

IPsec provides security services using two protocols, the Authentication Header and Encapsulating Security Payload. Both protocols use cryptographic algorithms for authenticated integrity services; Encapsulation Security Payload provides encryption services in combination with authenticated integrity.

  • A key point: Lab on IPsec between two ASAs. Site to Site IKEv1

In this lab, we will look at site-to-site IKEv1. Site-to-site IPsec VPNs are used to “bridge” two distant LANs together over the Internet.  So, we want IP reachability for R1 and R2, which are in the INSIDE interfaces of their respective ASAs. Generally, on the LAN, we use private addresses, so the two LANs cannot communicate without tunneling.

This lesson will teach you how to configure IKEv1 IPsec between two Cisco ASA firewalls to bridge two LANs. In the diagram below, you will see we have two ASAs. ASA1 and ASA2 are connected using their G0/1 interfaces to simulate the outside connection, which in the real world would be the WAN.

This is also set to the “OUTSIDE” security zone, so imagine this is their Internet connection. Each ASA has a G0/0 interface connected to the “INSIDE” security zone. R1 is on the network 192.168.1.0/24, while R2 is in 192.168.2.0/24. The goal of this lesson is to ensure that R1 and R2 can communicate with each other through the IPsec tunnel.

Site to Site VPN

IPsec and DMVPN

DMVPN builds tunnels between locations as needed, unlike IPsec VPN tunnels that are hard coded. As with SD-WAN, it uses standard routers without additional features. However, unlike hub-and-spoke networks, DMVPN tunnels are mesh networks. Organizations can choose from three basic DMVPN topologies when implementing a DMVPN network.

The first topology is the hub-and-spoke topology. The second topology is the Fully Masked topology. Finally, the third topology is the hub-and-spoke with Partial Mesh topology. To create these DMVPN topologies, we have phases, such as DMVPN Phase 3, that are the most flexible, enabling a pull mesh of on-demand tunnels that can use IPsec for security.

Concept of Reverse Routing Injection (RRI)

For network and host endpoints protected by a remote tunnel endpoint, reverse route injection (RRI) allows static routes to be automatically injected into the routing process. These protected hosts and networks are called remote proxy identities.

The next hop to the remote proxy network and mask is the remote tunnel endpoint, and each route is created based on these parameters. Traffic is encrypted using the remote Virtual Private Network (VPN) router as the next hop.

Static routes are created on the VPN router and propagated to upstream devices, allowing them to determine the appropriate VPN router to send returning traffic to maintain IPsec state flows. When multiple VPN routers provide load balancing or failover, or remote VPN devices cannot be accessed via a default route, choosing the right VPN router is crucial. Global routing tables or virtual route forwarding tables (VRFs) are used to create routes.

IPsec fault tolerance
Diagram: IPsec fault tolerance with multiple areas to consider.

The Networks Involved

Backbone network

IPsec uses an underlying backbone network for endpoint connectivity. It does not deploy its underlying packet-forwarding mechanism and relies on backbone IP packet-routing functions. Usually, the backbone is controlled by a 3rd-party provider, ensuring IPsec gateways trust redundancy and high availability methods applied by separate administrative domains.

Access link 

Adding a second link to terminate IPsec sessions and enabling both connections for IPsec termination improves redundant architectures. However, access link redundancy requires designers to deploy either Multiple IKE identities or Single IKE identities. Multiple IKE identity design involves two different peer IP addresses, one peer for each physical access link. The IKE identity of the initiator is derived from the source IP of the initial IKE message, and this will remain the same. Single IKE identity involves one peer neighbor, potentially terminating on a logical loopback address.

Physical interface redundancy

Design physical interface redundancy by terminating IPsec on logical interfaces instead of multiple physical interfaces. Useful when the router has multiple exit points, and you do not want the other side to use multiple peers’ addresses. A single IPsec session is terminating on loopback instead of multiple IPsec sessions terminating on physical interfaces. You still require the crypto map configured on two physical interfaces. Issue the command to terminate IPsec on the loopback: “crypto map VPN local-address lo0.”

  • A key point: Link failure

Phase 1 and 2 do not converge in the event of a single physical link failure. Convergence is based on an underlying network routing protocol. No IKE convergence occurs if one of the physical interfaces goes down.

Asymmetric Routing

Asymmetric routing may occur in multipath environments. For example, in the diagram below, traffic leaves spoke A, creating an IPsec tunnel to interface Se1/1:0 on Hub A. Asymmetric routing occurs when return traffic flows via Se0:0. The effect is a new IPsec SA between Se0:0 and Spoke A, introducing additional memory usage on peers. Overcome this with a proper routing mechanism and IPsec state replication ( discussed later ).

Asymmetric routing
Diagram: Asymmetric routing.

Design to ensure routing protocol convergence does not take longer than IKE dead peer detection. Routing protocols should not introduce repeated disruptions to IPsec processes. If you have control of the underlying routing protocol, deploy fast convergence techniques so that routing protocols converge faster than IKE detects a dead peer.

IPsec Fault Tolerance and IPsec Gateway

A redundant gateway involves a second IPsec gateway in standby mode. It does not have any IPsec state or replicate IPsec information between peers. Because either gateway may serve as an active gateway for spoke return traffic, you may experience asymmetric traffic flows. Also, due to the failure of the hub peer gateway, all traffic between sites drops until IKE and IPSec SAs are rebuilt on the standby peer.

Routing mechanism at gateway nodes

A common approach to overcome asymmetric routing is to deploy a routing mechanism at gateway nodes. IPsec’s high availability can be incorporated with HSRP, which pairs two devices with a single VIP address. VIP address terminates IPsec tunnel. HSRP and IPsec work perfectly fine as long as the traffic is symmetric.

Asymmetric traffic occurs when the return traffic does not flow via the active HSRP device. To prevent this, enable HSRP on the other side of IPsec peers, resulting in Front-end / Back-end HSRP design model. Or deploy Reverse Route Injection ( RRI ), and static routes are injected only by active IPsec peer. You no longer need Dead Peer Detection ( DPD ) as you use VIP for IPsec termination. In the event of a node failure, the IPsec peer does not change. A different method to resolve the asymmetric problem is implementing Reverse Route Injection. 

Reverse Route Injection
Diagram: Routing mechanisms and Reverse Route Injection.

Reverse Route Injection (RRI)

RRI is a method that synchronizes return routes for the spoke to the active gateway. The idea behind RRI is to make routing decisions that are dependent on the IPsec state. For end-to-end reachability, a route to a “secure” subnet must exist with a valid network hop. RRI inserts a route to the “secure” subnet in the RIB and associates it with an IPsec peer. Then, it injects based on the Proxy ACL; matches the destination address in the proxy ACL.

  •  RRI injects a static route for the upstream network.

 HSRPs’ or RRI IPsec is limited because it does not carry any state between the two IPsec peers. A better high-availability solution is to have state ( Security Association Database ) between the two gateways, offering stateful failover.

Implementing IPsec Fault Tolerance:

1. Redundant VPN Gateways: Deploying multiple VPN gateways in a high-availability configuration is fundamental to achieving IPsec fault tolerance. These gateways work in tandem, with one as the primary gateway and the others as backups. In case of a failure, the backup gateways seamlessly take over the traffic, guaranteeing uninterrupted, secure communication.

2. Load Balancing: Load balancing mechanisms distribute traffic across multiple VPN gateways, ensuring optimal resource utilization and preventing overloading of any single gateway. This improves performance and provides an additional layer of fault tolerance.

3. Automatic Failover: Implementing automatic failover mechanisms ensures that any failure or disruption in the primary VPN gateway triggers a swift and seamless switch to the backup gateway. This eliminates manual intervention, minimizing downtime and maintaining continuous network security.

4. Redundant Internet Connections: Organizations can establish redundant Internet connections to enhance fault tolerance further. This ensures that even if one connection fails, the IPsec infrastructure can continue operating using an alternate connection, guaranteeing uninterrupted, secure communication.

IPsec fault tolerance is a crucial aspect of maintaining uninterrupted network security. Organizations can ensure that their IPsec infrastructure remains operational despite failures or disruptions by implementing redundancy, failover, and load-balancing mechanisms. Such measures enhance reliability and enable seamless scalability as the organization’s network grows. With IPsec fault tolerance, organizations can rest assured that their sensitive information is protected and secure, irrespective of unforeseen circumstances.

 

Summary: IPSec Fault Tolerance

Maintaining secure connections is of utmost importance in the ever-evolving landscape of networking and data transmission. IPsec, or Internet Protocol Security, provides a reliable framework for securing data over IP networks. However, ensuring fault tolerance in IPsec is crucial to mitigate potential disruptions and guarantee uninterrupted communication. In this blog post, we explored the concept of IPsec fault tolerance and discuss strategies to enhance the resilience of IPsec connections.

Understanding IPsec Fault Tolerance

IPsec, at its core, is designed to provide confidentiality, integrity, and authenticity of network traffic. However, unforeseen circumstances such as hardware failures, network outages, or even cyber attacks can impact the availability of IPsec connections. To address these challenges, implementing fault tolerance mechanisms becomes essential.

Redundancy in IPsec Configuration

One key strategy to achieve fault tolerance in IPsec is through redundancy. By configuring redundant IPsec tunnels, network administrators can ensure that if one tunnel fails, traffic can seamlessly failover to an alternate tunnel. This redundancy can be implemented using various techniques, including dynamic routing protocols such as OSPF or BGP, or by utilizing VPN failover mechanisms provided by network devices.

Load Balancing for IPsec Connections

Load balancing plays a crucial role in distributing traffic across multiple IPsec tunnels. By evenly distributing the load, network resources can be effectively utilized, and the risk of congestion or overload on a single tunnel is mitigated. Load balancing algorithms such as round-robin, weighted round-robin, or even intelligent traffic analysis can be employed to achieve optimal utilization of IPsec connections.

Monitoring and Proactive Maintenance

Proactive monitoring and maintenance practices are paramount to ensure fault tolerance in IPsec. Network administrators should regularly monitor the health and performance of IPsec tunnels, including metrics such as latency, bandwidth utilization, and packet loss. By promptly identifying potential issues, proactive maintenance tasks such as firmware updates, patch installations, or hardware replacements can be scheduled to minimize downtime.

Conclusion:

In today’s interconnected world, where secure communication is vital, IPsec fault tolerance emerges as a critical aspect of network infrastructure. By implementing redundancy, load balancing, and proactive monitoring, organizations can enhance the resilience of their IPsec connections. Embracing fault tolerance measures safeguards against potential disruptions and ensures uninterrupted and secure data transmission over IP networks.