cloud-foundation-fabric/fast/stages/02-networking-peering/README.md

32 KiB

Networking with VPC Peering

This stage sets up the shared network infrastructure for the whole organization. It adopts the common “hub and spoke” reference design, which is well suited to multiple scenarios, and offers several advantages versus other designs:

  • the “hub” VPC centralizes external connectivity to on-prem or other cloud environments, and is ready to host cross-environment services like CI/CD, code repositories, and monitoring probes
  • the “spoke” VPCs allow partitioning workloads (e.g. by environment like in this setup), while still retaining controlled access to central connectivity and services
  • Shared VPC in both hub and spokes splits management of network resources in specific (host) projects, while still allowing them to be consumed from workload (service) projects
  • the design also lends itself to easy DNS centralization, both from on-prem to cloud and from cloud to on-prem

Connectivity between hub and spokes is established here via VPC Peering, which offers a complete isolation between environments, and no choke-points in the data plane. Different ways of implementing connectivity, and their respective pros and cons, are discussed below.

The following diagram illustrates the high-level design, and should be used as a reference for the following sections. The final number of subnets, and their IP addressing design will of course depend on customer-specific requirements, and can be easily changed via variables or external data files without having to edit the actual code.

Networking diagram

Design overview and choices

VPC design

The hub/landing VPC hosts external connectivity and shared services for spoke VPCs, which are connected to it via VPC peering. Spokes are used here to partition environments, which is a fairly common pattern:

  • one spoke VPC for the production environment
  • one spoke VPC for the development environment

Each VPC is created into its own project, and each project is configured as a Shared VPC host, so that network-related resources and access configurations via IAM are kept separate for each VPC.

The design easily lends itself to implementing additional environments, or adopting a different logical mapping for spokes (e.g. one spoke for each company entity, etc.). Adding spokes is a trivial operation, does not increase the design complexity, and is explained in operational terms in the following sections.

In multi-organization scenarios, where production and non-production resources use different Cloud Identity and GCP organizations, the hub/landing VPC is usually part of the production organization, and establishes connections with production spokes in its same organization, and non-production spokes in a different organization.

External connectivity

External connectivity to on-prem is implemented here via HA VPN (two tunnels per region), as this is the minimum common denominator often used directly, or as a stop-gap solution to validate routing and transfer data, while waiting for interconnects to be provisioned.

Connectivity to additional on-prem sites or other cloud providers should be implemented in a similar fashion, via VPN tunnels or interconnects in the landing VPC sharing the same regional router.

Internal connectivity

Each environment has full line of sight with the Landing VPC, and hence with any networks interconnected with it (e.g. your onprem environment). Environments cannot communicate with each other (prod to dev and viceversa). If this is a requirement, and according to your specific needs and constraints, solutions based on full-mesh peerings, VPNs or NVA should be added to this design.

As mentioned initially, there are of course other ways to implement internal connectivity other than VPC peering. These can be easily retrofitted with minimal code changes, but introduce additional considerations for service interoperability, quotas and management.

This is a summary of the main options:

  • HA VPN (implemented by 02-networking-vpn)
    • Pros: simple compatibility with GCP services that leverage peering internally, better control on routes, avoids peering groups shared quotas and limits
    • Cons: additional cost, marginal increase in latency, requires multiple tunnels for full bandwidth
  • VPC Peering (implemented here)
    • Pros: no additional costs, full bandwidth with no configurations, no extra latency, total environment isolation
    • Cons: no transitivity (e.g. to GKE masters, Cloud SQL, etc.), no selective exchange of routes, several quotas and limits shared between VPCs in a peering group
  • Multi-NIC appliances (implemented by 02-networking-nva)
    • Pros: additional security features (e.g. IPS), potentially better integration with on-prem systems by using the same vendor
    • Cons: complex HA/failover setup, limited by VM bandwidth and scale, additional costs for VMs and licenses, out of band management of a critical cloud component

IP ranges, subnetting, routing

Minimizing the number of routes (and subnets) in use on the cloud environment is an important consideration, as it simplifies management and avoids hitting Cloud Router and VPC quotas and limits. For this reason, we recommend careful planning of the IP space used in your cloud environment, to be able to use large IP CIDR blocks in routes whenever possible.

This stage uses a dedicated /16 block (which should of course be sized to your needs) for each region in each VPC, and subnets created in each VPC derive their ranges from the relevant block.

Spoke VPCs also define and reserve two "special" CIDR ranges dedicated to PSA (Private Service Access) and Internal HTTPs Load Balancers (L7ILB).

Routes in GCP are either automatically created for VPC subnets, manually created via static routes, or dynamically programmed by Cloud Routers via BGP sessions, which can be configured to advertise VPC ranges, and/or custom ranges via custom advertisements.

In this setup:

  • routes between multiple subnets within the same VPC are automatically programmed by GCP
  • each spoke exchanges routes with the hub/landing through VPC peering
  • spokes don't exchange routes, directly or indirectly
  • on-premises is connected to the landing VPC and dynamically exchanges BGP routes with GCP using HA VPN

Internet egress

The path of least resistance for Internet egress is using Cloud NAT, and that is what's implemented in this setup, with a NAT gateway configured for each VPC.

Several other scenarios are possible of course, with varying degrees of complexity:

  • a forward proxy, with optional URL filters
  • a default route to on-prem to leverage existing egress infrastructure
  • a full-fledged perimeter firewall to control egress and implement additional security features like IPS

Future pluggable modules will allow to easily experiment, or deploy the above scenarios.

VPC and Hierarchical Firewall

The GCP Firewall is a stateful, distributed feature that allows the creation of L4 policies, either via VPC-level rules or more recently via hierarchical policies applied on the resource hierarchy (organization, folders).

The current setup adopts both firewall types, and uses hierarchical rules on the Networking folder for common ingress rules (egress is open by default), e.g. from health check or IAP forwarders ranges, and VPC rules for the environment or workload-level ingress.

Rules and policies are defined in simple YAML files, described below.

DNS

DNS goes hand in hand with networking, especially on GCP where Cloud DNS zones and policies are associated at the VPC level. This setup implements both DNS flows:

  • on-prem to cloud via private zones for cloud-managed domains, and an inbound policy used as forwarding target or via delegation (requires some extra configuration) from on-prem DNS resolvers
  • cloud to on-prem via forwarding zones for the on-prem managed domains

DNS configuration is further centralized by leveraging peering zones, so that

  • the hub/landing Cloud DNS hosts configurations for on-prem forwarding, Google API domains, and the top-level private zone/s (e.g. gcp.example.com)
  • the spokes Cloud DNS host configurations for the environment-specific domains (e.g. prod.gcp.example.com), which are bound to the hub/landing leveraging cross-project binding; a peering zone for the . (root) zone is then created on each spoke, delegating all DNS resolution to hub/landing.
  • Private Google Access is enabled for a selection of the supported domains, namely
    • private.googleapis.com
    • restricted.googleapis.com
    • gcr.io
    • packages.cloud.google.com
    • pkg.dev
    • pki.goog

To complete the configuration, the 35.199.192.0/19 range should be routed on the VPN tunnels from on-prem, and the following names configured for DNS forwarding to cloud:

  • private.googleapis.com
  • restricted.googleapis.com
  • gcp.example.com (used as a placeholder)

From cloud, the example.com domain (used as a placeholder) is forwarded to on-prem.

This configuration is battle-tested, and flexible enough to lend itself to simple modifications without subverting its design, for example by forwarding and peering root zones to bypass Cloud DNS external resolution.

How to run this stage

This stage is meant to be executed after the resman stage has run, as it leverages the automation service account and bucket created there, and additional resources configured in the bootstrap stage.

It's of course possible to run this stage in isolation, but that's outside the scope of this document, and you would need to refer to the code for the previous stages for the environmental requirements.

Before running this stage, you need to make sure you have the correct credentials and permissions, and localize variables by assigning values that match your configuration.

Providers configuration

The default way of making sure you have the right permissions, is to use the identity of the service account pre-created for this stage during the resource management stage, and that you are a member of the group that can impersonate it via provider-level configuration (gcp-devops or organization-admins).

To simplify setup, the previous stage pre-configures a valid providers file in its output, and optionally writes it to a local file if the outputs_location variable is set to a valid path.

If you have set a valid value for outputs_location in the bootstrap stage, simply link the relevant providers.tf file from this stage's folder in the path you specified:

# `outputs_location` is set to `~/fast-config`
ln -s ~/fast-config/providers/02-networking-providers.tf .

If you have not configured outputs_location in bootstrap, you can derive the providers file from that stage's outputs:

cd ../01-resman
terraform output -json providers | jq -r '.["02-networking"]' \
  > ../02-networking/providers.tf

Variable configuration

There are two broad sets of variables you will need to fill in:

  • variables shared by other stages (org id, billing account id, etc.), or derived from a resource managed by a different stage (folder id, automation project id, etc.)
  • variables specific to resources managed by this stage

To avoid the tedious job of filling in the first group of variables with values derived from other stages' outputs, the same mechanism used above for the provider configuration can be used to leverage pre-configured .tfvars files.

If you have set a valid value for outputs_location in the bootstrap and in the resman stage, simply link the relevant *.auto.tfvars.json files from this stage's folder in the path you specified. The * above is set to the name of the stage that produced it, except for globals.auto.tfvars.json which is also generated by the bootstrap stage, containing global values compiled manually for the bootstrap stage. For this stage, link the following files:

# `outputs_location` is set to `~/fast-config`
ln -s ~/fast-config/tfvars/globals.auto.tfvars.json .
ln -s ~/fast-config/tfvars/00-bootstrap.auto.tfvars.json .
ln -s ~/fast-config/tfvars/01-resman.auto.tfvars.json .

A second set of variables is specific to this stage, they are all optional so if you need to customize them, create an extra terraform.tfvars file.

Please refer to the Variables table below for a map of the variable origins, and to the sections below on how to adapt this stage to your networking configuration.

VPCs

VPCs are defined in separate files, one for landing and one for each of prod and dev. Each file contains the same resources, described in the following paragraphs.

The project (project) contains the VPC, and enables the required APIs and sets itself as a "host project".

The VPC (net-vpc) manages the DNS inbound policy (for Landing), explicit routes for {private,restricted}.googleapis.com, and its subnets. Subnets are created leveraging a "resource factory" paradigm, where the configuration is separated from the module that implements it, and stored in a well-structured file. To add a new subnet, simply create a new file in the data_folder directory defined in the module, following the examples found in the Fabric net-vpc documentation. Sample subnets are shipped in data/subnets, and can be easily customised to fit your needs.

Subnets for L7 ILBs are handled differently, and defined in variable l7ilb_subnets, while ranges for PSA are configured by variable psa_ranges - such variables are consumed by spoke VPCs.

Cloud NAT (net-cloudnat) manages the networking infrastructure required to enable internet egress.

VPNs

Connectivity to on-prem is implemented with HA VPN (net-vpn) and defined in vpn-onprem.tf. The file provisionally implements a single logical connection between onprem and landing at europe-west1, and the relevant parameters for its configuration are found in variable vpn_onprem_configs.

Routing and BGP

Each VPC network (net-vpc) manages a separate routing table, which can define static routes (e.g. to private.googleapis.com) and receives dynamic routes from BGP sessions established with neighbor networks (i.e. landing receives routes from onprem and any other interconnected network). Spokes receive dynamic routes programmed on the Landing VPC from the VPC peering.

Static routes are defined in vpc-*.tf files, in the routes section of each net-vpc module.

Firewall

VPC firewall rules (net-vpc-firewall) are defined per-vpc on each vpc-*.tf file and leverage a resource factory to massively create rules. To add a new firewall rule, create a new file or edit an existing one in the data_folder directory defined in the module net-vpc-firewall, following the examples of the "Rules factory" section of the module documentation. Sample firewall rules are shipped in data/firewall-rules/landing and can be easily customised.

Hierarchical firewall policies (folder) are defined in main.tf, and managed through a policy factory implemented by the folder module, which applies the defined hierarchical to the Networking folder, which contains all the core networking infrastructure. Policies are defined in the rules_file file - to define a new one simply use the instructions found on "Firewall policy factory". Sample hierarchical firewall policies are shipped in data/hierarchical-policy-rules.yaml and can be easily customised.

DNS architecture

The DNS (dns) infrastructure is defined in the respective dns-xxx.tf files.

Cloud DNS manages onprem forwarding, the main GCP zone (in this example gcp.example.com) and environment-specific zones (i.e. dev.gcp.example.com and prod.gcp.example.com).

Cloud environment

Per the section above Landing acts as the source of truth for DNS within the Cloud environment. Resources defined in the spoke VPCs consume the Landing DNS infrastructure through DNS peering (e.g. prod-landing-root-dns-peering). Spokes can optionally define private zones (e.g. prod-dns-private-zone) - granting visibility to the Landing VPC ensures that the whole cloud environment can query such zones.

Cloud to on-prem

Leveraging the forwarding zones defined on Landing (e.g. onprem-example-dns-forwarding and reverse-10-dns-forwarding), the cloud environment can resolve in-addr.arpa. and onprem.example.com. using the on-premises DNS infrastructure. Onprem resolvers IPs are set in variable dns.onprem.

DNS queries sent to the on-premises infrastructure come from the 35.199.192.0/19 source range, which is only accessible from within a VPC or networks connected to one.

On-prem to cloud

The Inbound DNS Policy defined in module landing-vpc (landing.tf) automatically reserves the first available IP address on each created subnet (typically the third one in a CIDR) to expose the Cloud DNS service so that it can be consumed from outside of GCP.

Private Google Access

Private Google Access (or PGA) enables VMs and on-prem systems to consume Google APIs from within the Google network, and is already fully configured on this environment.

For PGA to work:

  • Private Google Access should be enabled on the subnet.
    Subnets created by the net-vpc module are PGA-enabled by default.

  • 199.36.153.4/30 (restricted.googleapis.com) and 199.36.153.8/30 (private.googleapis.com) should be routed from on-prem to VPC, and from there to the default-internet-gateway.
    Per variable vpn_onprem_configs such ranges are advertised to onprem - furthermore every VPC (e.g. see landing-vpc in landing.tf) has explicit routes set in case the 0.0.0.0/0 route is changed.

  • A private DNS zone for googleapis.com should be created and configured per this article, as implemented in module googleapis-private-zone in dns-landing.tf

Preliminar activities

Before running terraform apply on this stage, make sure to adapt all of variables.tf to your needs, to update all reference to regions (e.g. europe-west1 or ew1) in the whole directory to match your preferences.

If you're not using FAST, you'll also need to create a providers.tf file to configure the GCS backend and the service account to use to run the deployment.

You're now ready to run terraform init and apply.

Post-deployment activities

  • On-prem routers should be configured to advertise all relevant CIDRs to the GCP environments. To avoid hitting GCP quotas, we recomment aggregating routes as much as possible.
  • On-prem routers should accept BGP sessions from their cloud peers.
  • On-prem DNS servers should have forward zones for GCP-managed ones.

Customizations

Adding an environment

To create a new environment (e.g. staging), a few changes are required.

Create a spoke-staging.tf file by copying spoke-prod.tf file, and adapt the new file by replacing the value "prod" with the value "staging". Running diff spoke-dev.tf spoke-prod.tf can help to see how environment files differ.

The new VPC requires a set of dedicated CIDRs, one per region, added to variable custom_adv (for example as spoke_staging_ew1 and spoke_staging_ew4).

custom_adv is a map that "resolves" CIDR names to actual addresses, and will be used later to configure routing.

Variables managing L7 Interal Load Balancers (l7ilb_subnets) and Private Service Access (psa_ranges) should also be adapted, and subnets and firewall rules for the new spoke should be added as described above.

DNS configurations are centralised in the dns-*.tf files. Spokes delegate DNS resolution to Landing through DNS peering, and optionally define a private zone (e.g. dev.gcp.example.com) which the landing peers to. To configure DNS for a new environment, copy one of the other environments DNS files e.g. (dns-dev.tf) into a new dns-*.tf file suffixed with the environment name (e.g. dns-staging.tf), and update its content accordingly. Don't forget to add a peering zone from the landing to the newly created environment private zone.

Files

name description modules resources
dns-dev.tf Development spoke DNS zones and peerings setup. dns
dns-landing.tf Landing DNS zones and peerings setup. dns
dns-prod.tf Production spoke DNS zones and peerings setup. dns
landing.tf Landing VPC and related resources. net-cloudnat · net-vpc · net-vpc-firewall · project
main.tf Networking folder and hierarchical policy. folder
monitoring.tf Network monitoring dashboards. google_monitoring_dashboard
outputs.tf Module outputs. google_storage_bucket_object · local_file
peerings.tf None net-vpc-peering
spoke-dev.tf Dev spoke VPC and related resources. net-cloudnat · net-vpc · net-vpc-firewall · project google_project_iam_binding
spoke-prod.tf Production spoke VPC and related resources. net-cloudnat · net-vpc · net-vpc-firewall · project google_project_iam_binding
test-resources.tf temporary instances for testing compute-vm
variables-peerings.tf Peering related variables.
variables.tf Module variables.
vpn-onprem.tf VPN between landing and onprem. net-vpn-ha

Variables

name description type required default producer
automation Automation resources created by the bootstrap stage. object({…}) 00-bootstrap
billing_account Billing account id and organization id ('nnnnnnnn' or null). object({…}) 00-bootstrap
folder_ids Folders to be used for the networking resources in folders/nnnnnnnnnnn format. If null, folder will be created. object({…}) 01-resman
organization Organization details. object({…}) 00-bootstrap
prefix Prefix used for resources that need unique names. Use 9 characters or less. string 00-bootstrap
custom_adv Custom advertisement definitions in name => range format. map(string) {…}
custom_roles Custom roles defined at the org level, in key => id format. object({…}) null 00-bootstrap
data_dir Relative path for the folder storing configuration data for network resources. string "data"
dns Onprem DNS resolvers. map(list(string)) {…}
l7ilb_subnets Subnets used for L7 ILBs. map(list(object({…}))) {…}
outputs_location Path where providers and tfvars files for the following stages are written. Leave empty to disable. string null
peering_configs Peering configurations. map(object({…})) {…}
psa_ranges IP ranges used for Private Service Access (e.g. CloudSQL). object({…}) null
region_trigram Short names for GCP regions. map(string) {…}
router_onprem_configs Configurations for routers used for onprem connectivity. map(object({…})) {…}
service_accounts Automation service accounts in name => email format. object({…}) null 01-resman
vpn_onprem_configs VPN gateway configuration for onprem interconnection. map(object({…})) {…}

Outputs

name description sensitive consumers
cloud_dns_inbound_policy IP Addresses for Cloud DNS inbound policy.
host_project_ids Network project ids.
host_project_numbers Network project numbers.
shared_vpc_self_links Shared VPC host projects.
tfvars Terraform variables file for the following stages.
vpn_gateway_endpoints External IP Addresses for the GCP VPN gateways.