From a37a70b1338463b650f19646d1fad8402217eb4a Mon Sep 17 00:00:00 2001 From: Annie Hedgpeth Date: Thu, 15 Jun 2017 11:26:59 -0500 Subject: [PATCH] provider/azurerm: Example of Openshift origin (#15294) * initial commit - 101-vm-from-user-image * changed branch name * not deploying - storage problems * provisions vm but image not properly prepared * storage not correct * provisions properly * changed main.tf to azuredeploy.tf * added tfvars and info for README * tfvars ignored and corrected file ext * added CI config; added sane defaults for variables; updated deployment script, added mac specific deployment for local testing * deploy.sh to be executable * executable deploy files * added CI files; changed vars * prep for PR * removal of old folder * prep for PR * wrong args for travis * more PR prep * updated README * commented out variables in terraform.tfvars * Topic 101 vm from user image (#2) * initial commit - 101-vm-from-user-image * added tfvars and info for README * added CI config; added sane defaults for variables; updated deployment script, added mac specific deployment for local testing * prep for PR * added new template * oops, left off master * prep for PR * correct repository for destination * renamed scripts to be more intuitive; added check for docker * merge vm simple; vm from image * initial commit * deploys locally * updated deploy * consolidated deploy and after_deploy into a single script; simplified ci process; added os_profile_linux_config * added terraform show * changed to allow http & https (like ARM tmplt) * changed host_name & host_name variable desc * added az cli check * on this branch, only build test_dir; master will aggregate all the examples * merge master * added new constructs/naming for deploy scripts, etc. * suppress az login output * suppress az login output * forgot about line breaks * breaking build as an example * fixing broken build example * merge of CI config * fixed grammar in readme * prep for PR * took out armviz button and minor README changes * changed host_name * fixed merge conflicts * changed host_name variable * updating Hashicorp's changes to merged simple linux branch * updating files to merge w/master and prep for Hashicorp pr * Revert "updating files to merge w/master and prep for Hashicorp pr" This reverts commit b850cd5d2a858eff073fc5a1097a6813d0f8b362. * Revert "updating Hashicorp's changes to merged simple linux branch" This reverts commit dbaf8d14a9cdfcef0281919671357f6171ebd4e6. * removing vm from user image example from this branch * removed old branch * azure-2-vms-loadbalancer-lbrules (#13) * initial commit * need to change lb_rule & nic * deploys locally * updated README * updated travis and deploy scripts for Hari's repo * renamed deploy script * clean up * prep for PR * updated readme * fixing conflict in .travis.yml * add CI build tag * initial commit; in progress * in progress; merged Hashicorp master into this branch * in progress * in progress; created nsg * added vars to deploy; added vnet * chmod on deploy * edited vars * added var in travis * added var * added var to deploy * added storage accounts * fixed storage typos * removed storage tags * added PIPs * changed dns name vars * corrected PIP naming convention * added availability sets * added master-lb & rules * added infra lb & rules * added nics * added VMs, ready for VM extensions, can modularize in the future * added vm exts.; nsg is possibly broken; can't ssh * in progress * master ext succeeds * in progress, infra and nodes exts not succeeding * infra and node extensions fail * provisions with extensions * disabled password auth; ssh config added * changed ssh key vars * adding ssh var to deploy * commenting out validation * in progress; building openshift ext * troubleshooting openshift deploy script * changed vm names; added container * increased os disk size * in progress; troubleshooting deploy opnshft script * Updated the readme * updated deployment scripts; cleaned up variables, use remote-exec * more variable cleanup * more cleanup * simplified password; got rid of a needless comment * merge conflicts resolved --- examples/azure-openshift-origin/README.md | 114 +++ examples/azure-openshift-origin/deploy.ci.sh | 46 + examples/azure-openshift-origin/deploy.mac.sh | 24 + examples/azure-openshift-origin/main.tf | 826 ++++++++++++++++++ examples/azure-openshift-origin/outputs.tf | 23 + examples/azure-openshift-origin/variables.tf | 206 +++++ 6 files changed, 1239 insertions(+) create mode 100644 examples/azure-openshift-origin/README.md create mode 100755 examples/azure-openshift-origin/deploy.ci.sh create mode 100755 examples/azure-openshift-origin/deploy.mac.sh create mode 100644 examples/azure-openshift-origin/main.tf create mode 100644 examples/azure-openshift-origin/outputs.tf create mode 100644 examples/azure-openshift-origin/variables.tf diff --git a/examples/azure-openshift-origin/README.md b/examples/azure-openshift-origin/README.md new file mode 100644 index 000000000..0a5755512 --- /dev/null +++ b/examples/azure-openshift-origin/README.md @@ -0,0 +1,114 @@ +# OpenShift Origin Deployment Template + +This Terraform template was based on [this](https://github.com/Microsoft/openshift-origin) Azure Quickstart Template. Changes to the ARM template that may have occurred since the creation of this example may not be reflected here. + +## OpenShift Origin with Username / Password + +Current template deploys OpenShift Origin 1.5 RC0. + +This template deploys OpenShift Origin with basic username / password for authentication to OpenShift. You can select to use either CentOS or RHEL for the OS. It includes the following resources: + +|Resource |Properties | +|-------------------|------------------------------------------------------------------------------------------------------------------------------------| +|Virtual Network |**Address prefix:** 10.0.0.0/16
**Master subnet:** 10.0.0.0/24
**Node subnet:** 10.0.1.0/24 | +|Load Balancer |2 probes and two rules for TCP 80 and TCP 443 | +|Public IP Addresses|OpenShift Master public IP
OpenShift Router public IP attached to Load Balancer | +|Storage Accounts |2 Storage Accounts | +|Virtual Machines |Single master
User-defined number of nodes
All VMs include a single attached data disk for Docker thin pool logical volume| + +If you have a Red Hat subscription and would like to deploy an OpenShift Container Platform (formerly OpenShift Enterprise) cluster, please visit: https://github.com/Microsoft/openshift-container-platform + +### Generate SSH Keys + +You'll need to generate an SSH key pair in order to provision this template. Ensure that you do not include a passcode with the private key.
+If you are using a Windows computer, you can download `puttygen.exe`. You will need to export to OpenSSH (from Conversions menu) to get a valid Private Key for use in the Template.
+From a Linux or Mac, you can just use the `ssh-keygen` command. Once you are finished deploying the cluster, you can always generate a new key pair that uses a passphrase and replaces the original one used during initial deployment. + +### Create Key Vault to store SSH Private Key + +You will need to create a Key Vault to store your SSH Private Key that will then be used as part of the deployment. + +1. **Create Key Vault using Powershell**
+ a. Create new resource group: New-AzureRMResourceGroup -Name 'ResourceGroupName' -Location 'West US'
+ b. Create key vault: New-AzureRmKeyVault -VaultName 'KeyVaultName' -ResourceGroup 'ResourceGroupName' -Location 'West US'
+ c. Create variable with sshPrivateKey: $securesecret = ConvertTo-SecureString -String '[copy ssh Private Key here - including line feeds]' -AsPlainText -Force
+ d. Create Secret: Set-AzureKeyVaultSecret -Name 'SecretName' -SecretValue $securesecret -VaultName 'KeyVaultName'
+ e. Enable the Key Vault for Template Deployments: Set-AzureRmKeyVaultAccessPolicy -VaultName 'KeyVaultName' -ResourceGroupName 'ResourceGroupName' -EnabledForTemplateDeployment + +2. **Create Key Vault using Azure CLI 1.0**
+ a. Create new Resource Group: azure group create \ \
+ Ex: `azure group create ResourceGroupName 'East US'`
+ b. Create Key Vault: azure keyvault create -u \ -g \ -l \
+ Ex: `azure keyvault create -u KeyVaultName -g ResourceGroupName -l 'East US'`
+ c. Create Secret: azure keyvault secret set -u \ -s \ --file \
+ Ex: `azure keyvault secret set -u KeyVaultName -s SecretName --file ~/.ssh/id_rsa`
+ d. Enable the Keyvvault for Template Deployment: azure keyvault set-policy -u \ --enabled-for-template-deployment true
+ Ex: `azure keyvault set-policy -u KeyVaultName --enabled-for-template-deployment true`
+ +3. **Create Key Vault using Azure CLI 2.0**
+ a. Create new Resource Group: az group create -n \ -l \
+ Ex: `az group create -n ResourceGroupName -l 'East US'`
+ b. Create Key Vault: az keyvault create -n \ -g \ -l \ --enabled-for-template-deployment true
+ Ex: `az keyvault create -n KeyVaultName -g ResourceGroupName -l 'East US' --enabled-for-template-deployment true`
+ c. Create Secret: az keyvault secret set --vault-name \ -n \ --file \
+ Ex: `az keyvault secret set --vault-name KeyVaultName -n SecretName --file ~/.ssh/id_rsa`
+3. **Clone the Openshift repository [here](https://github.com/Microsoft/openshift-origin)**
+ a. Note the local script path, this will be needed for remote-execs on the remote machines.
+ +## Deploy Template + +Once you have collected all of the prerequisites for the template, you can deploy the template via terraform. + +Monitor deployment via Terraform and get the console URL from outputs of successful deployment which will look something like (if using sample parameters file and "West US 2" location): + +`https://me-master1.westus2.cloudapp.azure.com:8443/console` + +The cluster will use self-signed certificates. Accept the warning and proceed to the login page. + +### NOTE + +Ensure combination of openshiftMasterPublicIpDnsLabelPrefix, and nodeLbPublicIpDnsLabelPrefix parameters, combined with the deployment location give you globally unique URL for the cluster or deployment will fail at the step of allocating public IPs with fully-qualified-domain-names as above. + +### NOTE + +This template deploys a bastion host, merely for the connection provisioner and allowing remote-exec to run commands on machines without public IPs; notice the specific dependencies on the order in which VMs are created for this to work properly. + +### NOTE + +The OpenShift Ansible playbook does take a while to run when using VMs backed by Standard Storage. VMs backed by Premium Storage are faster. If you want Premimum Storage, select a DS or GS series VM. +
+Be sure to follow the OpenShift instructions to create the ncessary DNS entry for the OpenShift Router for access to applications. + +## Post-Deployment Operations + +This template creates an OpenShift user but does not make it a full OpenShift user. To do that, please perform the following. + +1. SSH in to master node +2. Execute the following command: + + ```sh + sudo oadm policy add-cluster-role-to-user cluster-admin + ``` +### Additional OpenShift Configuration Options + +You can configure additional settings per the official [OpenShift Origin Documentation](https://docs.openshift.org/latest/welcome/index.html). + +Few options you have + +1. Deployment Output + + a. openshiftConsoleUrl the openshift console url
+ b. openshiftMasterSsh ssh command for master node
+ c. openshiftNodeLoadBalancerFQDN node load balancer
+ + get the deployment output data + + a. portal.azure.com -> choose 'Resource groups' select your group select 'Deployments' and there the deployment 'Microsoft.Template'. As output from the deployment it contains information about the openshift console url, ssh command and load balancer url.
+ b. With the Azure CLI : azure group deployment list <resource group name> + +2. add additional users. you can find much detail about this in the openshift.org documentation under 'Cluster Administration' and 'Managing Users'. This installation uses htpasswd as the identity provider. To add more user ssh in to master node and execute following command: + + ```sh + sudo htpasswd /etc/origin/master/htpasswd user1 + ``` + Now this user can login with the 'oc' CLI tool or the openshift console url. diff --git a/examples/azure-openshift-origin/deploy.ci.sh b/examples/azure-openshift-origin/deploy.ci.sh new file mode 100755 index 000000000..33e9cbd70 --- /dev/null +++ b/examples/azure-openshift-origin/deploy.ci.sh @@ -0,0 +1,46 @@ +#!/bin/bash + +set -o errexit -o nounset + +docker run --rm -it \ + -e ARM_CLIENT_ID \ + -e ARM_CLIENT_SECRET \ + -e ARM_SUBSCRIPTION_ID \ + -e ARM_TENANT_ID \ + -e AAD_CLIENT_ID \ + -e AAD_CLIENT_SECRET \ + -e KEY_ENCRYPTION_KEY_URL \ + -e SSH_PUBLIC_KEY \ + -v /:/data \ + --workdir=/data/$(pwd) \ + --entrypoint "/bin/sh" \ + hashicorp/terraform:light \ + -c "/bin/terraform get; \ + /bin/terraform validate; \ + /bin/terraform plan -out=out.tfplan \ + -var subscription_id=$ARM_SUBSCRIPTION_ID \ + -var tenant_id=$ARM_TENANT_ID \ + -var aad_client_id=$ARM_CLIENT_ID \ + -var aad_client_secret=$ARM_CLIENT_SECRET \ + -var resource_group_name=$KEY \ + -var key_vault_name=$KEY_VAULT_NAME \ + -var key_vault_resource_group=$KEY_VAULT_RESOURCE_GROUP \ + -var key_vault_secret=$KEY_VAULT_SECRET \ + -var openshift_cluster_prefix=$KEY \ + -var openshift_password=$PASSWORD \ + -var openshift_script_path=$LOCAL_SCRIPT_PATH \ + -var ssh_public_key=\"$OS_PUBLIC_KEY\" \ + -var connection_private_ssh_key_path=$CONTAINER_PRIVATE_KEY_PATH \ + -var master_instance_count=$MASTER_COUNT \ + -var infra_instance_count=$INFRA_COUNT \ + -var node_instance_count=$NODE_COUNT; \ + /bin/terraform apply out.tfplan;" + +# cleanup deployed azure resources via azure-cli +# docker run --rm -it \ +# azuresdk/azure-cli-python \ +# sh -c "az login --service-principal -u $ARM_CLIENT_ID -p $ARM_CLIENT_SECRET --tenant $ARM_TENANT_ID > /dev/null; \ +# az vm show -g $KEY -n $KEY; \ +# az vm encryption show -g $KEY -n $KEY" + +# cleanup deployed azure resources via terraform \ No newline at end of file diff --git a/examples/azure-openshift-origin/deploy.mac.sh b/examples/azure-openshift-origin/deploy.mac.sh new file mode 100755 index 000000000..790993812 --- /dev/null +++ b/examples/azure-openshift-origin/deploy.mac.sh @@ -0,0 +1,24 @@ +#!/bin/bash + +set -o errexit -o nounset + +if docker -v; then + + # generate a unique string for CI deployment + export KEY=$(cat /dev/urandom | env LC_CTYPE=C tr -cd 'a-z' | head -c 12) + export PASSWORD="P4ssw0rd1" + export KEY_VAULT_RESOURCE_GROUP=permanent + export KEY_VAULT_NAME=TerraformVault + export KEY_VAULT_SECRET=OpenShiftSSH + export OS_PUBLIC_KEY='ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAACAQCvdOGspeeBhsOZw6OK2WmP3bBUOeZj1yaz6Dw+lzsRmjwNSmJIoGZPzlbdy1lzlkXIm2JaT4h/cUi39w+Q2RZRjxmr7TbLyuidJfFLvRJ35RDullUYLWEPx3csBroPkCv+0qgmTW/MqqjqS4yhlJ01uc9RNx9Jt3XZN7LNr8SUoBzdLCWJa1rpCTtUckO1Jyzi4VwZ2ek+nYPJuJ8hG0KeHnyXDXV4hQZTFtGvtbmgoyoybppFQMbM3a31KZeaWXUeZkZczBsdNRkX8XCDjb6zUmUMQUzZpalFlL1O+rZD0kaXKr0uZWiYOKu2LjnWeDW9x4tig1mf+L84vniP+lLKFW8na3Lzx11ysEpuhIJGPMMI8sjTCnu51PmiwHW2U9OR06skPUO7ZGD0QHg7jKXdz5bHT+1OqXeAStULDiPVRIPrxxpurPXiJRm7JPbPvPqrMqZJ3K7J9W6OGHG3CoDR5RfYlPWURTaVH10stb4hKevasCd+YoLStB1XgMaL/cG9bM0TIWmODV/+pfn800PgxeBn1vABpL0NF8K2POLs37vGJoh/RyGCDVd0HEKArpZj0/g+fv7tr3tFFOCY5bHSuDTZcY8sWPhxKXSismoApM3a+USF5HkDkWSTEiETs2wgUdTSt4MuN2maRXOK2JboQth1Qw+vCOvqcls0dMa0NQ== you@example.com' + export CONTAINER_PRIVATE_KEY_PATH="/data/Users/$USER/.ssh/id_rsa" + export LOCAL_SCRIPT_PATH="/data/Users/$USER/Code/10thmagnitude/openshift-origin/scripts" + export MASTER_COUNT=1 + export INFRA_COUNT=1 + export NODE_COUNT=1 + +/bin/sh ./deploy.ci.sh + +else + echo "Docker is used to run terraform commands, please install before run: https://docs.docker.com/docker-for-mac/install/" +fi \ No newline at end of file diff --git a/examples/azure-openshift-origin/main.tf b/examples/azure-openshift-origin/main.tf new file mode 100644 index 000000000..7237b8464 --- /dev/null +++ b/examples/azure-openshift-origin/main.tf @@ -0,0 +1,826 @@ +provider "azurerm" { + subscription_id = "${var.subscription_id}" + client_id = "${var.aad_client_id}" + client_secret = "${var.aad_client_secret}" + tenant_id = "${var.tenant_id}" +} + +resource "azurerm_resource_group" "rg" { + name = "${var.resource_group_name}" + location = "${var.resource_group_location}" +} + +# ******* NETWORK SECURITY GROUPS *********** + +resource "azurerm_network_security_group" "master_nsg" { + name = "${var.openshift_cluster_prefix}-master-nsg" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + + security_rule { + name = "allow_SSH_in_all" + description = "Allow SSH in from all locations" + priority = 100 + direction = "Inbound" + access = "Allow" + protocol = "Tcp" + source_port_range = "*" + destination_port_range = "22" + source_address_prefix = "*" + destination_address_prefix = "*" + } + + security_rule { + name = "allow_HTTPS_all" + description = "Allow HTTPS connections from all locations" + priority = 200 + direction = "Inbound" + access = "Allow" + protocol = "Tcp" + source_port_range = "*" + destination_port_range = "443" + source_address_prefix = "*" + destination_address_prefix = "*" + } + + security_rule { + name = "allow_OpenShift_console_in_all" + description = "Allow OpenShift Console connections from all locations" + priority = 300 + direction = "Inbound" + access = "Allow" + protocol = "Tcp" + source_port_range = "*" + destination_port_range = "8443" + source_address_prefix = "*" + destination_address_prefix = "*" + } +} + +resource "azurerm_network_security_group" "infra_nsg" { + name = "${var.openshift_cluster_prefix}-infra-nsg" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + + security_rule { + name = "allow_SSH_in_all" + description = "Allow SSH in from all locations" + priority = 100 + direction = "Inbound" + access = "Allow" + protocol = "Tcp" + source_port_range = "*" + destination_port_range = "22" + source_address_prefix = "*" + destination_address_prefix = "*" + } + + security_rule { + name = "allow_HTTPS_all" + description = "Allow HTTPS connections from all locations" + priority = 200 + direction = "Inbound" + access = "Allow" + protocol = "Tcp" + source_port_range = "*" + destination_port_range = "443" + source_address_prefix = "*" + destination_address_prefix = "*" + } + + security_rule { + name = "allow_HTTP_in_all" + description = "Allow HTTP connections from all locations" + priority = 300 + direction = "Inbound" + access = "Allow" + protocol = "Tcp" + source_port_range = "*" + destination_port_range = "80" + source_address_prefix = "*" + destination_address_prefix = "*" + } +} + +resource "azurerm_network_security_group" "node_nsg" { + name = "${var.openshift_cluster_prefix}-node-nsg" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + + security_rule { + name = "allow_SSH_in_all" + description = "Allow SSH in from all locations" + priority = 100 + direction = "Inbound" + access = "Allow" + protocol = "Tcp" + source_port_range = "*" + destination_port_range = "22" + source_address_prefix = "*" + destination_address_prefix = "*" + } + + security_rule { + name = "allow_HTTPS_all" + description = "Allow HTTPS connections from all locations" + priority = 200 + direction = "Inbound" + access = "Allow" + protocol = "Tcp" + source_port_range = "*" + destination_port_range = "443" + source_address_prefix = "*" + destination_address_prefix = "*" + } + + security_rule { + name = "allow_HTTP_in_all" + description = "Allow HTTP connections from all locations" + priority = 300 + direction = "Inbound" + access = "Allow" + protocol = "Tcp" + source_port_range = "*" + destination_port_range = "80" + source_address_prefix = "*" + destination_address_prefix = "*" + } +} + +# ******* STORAGE ACCOUNTS *********** + +resource "azurerm_storage_account" "bastion_storage_account" { + name = "${var.openshift_cluster_prefix}bsa" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + account_type = "${var.storage_account_type_map["${var.bastion_vm_size}"]}" +} + +resource "azurerm_storage_account" "master_storage_account" { + name = "${var.openshift_cluster_prefix}msa" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + account_type = "${var.storage_account_type_map["${var.master_vm_size}"]}" +} + +resource "azurerm_storage_account" "infra_storage_account" { + name = "${var.openshift_cluster_prefix}infrasa" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + account_type = "${var.storage_account_type_map["${var.infra_vm_size}"]}" +} + +resource "azurerm_storage_account" "nodeos_storage_account" { + name = "${var.openshift_cluster_prefix}nodeossa" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + account_type = "${var.storage_account_type_map["${var.node_vm_size}"]}" +} + +resource "azurerm_storage_account" "nodedata_storage_account" { + name = "${var.openshift_cluster_prefix}nodedatasa" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + account_type = "${var.storage_account_type_map["${var.node_vm_size}"]}" +} + +resource "azurerm_storage_account" "registry_storage_account" { + name = "${var.openshift_cluster_prefix}regsa" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + account_type = "Standard_LRS" +} + +resource "azurerm_storage_account" "persistent_volume_storage_account" { + name = "${var.openshift_cluster_prefix}pvsa" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + account_type = "Standard_LRS" +} + +# ******* AVAILABILITY SETS *********** + +resource "azurerm_availability_set" "master" { + name = "masteravailabilityset" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" +} + +resource "azurerm_availability_set" "infra" { + name = "infraavailabilityset" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" +} + +resource "azurerm_availability_set" "node" { + name = "nodeavailabilityset" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" +} + +# ******* IP ADDRESSES *********** + +resource "azurerm_public_ip" "bastion_pip" { + name = "bastionpip" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + public_ip_address_allocation = "Static" + domain_name_label = "${var.openshift_cluster_prefix}-bastion" +} + +resource "azurerm_public_ip" "openshift_master_pip" { + name = "masterpip" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + public_ip_address_allocation = "Static" + domain_name_label = "${var.openshift_cluster_prefix}" +} + +resource "azurerm_public_ip" "infra_lb_pip" { + name = "infraip" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + public_ip_address_allocation = "Static" + domain_name_label = "${var.openshift_cluster_prefix}infrapip" +} + +# ******* VNETS / SUBNETS *********** + +resource "azurerm_virtual_network" "vnet" { + name = "openshiftvnet" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + address_space = ["10.0.0.0/8"] + depends_on = ["azurerm_virtual_network.vnet"] +} + +resource "azurerm_subnet" "master_subnet" { + name = "mastersubnet" + virtual_network_name = "${azurerm_virtual_network.vnet.name}" + resource_group_name = "${azurerm_resource_group.rg.name}" + address_prefix = "10.1.0.0/16" + depends_on = ["azurerm_virtual_network.vnet"] +} + +resource "azurerm_subnet" "node_subnet" { + name = "nodesubnet" + virtual_network_name = "${azurerm_virtual_network.vnet.name}" + resource_group_name = "${azurerm_resource_group.rg.name}" + address_prefix = "10.2.0.0/16" +} + +# ******* MASTER LOAD BALANCER *********** + +resource "azurerm_lb" "master_lb" { + name = "masterloadbalancer" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + depends_on = ["azurerm_public_ip.openshift_master_pip"] + + frontend_ip_configuration { + name = "LoadBalancerFrontEnd" + public_ip_address_id = "${azurerm_public_ip.openshift_master_pip.id}" + } +} + +resource "azurerm_lb_backend_address_pool" "master_lb" { + resource_group_name = "${azurerm_resource_group.rg.name}" + name = "loadBalancerBackEnd" + loadbalancer_id = "${azurerm_lb.master_lb.id}" + depends_on = ["azurerm_lb.master_lb"] +} + +resource "azurerm_lb_probe" "master_lb" { + resource_group_name = "${azurerm_resource_group.rg.name}" + loadbalancer_id = "${azurerm_lb.master_lb.id}" + name = "8443Probe" + port = 8443 + interval_in_seconds = 5 + number_of_probes = 2 + protocol = "Tcp" + depends_on = ["azurerm_lb.master_lb"] +} + +resource "azurerm_lb_rule" "master_lb" { + resource_group_name = "${azurerm_resource_group.rg.name}" + loadbalancer_id = "${azurerm_lb.master_lb.id}" + name = "OpenShiftAdminConsole" + protocol = "Tcp" + frontend_port = 8443 + backend_port = 8443 + frontend_ip_configuration_name = "LoadBalancerFrontEnd" + backend_address_pool_id = "${azurerm_lb_backend_address_pool.master_lb.id}" + load_distribution = "SourceIP" + idle_timeout_in_minutes = 30 + probe_id = "${azurerm_lb_probe.master_lb.id}" + enable_floating_ip = false + depends_on = ["azurerm_lb_probe.master_lb", "azurerm_lb.master_lb", "azurerm_lb_backend_address_pool.master_lb"] +} + +resource "azurerm_lb_nat_rule" "master_lb" { + resource_group_name = "${azurerm_resource_group.rg.name}" + loadbalancer_id = "${azurerm_lb.master_lb.id}" + name = "${azurerm_lb.master_lb.name}-SSH-${count.index}" + protocol = "Tcp" + frontend_port = "${count.index + 2200}" + backend_port = 22 + frontend_ip_configuration_name = "LoadBalancerFrontEnd" + count = "${var.master_instance_count}" + depends_on = ["azurerm_lb.master_lb"] +} + +# ******* INFRA LOAD BALANCER *********** + +resource "azurerm_lb" "infra_lb" { + name = "infraloadbalancer" + resource_group_name = "${azurerm_resource_group.rg.name}" + location = "${azurerm_resource_group.rg.location}" + depends_on = ["azurerm_public_ip.infra_lb_pip"] + + frontend_ip_configuration { + name = "LoadBalancerFrontEnd" + public_ip_address_id = "${azurerm_public_ip.infra_lb_pip.id}" + } +} + +resource "azurerm_lb_backend_address_pool" "infra_lb" { + resource_group_name = "${azurerm_resource_group.rg.name}" + name = "loadBalancerBackEnd" + loadbalancer_id = "${azurerm_lb.infra_lb.id}" + depends_on = ["azurerm_lb.infra_lb"] +} + +resource "azurerm_lb_probe" "infra_lb_http_probe" { + resource_group_name = "${azurerm_resource_group.rg.name}" + loadbalancer_id = "${azurerm_lb.infra_lb.id}" + name = "httpProbe" + port = 80 + interval_in_seconds = 5 + number_of_probes = 2 + protocol = "Tcp" + depends_on = ["azurerm_lb.infra_lb"] +} + +resource "azurerm_lb_probe" "infra_lb_https_probe" { + resource_group_name = "${azurerm_resource_group.rg.name}" + loadbalancer_id = "${azurerm_lb.infra_lb.id}" + name = "httpsProbe" + port = 443 + interval_in_seconds = 5 + number_of_probes = 2 + protocol = "Tcp" +} + +resource "azurerm_lb_rule" "infra_lb_http" { + resource_group_name = "${azurerm_resource_group.rg.name}" + loadbalancer_id = "${azurerm_lb.infra_lb.id}" + name = "OpenShiftRouterHTTP" + protocol = "Tcp" + frontend_port = 80 + backend_port = 80 + frontend_ip_configuration_name = "LoadBalancerFrontEnd" + backend_address_pool_id = "${azurerm_lb_backend_address_pool.infra_lb.id}" + probe_id = "${azurerm_lb_probe.infra_lb_http_probe.id}" + depends_on = ["azurerm_lb_probe.infra_lb_http_probe", "azurerm_lb.infra_lb", "azurerm_lb_backend_address_pool.infra_lb"] +} + +resource "azurerm_lb_rule" "infra_lb_https" { + resource_group_name = "${azurerm_resource_group.rg.name}" + loadbalancer_id = "${azurerm_lb.infra_lb.id}" + name = "OpenShiftRouterHTTPS" + protocol = "Tcp" + frontend_port = 443 + backend_port = 443 + frontend_ip_configuration_name = "LoadBalancerFrontEnd" + backend_address_pool_id = "${azurerm_lb_backend_address_pool.infra_lb.id}" + probe_id = "${azurerm_lb_probe.infra_lb_https_probe.id}" + depends_on = ["azurerm_lb_probe.infra_lb_https_probe", "azurerm_lb_backend_address_pool.infra_lb"] +} + +# ******* NETWORK INTERFACES *********** + +resource "azurerm_network_interface" "bastion_nic" { + name = "bastionnic${count.index}" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + network_security_group_id = "${azurerm_network_security_group.master_nsg.id}" + + ip_configuration { + name = "bastionip${count.index}" + subnet_id = "${azurerm_subnet.master_subnet.id}" + private_ip_address_allocation = "Dynamic" + public_ip_address_id = "${azurerm_public_ip.bastion_pip.id}" + } +} + +resource "azurerm_network_interface" "master_nic" { + name = "masternic${count.index}" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + network_security_group_id = "${azurerm_network_security_group.master_nsg.id}" + count = "${var.master_instance_count}" + + ip_configuration { + name = "masterip${count.index}" + subnet_id = "${azurerm_subnet.master_subnet.id}" + private_ip_address_allocation = "Dynamic" + load_balancer_backend_address_pools_ids = ["${azurerm_lb_backend_address_pool.master_lb.id}"] + load_balancer_inbound_nat_rules_ids = ["${element(azurerm_lb_nat_rule.master_lb.*.id, count.index)}"] + } +} + +resource "azurerm_network_interface" "infra_nic" { + name = "infra_nic${count.index}" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + network_security_group_id = "${azurerm_network_security_group.infra_nsg.id}" + count = "${var.infra_instance_count}" + + ip_configuration { + name = "infraip${count.index}" + subnet_id = "${azurerm_subnet.master_subnet.id}" + private_ip_address_allocation = "Dynamic" + load_balancer_backend_address_pools_ids = ["${azurerm_lb_backend_address_pool.infra_lb.id}"] + } +} + +resource "azurerm_network_interface" "node_nic" { + name = "node_nic${count.index}" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + network_security_group_id = "${azurerm_network_security_group.node_nsg.id}" + count = "${var.node_instance_count}" + + ip_configuration { + name = "nodeip${count.index}" + subnet_id = "${azurerm_subnet.node_subnet.id}" + private_ip_address_allocation = "Dynamic" + } +} + +# ******* Bastion Host ******* + +resource "azurerm_virtual_machine" "bastion" { + name = "${var.openshift_cluster_prefix}-bastion-1" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + network_interface_ids = ["${azurerm_network_interface.bastion_nic.id}"] + vm_size = "${var.bastion_vm_size}" + delete_os_disk_on_termination = true + delete_data_disks_on_termination = true + + tags { + displayName = "${var.openshift_cluster_prefix}-bastion VM Creation" + } + + os_profile { + computer_name = "${var.openshift_cluster_prefix}-bastion-${count.index}" + admin_username = "${var.admin_username}" + admin_password = "${var.openshift_password}" + } + + os_profile_linux_config { + disable_password_authentication = true + + ssh_keys { + path = "/home/${var.admin_username}/.ssh/authorized_keys" + key_data = "${var.ssh_public_key}" + } + } + + storage_image_reference { + publisher = "${lookup(var.os_image_map, join("_publisher", list(var.os_image, "")))}" + offer = "${lookup(var.os_image_map, join("_offer", list(var.os_image, "")))}" + sku = "${lookup(var.os_image_map, join("_sku", list(var.os_image, "")))}" + version = "${lookup(var.os_image_map, join("_version", list(var.os_image, "")))}" + } + + storage_os_disk { + name = "${var.openshift_cluster_prefix}-master-osdisk${count.index}" + vhd_uri = "${azurerm_storage_account.bastion_storage_account.primary_blob_endpoint}vhds/${var.openshift_cluster_prefix}-bastion-osdisk.vhd" + caching = "ReadWrite" + create_option = "FromImage" + disk_size_gb = 60 + } +} + +# ******* Master VMs ******* + +resource "azurerm_virtual_machine" "master" { + name = "${var.openshift_cluster_prefix}-master-${count.index}" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + availability_set_id = "${azurerm_availability_set.master.id}" + network_interface_ids = ["${element(azurerm_network_interface.master_nic.*.id, count.index)}"] + vm_size = "${var.master_vm_size}" + delete_os_disk_on_termination = true + delete_data_disks_on_termination = true + count = "${var.master_instance_count}" + depends_on = ["azurerm_virtual_machine.infra", "azurerm_virtual_machine.node"] + + tags { + displayName = "${var.openshift_cluster_prefix}-master VM Creation" + } + + connection { + host = "${azurerm_public_ip.openshift_master_pip.fqdn}" + user = "${var.admin_username}" + port = 2200 + private_key = "${file(var.connection_private_ssh_key_path)}" + } + + provisioner "file" { + source = "${var.openshift_script_path}/masterPrep.sh" + destination = "masterPrep.sh" + } + + provisioner "file" { + source = "${var.openshift_script_path}/deployOpenShift.sh" + destination = "deployOpenShift.sh" + } + + provisioner "remote-exec" { + inline = [ + "chmod +x masterPrep.sh", + "chmod +x deployOpenShift.sh", + "sudo bash masterPrep.sh \"${azurerm_storage_account.persistent_volume_storage_account.name}\" \"${var.admin_username}\" && sudo bash deployOpenShift.sh \"${var.admin_username}\" \"${var.openshift_password}\" \"${var.key_vault_secret}\" \"${var.openshift_cluster_prefix}-master\" \"${azurerm_public_ip.openshift_master_pip.fqdn}\" \"${azurerm_public_ip.openshift_master_pip.ip_address}\" \"${var.openshift_cluster_prefix}-infra\" \"${var.openshift_cluster_prefix}-node\" \"${var.node_instance_count}\" \"${var.infra_instance_count}\" \"${var.master_instance_count}\" \"${var.default_sub_domain_type}\" \"${azurerm_storage_account.registry_storage_account.name}\" \"${azurerm_storage_account.registry_storage_account.primary_access_key}\" \"${var.tenant_id}\" \"${var.subscription_id}\" \"${var.aad_client_id}\" \"${var.aad_client_secret}\" \"${azurerm_resource_group.rg.name}\" \"${azurerm_resource_group.rg.location}\" \"${var.key_vault_name}\"" + ] + } + + os_profile { + computer_name = "${var.openshift_cluster_prefix}-master-${count.index}" + admin_username = "${var.admin_username}" + admin_password = "${var.openshift_password}" + } + + os_profile_linux_config { + disable_password_authentication = true + + ssh_keys { + path = "/home/${var.admin_username}/.ssh/authorized_keys" + key_data = "${var.ssh_public_key}" + } + } + + storage_image_reference { + publisher = "${lookup(var.os_image_map, join("_publisher", list(var.os_image, "")))}" + offer = "${lookup(var.os_image_map, join("_offer", list(var.os_image, "")))}" + sku = "${lookup(var.os_image_map, join("_sku", list(var.os_image, "")))}" + version = "${lookup(var.os_image_map, join("_version", list(var.os_image, "")))}" + } + + storage_os_disk { + name = "${var.openshift_cluster_prefix}-master-osdisk${count.index}" + vhd_uri = "${azurerm_storage_account.master_storage_account.primary_blob_endpoint}vhds/${var.openshift_cluster_prefix}-master-osdisk${count.index}.vhd" + caching = "ReadWrite" + create_option = "FromImage" + disk_size_gb = 60 + } + + storage_data_disk { + name = "${var.openshift_cluster_prefix}-master-docker-pool${count.index}" + vhd_uri = "${azurerm_storage_account.master_storage_account.primary_blob_endpoint}vhds/${var.openshift_cluster_prefix}-master-docker-pool${count.index}.vhd" + disk_size_gb = "${var.data_disk_size}" + create_option = "Empty" + lun = 0 + } +} + +# ******* Infra VMs ******* + +resource "azurerm_virtual_machine" "infra" { + name = "${var.openshift_cluster_prefix}-infra-${count.index}" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + availability_set_id = "${azurerm_availability_set.infra.id}" + network_interface_ids = ["${element(azurerm_network_interface.infra_nic.*.id, count.index)}"] + vm_size = "${var.infra_vm_size}" + delete_os_disk_on_termination = true + delete_data_disks_on_termination = true + count = "${var.infra_instance_count}" + + tags { + displayName = "${var.openshift_cluster_prefix}-infra VM Creation" + } + + connection { + type = "ssh" + bastion_host = "${azurerm_public_ip.bastion_pip.fqdn}" + bastion_user = "${var.admin_username}" + bastion_private_key = "${file(var.connection_private_ssh_key_path)}" + host = "${element(azurerm_network_interface.infra_nic.*.private_ip_address, count.index)}" + user = "${var.admin_username}" + private_key = "${file(var.connection_private_ssh_key_path)}" + } + + provisioner "file" { + source = "${var.openshift_script_path}/nodePrep.sh" + destination = "nodePrep.sh" + } + + provisioner "remote-exec" { + inline = [ + "chmod +x nodePrep.sh", + "sudo bash nodePrep.sh", + ] + } + + os_profile { + computer_name = "${var.openshift_cluster_prefix}-infra-${count.index}" + admin_username = "${var.admin_username}" + admin_password = "${var.openshift_password}" + } + + os_profile_linux_config { + disable_password_authentication = true + + ssh_keys { + path = "/home/${var.admin_username}/.ssh/authorized_keys" + key_data = "${var.ssh_public_key}" + } + } + + storage_image_reference { + publisher = "${lookup(var.os_image_map, join("_publisher", list(var.os_image, "")))}" + offer = "${lookup(var.os_image_map, join("_offer", list(var.os_image, "")))}" + sku = "${lookup(var.os_image_map, join("_sku", list(var.os_image, "")))}" + version = "${lookup(var.os_image_map, join("_version", list(var.os_image, "")))}" + } + + storage_os_disk { + name = "${var.openshift_cluster_prefix}-infra-osdisk${count.index}" + vhd_uri = "${azurerm_storage_account.infra_storage_account.primary_blob_endpoint}vhds/${var.openshift_cluster_prefix}-infra-osdisk${count.index}.vhd" + caching = "ReadWrite" + create_option = "FromImage" + } + + storage_data_disk { + name = "${var.openshift_cluster_prefix}-infra-docker-pool" + vhd_uri = "${azurerm_storage_account.infra_storage_account.primary_blob_endpoint}vhds/${var.openshift_cluster_prefix}-infra-docker-pool${count.index}.vhd" + disk_size_gb = "${var.data_disk_size}" + create_option = "Empty" + lun = 0 + } +} + +# ******* Node VMs ******* + +resource "azurerm_virtual_machine" "node" { + name = "${var.openshift_cluster_prefix}-node-${count.index}" + location = "${azurerm_resource_group.rg.location}" + resource_group_name = "${azurerm_resource_group.rg.name}" + availability_set_id = "${azurerm_availability_set.node.id}" + network_interface_ids = ["${element(azurerm_network_interface.node_nic.*.id, count.index)}"] + vm_size = "${var.node_vm_size}" + delete_os_disk_on_termination = true + delete_data_disks_on_termination = true + count = "${var.node_instance_count}" + + tags { + displayName = "${var.openshift_cluster_prefix}-node VM Creation" + } + + connection { + type = "ssh" + bastion_host = "${azurerm_public_ip.bastion_pip.fqdn}" + bastion_user = "${var.admin_username}" + bastion_private_key = "${file(var.connection_private_ssh_key_path)}" + host = "${element(azurerm_network_interface.node_nic.*.private_ip_address, count.index)}" + user = "${var.admin_username}" + private_key = "${file(var.connection_private_ssh_key_path)}" + } + + provisioner "file" { + source = "${var.openshift_script_path}/nodePrep.sh" + destination = "nodePrep.sh" + } + + provisioner "remote-exec" { + inline = [ + "chmod +x nodePrep.sh", + "sudo bash nodePrep.sh", + ] + } + + os_profile { + computer_name = "${var.openshift_cluster_prefix}-node-${count.index}" + admin_username = "${var.admin_username}" + admin_password = "${var.openshift_password}" + } + + os_profile_linux_config { + disable_password_authentication = true + + ssh_keys { + path = "/home/${var.admin_username}/.ssh/authorized_keys" + key_data = "${var.ssh_public_key}" + } + } + + storage_image_reference { + publisher = "${lookup(var.os_image_map, join("_publisher", list(var.os_image, "")))}" + offer = "${lookup(var.os_image_map, join("_offer", list(var.os_image, "")))}" + sku = "${lookup(var.os_image_map, join("_sku", list(var.os_image, "")))}" + version = "${lookup(var.os_image_map, join("_version", list(var.os_image, "")))}" + } + + storage_os_disk { + name = "${var.openshift_cluster_prefix}-node-osdisk" + vhd_uri = "${azurerm_storage_account.nodeos_storage_account.primary_blob_endpoint}vhds/${var.openshift_cluster_prefix}-node-osdisk${count.index}.vhd" + caching = "ReadWrite" + create_option = "FromImage" + } + + storage_data_disk { + name = "${var.openshift_cluster_prefix}-node-docker-pool${count.index}" + vhd_uri = "${azurerm_storage_account.nodeos_storage_account.primary_blob_endpoint}vhds/${var.openshift_cluster_prefix}-node-docker-pool${count.index}.vhd" + disk_size_gb = "${var.data_disk_size}" + create_option = "Empty" + lun = 0 + } +} + +# ******* VM EXTENSIONS ******* + + +# resource "azurerm_virtual_machine_extension" "deploy_open_shift_master" { +# name = "masterOpShExt${count.index}" +# location = "${azurerm_resource_group.rg.location}" +# resource_group_name = "${azurerm_resource_group.rg.name}" +# virtual_machine_name = "${element(azurerm_virtual_machine.master.*.name, count.index)}" +# publisher = "Microsoft.Azure.Extensions" +# type = "CustomScript" +# type_handler_version = "2.0" +# auto_upgrade_minor_version = true +# depends_on = ["azurerm_virtual_machine.master", "azurerm_virtual_machine_extension.node_prep", "azurerm_storage_container.vhds", "azurerm_virtual_machine_extension.deploy_infra"] +# +# settings = <