package vsphere import ( "fmt" "log" "net" "strconv" "strings" "github.com/hashicorp/terraform/helper/schema" "github.com/vmware/govmomi" "github.com/vmware/govmomi/find" "github.com/vmware/govmomi/object" "github.com/vmware/govmomi/property" "github.com/vmware/govmomi/vim25/mo" "github.com/vmware/govmomi/vim25/types" "golang.org/x/net/context" ) var DefaultDNSSuffixes = []string{ "vsphere.local", } var DefaultDNSServers = []string{ "8.8.8.8", "8.8.4.4", } var DiskControllerTypes = []string{ "scsi", "scsi-lsi-parallel", "scsi-buslogic", "scsi-paravirtual", "scsi-lsi-sas", "ide", } type networkInterface struct { deviceName string label string ipv4Address string ipv4PrefixLength int ipv4Gateway string ipv6Address string ipv6PrefixLength int ipv6Gateway string adapterType string // TODO: Make "adapter_type" argument macAddress string } type hardDisk struct { name string size int64 iops int64 initType string vmdkPath string controller string bootable bool } //Additional options Vsphere can use clones of windows machines type windowsOptConfig struct { productKey string adminPassword string domainUser string domain string domainUserPassword string } type cdrom struct { datastore string path string } type memoryAllocation struct { reservation int64 } type virtualMachine struct { name string folder string datacenter string cluster string resourcePool string datastore string vcpu int32 memoryMb int64 memoryAllocation memoryAllocation template string networkInterfaces []networkInterface hardDisks []hardDisk cdroms []cdrom domain string timeZone string dnsSuffixes []string dnsServers []string hasBootableVmdk bool linkedClone bool skipCustomization bool enableDiskUUID bool windowsOptionalConfig windowsOptConfig customConfigurations map[string](types.AnyType) } func (v virtualMachine) Path() string { return vmPath(v.folder, v.name) } func vmPath(folder string, name string) string { var path string if len(folder) > 0 { path += folder + "/" } return path + name } func resourceVSphereVirtualMachine() *schema.Resource { return &schema.Resource{ Create: resourceVSphereVirtualMachineCreate, Read: resourceVSphereVirtualMachineRead, Update: resourceVSphereVirtualMachineUpdate, Delete: resourceVSphereVirtualMachineDelete, SchemaVersion: 1, MigrateState: resourceVSphereVirtualMachineMigrateState, Schema: map[string]*schema.Schema{ "name": &schema.Schema{ Type: schema.TypeString, Required: true, ForceNew: true, }, "folder": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, }, "vcpu": &schema.Schema{ Type: schema.TypeInt, Required: true, }, "memory": &schema.Schema{ Type: schema.TypeInt, Required: true, }, "memory_reservation": &schema.Schema{ Type: schema.TypeInt, Optional: true, Default: 0, ForceNew: true, }, "datacenter": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, }, "cluster": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, }, "resource_pool": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, }, "linked_clone": &schema.Schema{ Type: schema.TypeBool, Optional: true, Default: false, ForceNew: true, }, "gateway": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, Deprecated: "Please use network_interface.ipv4_gateway", }, "domain": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, Default: "vsphere.local", }, "time_zone": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, Default: "Etc/UTC", }, "dns_suffixes": &schema.Schema{ Type: schema.TypeList, Optional: true, Elem: &schema.Schema{Type: schema.TypeString}, ForceNew: true, }, "dns_servers": &schema.Schema{ Type: schema.TypeList, Optional: true, Elem: &schema.Schema{Type: schema.TypeString}, ForceNew: true, }, "skip_customization": &schema.Schema{ Type: schema.TypeBool, Optional: true, ForceNew: true, Default: false, }, "enable_disk_uuid": &schema.Schema{ Type: schema.TypeBool, Optional: true, ForceNew: true, Default: false, }, "uuid": &schema.Schema{ Type: schema.TypeString, Computed: true, }, "custom_configuration_parameters": &schema.Schema{ Type: schema.TypeMap, Optional: true, ForceNew: true, }, "windows_opt_config": &schema.Schema{ Type: schema.TypeList, Optional: true, ForceNew: true, Elem: &schema.Resource{ Schema: map[string]*schema.Schema{ "product_key": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, }, "admin_password": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, }, "domain_user": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, }, "domain": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, }, "domain_user_password": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, }, }, }, }, "network_interface": &schema.Schema{ Type: schema.TypeList, Required: true, ForceNew: true, Elem: &schema.Resource{ Schema: map[string]*schema.Schema{ "label": &schema.Schema{ Type: schema.TypeString, Required: true, ForceNew: true, }, "ip_address": &schema.Schema{ Type: schema.TypeString, Optional: true, Computed: true, Deprecated: "Please use ipv4_address", }, "subnet_mask": &schema.Schema{ Type: schema.TypeString, Optional: true, Computed: true, Deprecated: "Please use ipv4_prefix_length", }, "ipv4_address": &schema.Schema{ Type: schema.TypeString, Optional: true, Computed: true, }, "ipv4_prefix_length": &schema.Schema{ Type: schema.TypeInt, Optional: true, Computed: true, }, "ipv4_gateway": &schema.Schema{ Type: schema.TypeString, Optional: true, Computed: true, }, "ipv6_address": &schema.Schema{ Type: schema.TypeString, Optional: true, Computed: true, }, "ipv6_prefix_length": &schema.Schema{ Type: schema.TypeInt, Optional: true, Computed: true, }, "ipv6_gateway": &schema.Schema{ Type: schema.TypeString, Optional: true, Computed: true, }, "adapter_type": &schema.Schema{ Type: schema.TypeString, Optional: true, ForceNew: true, }, "mac_address": &schema.Schema{ Type: schema.TypeString, Optional: true, Computed: true, }, }, }, }, "disk": &schema.Schema{ Type: schema.TypeSet, Required: true, Elem: &schema.Resource{ Schema: map[string]*schema.Schema{ "uuid": &schema.Schema{ Type: schema.TypeString, Computed: true, }, "key": &schema.Schema{ Type: schema.TypeInt, Computed: true, }, "template": &schema.Schema{ Type: schema.TypeString, Optional: true, }, "type": &schema.Schema{ Type: schema.TypeString, Optional: true, Default: "eager_zeroed", ValidateFunc: func(v interface{}, k string) (ws []string, errors []error) { value := v.(string) if value != "thin" && value != "eager_zeroed" && value != "lazy" { errors = append(errors, fmt.Errorf( "only 'thin', 'eager_zeroed', and 'lazy' are supported values for 'type'")) } return }, }, "datastore": &schema.Schema{ Type: schema.TypeString, Optional: true, }, "size": &schema.Schema{ Type: schema.TypeInt, Optional: true, }, "name": &schema.Schema{ Type: schema.TypeString, Optional: true, }, "iops": &schema.Schema{ Type: schema.TypeInt, Optional: true, }, "vmdk": &schema.Schema{ // TODO: Add ValidateFunc to confirm path exists Type: schema.TypeString, Optional: true, }, "bootable": &schema.Schema{ Type: schema.TypeBool, Optional: true, }, "keep_on_remove": &schema.Schema{ Type: schema.TypeBool, Optional: true, }, "controller_type": &schema.Schema{ Type: schema.TypeString, Optional: true, Default: "scsi", ValidateFunc: func(v interface{}, k string) (ws []string, errors []error) { value := v.(string) found := false for _, t := range DiskControllerTypes { if t == value { found = true } } if !found { errors = append(errors, fmt.Errorf( "Supported values for 'controller_type' are %v", strings.Join(DiskControllerTypes, ", "))) } return }, }, }, }, }, "detach_unknown_disks_on_delete": &schema.Schema{ Type: schema.TypeBool, Optional: true, Default: false, }, "cdrom": &schema.Schema{ Type: schema.TypeList, Optional: true, ForceNew: true, Elem: &schema.Resource{ Schema: map[string]*schema.Schema{ "datastore": &schema.Schema{ Type: schema.TypeString, Required: true, ForceNew: true, }, "path": &schema.Schema{ Type: schema.TypeString, Required: true, ForceNew: true, }, }, }, }, }, } } func resourceVSphereVirtualMachineUpdate(d *schema.ResourceData, meta interface{}) error { // flag if changes have to be applied hasChanges := false // flag if changes have to be done when powered off rebootRequired := false // make config spec configSpec := types.VirtualMachineConfigSpec{} if d.HasChange("vcpu") { configSpec.NumCPUs = int32(d.Get("vcpu").(int)) hasChanges = true rebootRequired = true } if d.HasChange("memory") { configSpec.MemoryMB = int64(d.Get("memory").(int)) hasChanges = true rebootRequired = true } client := meta.(*govmomi.Client) dc, err := getDatacenter(client, d.Get("datacenter").(string)) if err != nil { return err } finder := find.NewFinder(client.Client, true) finder = finder.SetDatacenter(dc) vm, err := finder.VirtualMachine(context.TODO(), vmPath(d.Get("folder").(string), d.Get("name").(string))) if err != nil { return err } if d.HasChange("disk") { hasChanges = true oldDisks, newDisks := d.GetChange("disk") oldDiskSet := oldDisks.(*schema.Set) newDiskSet := newDisks.(*schema.Set) addedDisks := newDiskSet.Difference(oldDiskSet) removedDisks := oldDiskSet.Difference(newDiskSet) // Removed disks for _, diskRaw := range removedDisks.List() { if disk, ok := diskRaw.(map[string]interface{}); ok { devices, err := vm.Device(context.TODO()) if err != nil { return fmt.Errorf("[ERROR] Update Remove Disk - Could not get virtual device list: %v", err) } virtualDisk := devices.FindByKey(int32(disk["key"].(int))) keep := false if v, ok := disk["keep_on_remove"].(bool); ok { keep = v } err = vm.RemoveDevice(context.TODO(), keep, virtualDisk) if err != nil { return fmt.Errorf("[ERROR] Update Remove Disk - Error removing disk: %v", err) } } } // Added disks for _, diskRaw := range addedDisks.List() { if disk, ok := diskRaw.(map[string]interface{}); ok { var datastore *object.Datastore if disk["datastore"] == "" { datastore, err = finder.DefaultDatastore(context.TODO()) if err != nil { return fmt.Errorf("[ERROR] Update Remove Disk - Error finding datastore: %v", err) } } else { datastore, err = finder.Datastore(context.TODO(), disk["datastore"].(string)) if err != nil { log.Printf("[ERROR] Couldn't find datastore %v. %s", disk["datastore"].(string), err) return err } } var size int64 if disk["size"] == 0 { size = 0 } else { size = int64(disk["size"].(int)) } iops := int64(disk["iops"].(int)) controller_type := disk["controller_type"].(string) var mo mo.VirtualMachine vm.Properties(context.TODO(), vm.Reference(), []string{"summary", "config"}, &mo) var diskPath string switch { case disk["vmdk"] != "": diskPath = disk["vmdk"].(string) case disk["name"] != "": snapshotFullDir := mo.Config.Files.SnapshotDirectory split := strings.Split(snapshotFullDir, " ") if len(split) != 2 { return fmt.Errorf("[ERROR] createVirtualMachine - failed to split snapshot directory: %v", snapshotFullDir) } vmWorkingPath := split[1] diskPath = vmWorkingPath + disk["name"].(string) default: return fmt.Errorf("[ERROR] resourceVSphereVirtualMachineUpdate - Neither vmdk path nor vmdk name was given") } var initType string if disk["type"] != "" { initType = disk["type"].(string) } else { initType = "thin" } log.Printf("[INFO] Attaching disk: %v", diskPath) err = addHardDisk(vm, size, iops, initType, datastore, diskPath, controller_type) if err != nil { log.Printf("[ERROR] Add Hard Disk Failed: %v", err) return err } } if err != nil { return err } } } // do nothing if there are no changes if !hasChanges { return nil } log.Printf("[DEBUG] virtual machine config spec: %v", configSpec) if rebootRequired { log.Printf("[INFO] Shutting down virtual machine: %s", d.Id()) task, err := vm.PowerOff(context.TODO()) if err != nil { return err } err = task.Wait(context.TODO()) if err != nil { return err } } log.Printf("[INFO] Reconfiguring virtual machine: %s", d.Id()) task, err := vm.Reconfigure(context.TODO(), configSpec) if err != nil { log.Printf("[ERROR] %s", err) } err = task.Wait(context.TODO()) if err != nil { log.Printf("[ERROR] %s", err) } if rebootRequired { task, err = vm.PowerOn(context.TODO()) if err != nil { return err } err = task.Wait(context.TODO()) if err != nil { log.Printf("[ERROR] %s", err) } } return resourceVSphereVirtualMachineRead(d, meta) } func resourceVSphereVirtualMachineCreate(d *schema.ResourceData, meta interface{}) error { client := meta.(*govmomi.Client) vm := virtualMachine{ name: d.Get("name").(string), vcpu: int32(d.Get("vcpu").(int)), memoryMb: int64(d.Get("memory").(int)), memoryAllocation: memoryAllocation{ reservation: int64(d.Get("memory_reservation").(int)), }, } if v, ok := d.GetOk("folder"); ok { vm.folder = v.(string) } if v, ok := d.GetOk("datacenter"); ok { vm.datacenter = v.(string) } if v, ok := d.GetOk("cluster"); ok { vm.cluster = v.(string) } if v, ok := d.GetOk("resource_pool"); ok { vm.resourcePool = v.(string) } if v, ok := d.GetOk("domain"); ok { vm.domain = v.(string) } if v, ok := d.GetOk("time_zone"); ok { vm.timeZone = v.(string) } if v, ok := d.GetOk("linked_clone"); ok { vm.linkedClone = v.(bool) } if v, ok := d.GetOk("skip_customization"); ok { vm.skipCustomization = v.(bool) } if v, ok := d.GetOk("enable_disk_uuid"); ok { vm.enableDiskUUID = v.(bool) } if raw, ok := d.GetOk("dns_suffixes"); ok { for _, v := range raw.([]interface{}) { vm.dnsSuffixes = append(vm.dnsSuffixes, v.(string)) } } else { vm.dnsSuffixes = DefaultDNSSuffixes } if raw, ok := d.GetOk("dns_servers"); ok { for _, v := range raw.([]interface{}) { vm.dnsServers = append(vm.dnsServers, v.(string)) } } else { vm.dnsServers = DefaultDNSServers } if vL, ok := d.GetOk("custom_configuration_parameters"); ok { if custom_configs, ok := vL.(map[string]interface{}); ok { custom := make(map[string]types.AnyType) for k, v := range custom_configs { custom[k] = v } vm.customConfigurations = custom log.Printf("[DEBUG] custom_configuration_parameters init: %v", vm.customConfigurations) } } if vL, ok := d.GetOk("network_interface"); ok { networks := make([]networkInterface, len(vL.([]interface{}))) for i, v := range vL.([]interface{}) { network := v.(map[string]interface{}) networks[i].label = network["label"].(string) if v, ok := network["ip_address"].(string); ok && v != "" { networks[i].ipv4Address = v } if v, ok := d.GetOk("gateway"); ok { networks[i].ipv4Gateway = v.(string) } if v, ok := network["subnet_mask"].(string); ok && v != "" { ip := net.ParseIP(v).To4() if ip != nil { mask := net.IPv4Mask(ip[0], ip[1], ip[2], ip[3]) pl, _ := mask.Size() networks[i].ipv4PrefixLength = pl } else { return fmt.Errorf("subnet_mask parameter is invalid.") } } if v, ok := network["ipv4_address"].(string); ok && v != "" { networks[i].ipv4Address = v } if v, ok := network["ipv4_prefix_length"].(int); ok && v != 0 { networks[i].ipv4PrefixLength = v } if v, ok := network["ipv4_gateway"].(string); ok && v != "" { networks[i].ipv4Gateway = v } if v, ok := network["ipv6_address"].(string); ok && v != "" { networks[i].ipv6Address = v } if v, ok := network["ipv6_prefix_length"].(int); ok && v != 0 { networks[i].ipv6PrefixLength = v } if v, ok := network["ipv6_gateway"].(string); ok && v != "" { networks[i].ipv6Gateway = v } if v, ok := network["mac_address"].(string); ok && v != "" { networks[i].macAddress = v } } vm.networkInterfaces = networks log.Printf("[DEBUG] network_interface init: %v", networks) } if vL, ok := d.GetOk("windows_opt_config"); ok { var winOpt windowsOptConfig custom_configs := (vL.([]interface{}))[0].(map[string]interface{}) if v, ok := custom_configs["admin_password"].(string); ok && v != "" { winOpt.adminPassword = v } if v, ok := custom_configs["domain"].(string); ok && v != "" { winOpt.domain = v } if v, ok := custom_configs["domain_user"].(string); ok && v != "" { winOpt.domainUser = v } if v, ok := custom_configs["product_key"].(string); ok && v != "" { winOpt.productKey = v } if v, ok := custom_configs["domain_user_password"].(string); ok && v != "" { winOpt.domainUserPassword = v } vm.windowsOptionalConfig = winOpt log.Printf("[DEBUG] windows config init: %v", winOpt) } if vL, ok := d.GetOk("disk"); ok { if diskSet, ok := vL.(*schema.Set); ok { disks := []hardDisk{} for _, value := range diskSet.List() { disk := value.(map[string]interface{}) newDisk := hardDisk{} if v, ok := disk["template"].(string); ok && v != "" { if v, ok := disk["name"].(string); ok && v != "" { return fmt.Errorf("Cannot specify name of a template") } vm.template = v if vm.hasBootableVmdk { return fmt.Errorf("[ERROR] Only one bootable disk or template may be given") } vm.hasBootableVmdk = true } if v, ok := disk["type"].(string); ok && v != "" { newDisk.initType = v } if v, ok := disk["datastore"].(string); ok && v != "" { vm.datastore = v } if v, ok := disk["size"].(int); ok && v != 0 { if v, ok := disk["template"].(string); ok && v != "" { return fmt.Errorf("Cannot specify size of a template") } if v, ok := disk["name"].(string); ok && v != "" { newDisk.name = v } else { return fmt.Errorf("[ERROR] Disk name must be provided when creating a new disk") } newDisk.size = int64(v) } if v, ok := disk["iops"].(int); ok && v != 0 { newDisk.iops = int64(v) } if v, ok := disk["controller_type"].(string); ok && v != "" { newDisk.controller = v } if vVmdk, ok := disk["vmdk"].(string); ok && vVmdk != "" { if v, ok := disk["template"].(string); ok && v != "" { return fmt.Errorf("Cannot specify a vmdk for a template") } if v, ok := disk["size"].(string); ok && v != "" { return fmt.Errorf("Cannot specify size of a vmdk") } if v, ok := disk["name"].(string); ok && v != "" { return fmt.Errorf("Cannot specify name of a vmdk") } if vBootable, ok := disk["bootable"].(bool); ok { if vBootable && vm.hasBootableVmdk { return fmt.Errorf("[ERROR] Only one bootable disk or template may be given") } newDisk.bootable = vBootable vm.hasBootableVmdk = vm.hasBootableVmdk || vBootable } newDisk.vmdkPath = vVmdk } // Preserves order so bootable disk is first if newDisk.bootable == true || disk["template"] != "" { disks = append([]hardDisk{newDisk}, disks...) } else { disks = append(disks, newDisk) } } vm.hardDisks = disks log.Printf("[DEBUG] disk init: %v", disks) } } if vL, ok := d.GetOk("cdrom"); ok { cdroms := make([]cdrom, len(vL.([]interface{}))) for i, v := range vL.([]interface{}) { c := v.(map[string]interface{}) if v, ok := c["datastore"].(string); ok && v != "" { cdroms[i].datastore = v } else { return fmt.Errorf("Datastore argument must be specified when attaching a cdrom image.") } if v, ok := c["path"].(string); ok && v != "" { cdroms[i].path = v } else { return fmt.Errorf("Path argument must be specified when attaching a cdrom image.") } } vm.cdroms = cdroms log.Printf("[DEBUG] cdrom init: %v", cdroms) } err := vm.setupVirtualMachine(client) if err != nil { return err } d.SetId(vm.Path()) log.Printf("[INFO] Created virtual machine: %s", d.Id()) return resourceVSphereVirtualMachineRead(d, meta) } func resourceVSphereVirtualMachineRead(d *schema.ResourceData, meta interface{}) error { log.Printf("[DEBUG] virtual machine resource data: %#v", d) client := meta.(*govmomi.Client) dc, err := getDatacenter(client, d.Get("datacenter").(string)) if err != nil { return err } finder := find.NewFinder(client.Client, true) finder = finder.SetDatacenter(dc) vm, err := finder.VirtualMachine(context.TODO(), d.Id()) if err != nil { d.SetId("") return nil } state, err := vm.PowerState(context.TODO()) if err != nil { return err } if state == types.VirtualMachinePowerStatePoweredOn { // wait for interfaces to appear log.Printf("[DEBUG] Waiting for interfaces to appear") _, err = vm.WaitForNetIP(context.TODO(), false) if err != nil { return err } log.Printf("[DEBUG] Successfully waited for interfaces to appear") } var mvm mo.VirtualMachine collector := property.DefaultCollector(client.Client) if err := collector.RetrieveOne(context.TODO(), vm.Reference(), []string{"guest", "summary", "datastore", "config"}, &mvm); err != nil { return err } log.Printf("[DEBUG] Datacenter - %#v", dc) log.Printf("[DEBUG] mvm.Summary.Config - %#v", mvm.Summary.Config) log.Printf("[DEBUG] mvm.Summary.Config - %#v", mvm.Config) log.Printf("[DEBUG] mvm.Guest.Net - %#v", mvm.Guest.Net) disks := make([]map[string]interface{}, 0) templateDisk := make(map[string]interface{}, 1) for _, device := range mvm.Config.Hardware.Device { if vd, ok := device.(*types.VirtualDisk); ok { virtualDevice := vd.GetVirtualDevice() backingInfo := virtualDevice.Backing var diskFullPath string var diskUuid string if v, ok := backingInfo.(*types.VirtualDiskFlatVer2BackingInfo); ok { diskFullPath = v.FileName diskUuid = v.Uuid } else if v, ok := backingInfo.(*types.VirtualDiskSparseVer2BackingInfo); ok { diskFullPath = v.FileName diskUuid = v.Uuid } log.Printf("[DEBUG] resourceVSphereVirtualMachineRead - Analyzing disk: %v", diskFullPath) // Separate datastore and path diskFullPathSplit := strings.Split(diskFullPath, " ") if len(diskFullPathSplit) != 2 { return fmt.Errorf("[ERROR] Failed trying to parse disk path: %v", diskFullPath) } diskPath := diskFullPathSplit[1] // Isolate filename diskNameSplit := strings.Split(diskPath, "/") diskName := diskNameSplit[len(diskNameSplit)-1] // Remove possible extension diskName = strings.Split(diskName, ".")[0] if prevDisks, ok := d.GetOk("disk"); ok { if prevDiskSet, ok := prevDisks.(*schema.Set); ok { for _, v := range prevDiskSet.List() { prevDisk := v.(map[string]interface{}) // We're guaranteed only one template disk. Passing value directly through since templates should be immutable if prevDisk["template"] != "" { if len(templateDisk) == 0 { templateDisk = prevDisk disks = append(disks, templateDisk) break } } // It is enforced that prevDisk["name"] should only be set in the case // of creating a new disk for the user. // size case: name was set by user, compare parsed filename from mo.filename (without path or .vmdk extension) with name // vmdk case: compare prevDisk["vmdk"] and mo.Filename if diskName == prevDisk["name"] || diskPath == prevDisk["vmdk"] { prevDisk["key"] = virtualDevice.Key prevDisk["uuid"] = diskUuid disks = append(disks, prevDisk) break } } } } log.Printf("[DEBUG] disks: %#v", disks) } } err = d.Set("disk", disks) if err != nil { return fmt.Errorf("Invalid disks to set: %#v", disks) } networkInterfaces := make([]map[string]interface{}, 0) for _, v := range mvm.Guest.Net { if v.DeviceConfigId >= 0 { log.Printf("[DEBUG] v.Network - %#v", v.Network) networkInterface := make(map[string]interface{}) networkInterface["label"] = v.Network networkInterface["mac_address"] = v.MacAddress for _, ip := range v.IpConfig.IpAddress { p := net.ParseIP(ip.IpAddress) if p.To4() != nil { log.Printf("[DEBUG] p.String - %#v", p.String()) log.Printf("[DEBUG] ip.PrefixLength - %#v", ip.PrefixLength) networkInterface["ipv4_address"] = p.String() networkInterface["ipv4_prefix_length"] = ip.PrefixLength } else if p.To16() != nil { log.Printf("[DEBUG] p.String - %#v", p.String()) log.Printf("[DEBUG] ip.PrefixLength - %#v", ip.PrefixLength) networkInterface["ipv6_address"] = p.String() networkInterface["ipv6_prefix_length"] = ip.PrefixLength } log.Printf("[DEBUG] networkInterface: %#v", networkInterface) } log.Printf("[DEBUG] networkInterface: %#v", networkInterface) networkInterfaces = append(networkInterfaces, networkInterface) } } if mvm.Guest.IpStack != nil { for _, v := range mvm.Guest.IpStack { if v.IpRouteConfig != nil && v.IpRouteConfig.IpRoute != nil { for _, route := range v.IpRouteConfig.IpRoute { if route.Gateway.Device != "" { gatewaySetting := "" if route.Network == "::" { gatewaySetting = "ipv6_gateway" } else if route.Network == "0.0.0.0" { gatewaySetting = "ipv4_gateway" } if gatewaySetting != "" { deviceID, err := strconv.Atoi(route.Gateway.Device) if len(networkInterfaces) == 1 { deviceID = 0 } if err != nil { log.Printf("[WARN] error at processing %s of device id %#v: %#v", gatewaySetting, route.Gateway.Device, err) } else { log.Printf("[DEBUG] %s of device id %d: %s", gatewaySetting, deviceID, route.Gateway.IpAddress) networkInterfaces[deviceID][gatewaySetting] = route.Gateway.IpAddress } } } } } } } log.Printf("[DEBUG] networkInterfaces: %#v", networkInterfaces) err = d.Set("network_interface", networkInterfaces) if err != nil { return fmt.Errorf("Invalid network interfaces to set: %#v", networkInterfaces) } if len(networkInterfaces) > 0 { if _, ok := networkInterfaces[0]["ipv4_address"]; ok { log.Printf("[DEBUG] ip address: %v", networkInterfaces[0]["ipv4_address"].(string)) d.SetConnInfo(map[string]string{ "type": "ssh", "host": networkInterfaces[0]["ipv4_address"].(string), }) } } var rootDatastore string for _, v := range mvm.Datastore { var md mo.Datastore if err := collector.RetrieveOne(context.TODO(), v, []string{"name", "parent"}, &md); err != nil { return err } if md.Parent.Type == "StoragePod" { var msp mo.StoragePod if err := collector.RetrieveOne(context.TODO(), *md.Parent, []string{"name"}, &msp); err != nil { return err } rootDatastore = msp.Name log.Printf("[DEBUG] %#v", msp.Name) } else { rootDatastore = md.Name log.Printf("[DEBUG] %#v", md.Name) } break } d.Set("datacenter", dc) d.Set("memory", mvm.Summary.Config.MemorySizeMB) d.Set("memory_reservation", mvm.Summary.Config.MemoryReservation) d.Set("cpu", mvm.Summary.Config.NumCpu) d.Set("datastore", rootDatastore) d.Set("uuid", mvm.Summary.Config.Uuid) return nil } func resourceVSphereVirtualMachineDelete(d *schema.ResourceData, meta interface{}) error { client := meta.(*govmomi.Client) dc, err := getDatacenter(client, d.Get("datacenter").(string)) if err != nil { return err } finder := find.NewFinder(client.Client, true) finder = finder.SetDatacenter(dc) vm, err := finder.VirtualMachine(context.TODO(), vmPath(d.Get("folder").(string), d.Get("name").(string))) if err != nil { return err } devices, err := vm.Device(context.TODO()) if err != nil { log.Printf("[DEBUG] resourceVSphereVirtualMachineDelete - Failed to get device list: %v", err) return err } log.Printf("[INFO] Deleting virtual machine: %s", d.Id()) state, err := vm.PowerState(context.TODO()) if err != nil { return err } if state == types.VirtualMachinePowerStatePoweredOn { task, err := vm.PowerOff(context.TODO()) if err != nil { return err } err = task.Wait(context.TODO()) if err != nil { return err } } // Safely eject any disks the user marked as keep_on_remove var diskSetList []interface{} if vL, ok := d.GetOk("disk"); ok { if diskSet, ok := vL.(*schema.Set); ok { diskSetList = diskSet.List() for _, value := range diskSetList { disk := value.(map[string]interface{}) if v, ok := disk["keep_on_remove"].(bool); ok && v == true { log.Printf("[DEBUG] not destroying %v", disk["name"]) virtualDisk := devices.FindByKey(int32(disk["key"].(int))) err = vm.RemoveDevice(context.TODO(), true, virtualDisk) if err != nil { log.Printf("[ERROR] Update Remove Disk - Error removing disk: %v", err) return err } } } } } // Safely eject any disks that are not managed by this resource if v, ok := d.GetOk("detach_unknown_disks_on_delete"); ok && v.(bool) { var disksToRemove object.VirtualDeviceList for _, device := range devices { if devices.TypeName(device) != "VirtualDisk" { continue } vd := device.GetVirtualDevice() var skip bool for _, value := range diskSetList { disk := value.(map[string]interface{}) if int32(disk["key"].(int)) == vd.Key { skip = true break } } if skip { continue } disksToRemove = append(disksToRemove, device) } if len(disksToRemove) != 0 { err = vm.RemoveDevice(context.TODO(), true, disksToRemove...) if err != nil { log.Printf("[ERROR] Update Remove Disk - Error removing disk: %v", err) return err } } } task, err := vm.Destroy(context.TODO()) if err != nil { return err } err = task.Wait(context.TODO()) if err != nil { return err } d.SetId("") return nil } // addHardDisk adds a new Hard Disk to the VirtualMachine. func addHardDisk(vm *object.VirtualMachine, size, iops int64, diskType string, datastore *object.Datastore, diskPath string, controller_type string) error { devices, err := vm.Device(context.TODO()) if err != nil { return err } log.Printf("[DEBUG] vm devices: %#v\n", devices) var controller types.BaseVirtualController switch controller_type { case "scsi": controller, err = devices.FindDiskController(controller_type) case "scsi-lsi-parallel": controller = devices.PickController(&types.VirtualLsiLogicController{}) case "scsi-buslogic": controller = devices.PickController(&types.VirtualBusLogicController{}) case "scsi-paravirtual": controller = devices.PickController(&types.ParaVirtualSCSIController{}) case "scsi-lsi-sas": controller = devices.PickController(&types.VirtualLsiLogicSASController{}) case "ide": controller, err = devices.FindDiskController(controller_type) default: return fmt.Errorf("[ERROR] Unsupported disk controller provided: %v", controller_type) } if err != nil || controller == nil { // Check if max number of scsi controller are already used diskControllers := getSCSIControllers(devices) if len(diskControllers) >= 4 { return fmt.Errorf("[ERROR] Maximum number of SCSI controllers created") } log.Printf("[DEBUG] Couldn't find a %v controller. Creating one..", controller_type) var c types.BaseVirtualDevice switch controller_type { case "scsi": // Create scsi controller c, err = devices.CreateSCSIController("scsi") if err != nil { return fmt.Errorf("[ERROR] Failed creating SCSI controller: %v", err) } case "scsi-lsi-parallel": // Create scsi controller c, err = devices.CreateSCSIController("lsilogic") if err != nil { return fmt.Errorf("[ERROR] Failed creating SCSI controller: %v", err) } case "scsi-buslogic": // Create scsi controller c, err = devices.CreateSCSIController("buslogic") if err != nil { return fmt.Errorf("[ERROR] Failed creating SCSI controller: %v", err) } case "scsi-paravirtual": // Create scsi controller c, err = devices.CreateSCSIController("pvscsi") if err != nil { return fmt.Errorf("[ERROR] Failed creating SCSI controller: %v", err) } case "scsi-lsi-sas": // Create scsi controller c, err = devices.CreateSCSIController("lsilogic-sas") if err != nil { return fmt.Errorf("[ERROR] Failed creating SCSI controller: %v", err) } case "ide": // Create ide controller c, err = devices.CreateIDEController() if err != nil { return fmt.Errorf("[ERROR] Failed creating IDE controller: %v", err) } default: return fmt.Errorf("[ERROR] Unsupported disk controller provided: %v", controller_type) } vm.AddDevice(context.TODO(), c) // Update our devices list devices, err := vm.Device(context.TODO()) if err != nil { return err } controller = devices.PickController(c.(types.BaseVirtualController)) if controller == nil { log.Printf("[ERROR] Could not find the new %v controller", controller_type) return fmt.Errorf("Could not find the new %v controller", controller_type) } } log.Printf("[DEBUG] disk controller: %#v\n", controller) // TODO Check if diskPath & datastore exist // If diskPath is not specified, pass empty string to CreateDisk() if diskPath == "" { return fmt.Errorf("[ERROR] addHardDisk - No path proided") } else { diskPath = datastore.Path(diskPath) } log.Printf("[DEBUG] addHardDisk - diskPath: %v", diskPath) disk := devices.CreateDisk(controller, datastore.Reference(), diskPath) if strings.Contains(controller_type, "scsi") { unitNumber, err := getNextUnitNumber(devices, controller) if err != nil { return err } *disk.UnitNumber = unitNumber } existing := devices.SelectByBackingInfo(disk.Backing) log.Printf("[DEBUG] disk: %#v\n", disk) if len(existing) == 0 { disk.CapacityInKB = int64(size * 1024 * 1024) if iops != 0 { disk.StorageIOAllocation = &types.StorageIOAllocationInfo{ Limit: iops, } } backing := disk.Backing.(*types.VirtualDiskFlatVer2BackingInfo) if diskType == "eager_zeroed" { // eager zeroed thick virtual disk backing.ThinProvisioned = types.NewBool(false) backing.EagerlyScrub = types.NewBool(true) } else if diskType == "lazy" { // lazy zeroed thick virtual disk backing.ThinProvisioned = types.NewBool(false) backing.EagerlyScrub = types.NewBool(false) } else if diskType == "thin" { // thin provisioned virtual disk backing.ThinProvisioned = types.NewBool(true) } log.Printf("[DEBUG] addHardDisk: %#v\n", disk) log.Printf("[DEBUG] addHardDisk capacity: %#v\n", disk.CapacityInKB) return vm.AddDevice(context.TODO(), disk) } else { log.Printf("[DEBUG] addHardDisk: Disk already present.\n") return nil } } func getSCSIControllers(vmDevices object.VirtualDeviceList) []*types.VirtualController { // get virtual scsi controllers of all supported types var scsiControllers []*types.VirtualController for _, device := range vmDevices { devType := vmDevices.Type(device) switch devType { case "scsi", "lsilogic", "buslogic", "pvscsi", "lsilogic-sas": if c, ok := device.(types.BaseVirtualController); ok { scsiControllers = append(scsiControllers, c.GetVirtualController()) } } } return scsiControllers } func getNextUnitNumber(devices object.VirtualDeviceList, c types.BaseVirtualController) (int32, error) { key := c.GetVirtualController().Key var unitNumbers [16]bool unitNumbers[7] = true for _, device := range devices { d := device.GetVirtualDevice() if d.ControllerKey == key { if d.UnitNumber != nil { unitNumbers[*d.UnitNumber] = true } } } for i, taken := range unitNumbers { if !taken { return int32(i), nil } } return -1, fmt.Errorf("[ERROR] getNextUnitNumber - controller is full") } // addCdrom adds a new virtual cdrom drive to the VirtualMachine and attaches an image (ISO) to it from a datastore path. func addCdrom(client *govmomi.Client, vm *object.VirtualMachine, datacenter *object.Datacenter, datastore, path string) error { devices, err := vm.Device(context.TODO()) if err != nil { return err } log.Printf("[DEBUG] vm devices: %#v", devices) var controller *types.VirtualIDEController controller, err = devices.FindIDEController("") if err != nil { log.Printf("[DEBUG] Couldn't find a ide controller. Creating one..") var c types.BaseVirtualDevice c, err := devices.CreateIDEController() if err != nil { return fmt.Errorf("[ERROR] Failed creating IDE controller: %v", err) } if v, ok := c.(*types.VirtualIDEController); ok { controller = v } else { return fmt.Errorf("[ERROR] Controller type could not be asserted") } vm.AddDevice(context.TODO(), c) // Update our devices list devices, err := vm.Device(context.TODO()) if err != nil { return err } controller, err = devices.FindIDEController("") if err != nil { log.Printf("[ERROR] Could not find the new disk IDE controller: %v", err) return err } } log.Printf("[DEBUG] ide controller: %#v", controller) c, err := devices.CreateCdrom(controller) if err != nil { return err } finder := find.NewFinder(client.Client, true) finder = finder.SetDatacenter(datacenter) ds, err := getDatastore(finder, datastore) if err != nil { return err } c = devices.InsertIso(c, ds.Path(path)) log.Printf("[DEBUG] addCdrom: %#v", c) return vm.AddDevice(context.TODO(), c) } // buildNetworkDevice builds VirtualDeviceConfigSpec for Network Device. func buildNetworkDevice(f *find.Finder, label, adapterType string, macAddress string) (*types.VirtualDeviceConfigSpec, error) { network, err := f.Network(context.TODO(), "*"+label) if err != nil { return nil, err } backing, err := network.EthernetCardBackingInfo(context.TODO()) if err != nil { return nil, err } var address_type string if macAddress == "" { address_type = string(types.VirtualEthernetCardMacTypeGenerated) } else { address_type = string(types.VirtualEthernetCardMacTypeManual) } if adapterType == "vmxnet3" { return &types.VirtualDeviceConfigSpec{ Operation: types.VirtualDeviceConfigSpecOperationAdd, Device: &types.VirtualVmxnet3{ VirtualVmxnet: types.VirtualVmxnet{ VirtualEthernetCard: types.VirtualEthernetCard{ VirtualDevice: types.VirtualDevice{ Key: -1, Backing: backing, }, AddressType: address_type, MacAddress: macAddress, }, }, }, }, nil } else if adapterType == "e1000" { return &types.VirtualDeviceConfigSpec{ Operation: types.VirtualDeviceConfigSpecOperationAdd, Device: &types.VirtualE1000{ VirtualEthernetCard: types.VirtualEthernetCard{ VirtualDevice: types.VirtualDevice{ Key: -1, Backing: backing, }, AddressType: address_type, MacAddress: macAddress, }, }, }, nil } else { return nil, fmt.Errorf("Invalid network adapter type.") } } // buildVMRelocateSpec builds VirtualMachineRelocateSpec to set a place for a new VirtualMachine. func buildVMRelocateSpec(rp *object.ResourcePool, ds *object.Datastore, vm *object.VirtualMachine, linkedClone bool, initType string) (types.VirtualMachineRelocateSpec, error) { var key int32 var moveType string if linkedClone { moveType = "createNewChildDiskBacking" } else { moveType = "moveAllDiskBackingsAndDisallowSharing" } log.Printf("[DEBUG] relocate type: [%s]", moveType) devices, err := vm.Device(context.TODO()) if err != nil { return types.VirtualMachineRelocateSpec{}, err } for _, d := range devices { if devices.Type(d) == "disk" { key = int32(d.GetVirtualDevice().Key) } } isThin := initType == "thin" eagerScrub := initType == "eager_zeroed" rpr := rp.Reference() dsr := ds.Reference() return types.VirtualMachineRelocateSpec{ Datastore: &dsr, Pool: &rpr, DiskMoveType: moveType, Disk: []types.VirtualMachineRelocateSpecDiskLocator{ { Datastore: dsr, DiskBackingInfo: &types.VirtualDiskFlatVer2BackingInfo{ DiskMode: "persistent", ThinProvisioned: types.NewBool(isThin), EagerlyScrub: types.NewBool(eagerScrub), }, DiskId: key, }, }, }, nil } // getDatastoreObject gets datastore object. func getDatastoreObject(client *govmomi.Client, f *object.DatacenterFolders, name string) (types.ManagedObjectReference, error) { s := object.NewSearchIndex(client.Client) ref, err := s.FindChild(context.TODO(), f.DatastoreFolder, name) if err != nil { return types.ManagedObjectReference{}, err } if ref == nil { return types.ManagedObjectReference{}, fmt.Errorf("Datastore '%s' not found.", name) } log.Printf("[DEBUG] getDatastoreObject: reference: %#v", ref) return ref.Reference(), nil } // buildStoragePlacementSpecCreate builds StoragePlacementSpec for create action. func buildStoragePlacementSpecCreate(f *object.DatacenterFolders, rp *object.ResourcePool, storagePod object.StoragePod, configSpec types.VirtualMachineConfigSpec) types.StoragePlacementSpec { vmfr := f.VmFolder.Reference() rpr := rp.Reference() spr := storagePod.Reference() sps := types.StoragePlacementSpec{ Type: "create", ConfigSpec: &configSpec, PodSelectionSpec: types.StorageDrsPodSelectionSpec{ StoragePod: &spr, }, Folder: &vmfr, ResourcePool: &rpr, } log.Printf("[DEBUG] findDatastore: StoragePlacementSpec: %#v\n", sps) return sps } // buildStoragePlacementSpecClone builds StoragePlacementSpec for clone action. func buildStoragePlacementSpecClone(c *govmomi.Client, f *object.DatacenterFolders, vm *object.VirtualMachine, rp *object.ResourcePool, storagePod object.StoragePod) types.StoragePlacementSpec { vmr := vm.Reference() vmfr := f.VmFolder.Reference() rpr := rp.Reference() spr := storagePod.Reference() var o mo.VirtualMachine err := vm.Properties(context.TODO(), vmr, []string{"datastore"}, &o) if err != nil { return types.StoragePlacementSpec{} } ds := object.NewDatastore(c.Client, o.Datastore[0]) log.Printf("[DEBUG] findDatastore: datastore: %#v\n", ds) devices, err := vm.Device(context.TODO()) if err != nil { return types.StoragePlacementSpec{} } var key int32 for _, d := range devices.SelectByType((*types.VirtualDisk)(nil)) { key = int32(d.GetVirtualDevice().Key) log.Printf("[DEBUG] findDatastore: virtual devices: %#v\n", d.GetVirtualDevice()) } sps := types.StoragePlacementSpec{ Type: "clone", Vm: &vmr, PodSelectionSpec: types.StorageDrsPodSelectionSpec{ StoragePod: &spr, }, CloneSpec: &types.VirtualMachineCloneSpec{ Location: types.VirtualMachineRelocateSpec{ Disk: []types.VirtualMachineRelocateSpecDiskLocator{ { Datastore: ds.Reference(), DiskBackingInfo: &types.VirtualDiskFlatVer2BackingInfo{}, DiskId: key, }, }, Pool: &rpr, }, PowerOn: false, Template: false, }, CloneName: "dummy", Folder: &vmfr, } return sps } // findDatastore finds Datastore object. func findDatastore(c *govmomi.Client, sps types.StoragePlacementSpec) (*object.Datastore, error) { var datastore *object.Datastore log.Printf("[DEBUG] findDatastore: StoragePlacementSpec: %#v\n", sps) srm := object.NewStorageResourceManager(c.Client) rds, err := srm.RecommendDatastores(context.TODO(), sps) if err != nil { return nil, err } log.Printf("[DEBUG] findDatastore: recommendDatastores: %#v\n", rds) spa := rds.Recommendations[0].Action[0].(*types.StoragePlacementAction) datastore = object.NewDatastore(c.Client, spa.Destination) log.Printf("[DEBUG] findDatastore: datastore: %#v", datastore) return datastore, nil } // createCdroms is a helper function to attach virtual cdrom devices (and their attached disk images) to a virtual IDE controller. func createCdroms(client *govmomi.Client, vm *object.VirtualMachine, datacenter *object.Datacenter, cdroms []cdrom) error { log.Printf("[DEBUG] add cdroms: %v", cdroms) for _, cd := range cdroms { log.Printf("[DEBUG] add cdrom (datastore): %v", cd.datastore) log.Printf("[DEBUG] add cdrom (cd path): %v", cd.path) err := addCdrom(client, vm, datacenter, cd.datastore, cd.path) if err != nil { return err } } return nil } func (vm *virtualMachine) setupVirtualMachine(c *govmomi.Client) error { dc, err := getDatacenter(c, vm.datacenter) if err != nil { return err } finder := find.NewFinder(c.Client, true) finder = finder.SetDatacenter(dc) var template *object.VirtualMachine var template_mo mo.VirtualMachine var vm_mo mo.VirtualMachine if vm.template != "" { template, err = finder.VirtualMachine(context.TODO(), vm.template) if err != nil { return err } log.Printf("[DEBUG] template: %#v", template) err = template.Properties(context.TODO(), template.Reference(), []string{"parent", "config.template", "config.guestId", "resourcePool", "snapshot", "guest.toolsVersionStatus2", "config.guestFullName"}, &template_mo) if err != nil { return err } } var resourcePool *object.ResourcePool if vm.resourcePool == "" { if vm.cluster == "" { resourcePool, err = finder.DefaultResourcePool(context.TODO()) if err != nil { return err } } else { resourcePool, err = finder.ResourcePool(context.TODO(), "*"+vm.cluster+"/Resources") if err != nil { return err } } } else { resourcePool, err = finder.ResourcePool(context.TODO(), vm.resourcePool) if err != nil { return err } } log.Printf("[DEBUG] resource pool: %#v", resourcePool) dcFolders, err := dc.Folders(context.TODO()) if err != nil { return err } log.Printf("[DEBUG] folder: %#v", vm.folder) folder := dcFolders.VmFolder if len(vm.folder) > 0 { si := object.NewSearchIndex(c.Client) folderRef, err := si.FindByInventoryPath( context.TODO(), fmt.Sprintf("%v/vm/%v", vm.datacenter, vm.folder)) if err != nil { return fmt.Errorf("Error reading folder %s: %s", vm.folder, err) } else if folderRef == nil { return fmt.Errorf("Cannot find folder %s", vm.folder) } else { folder = folderRef.(*object.Folder) } } // make config spec configSpec := types.VirtualMachineConfigSpec{ Name: vm.name, NumCPUs: vm.vcpu, NumCoresPerSocket: 1, MemoryMB: vm.memoryMb, MemoryAllocation: &types.ResourceAllocationInfo{ Reservation: vm.memoryAllocation.reservation, }, Flags: &types.VirtualMachineFlagInfo{ DiskUuidEnabled: &vm.enableDiskUUID, }, } if vm.template == "" { configSpec.GuestId = "otherLinux64Guest" } log.Printf("[DEBUG] virtual machine config spec: %v", configSpec) // make ExtraConfig log.Printf("[DEBUG] virtual machine Extra Config spec start") if len(vm.customConfigurations) > 0 { var ov []types.BaseOptionValue for k, v := range vm.customConfigurations { key := k value := v o := types.OptionValue{ Key: key, Value: &value, } log.Printf("[DEBUG] virtual machine Extra Config spec: %s,%s", k, v) ov = append(ov, &o) } configSpec.ExtraConfig = ov log.Printf("[DEBUG] virtual machine Extra Config spec: %v", configSpec.ExtraConfig) } var datastore *object.Datastore if vm.datastore == "" { datastore, err = finder.DefaultDatastore(context.TODO()) if err != nil { return err } } else { datastore, err = finder.Datastore(context.TODO(), vm.datastore) if err != nil { // TODO: datastore cluster support in govmomi finder function d, err := getDatastoreObject(c, dcFolders, vm.datastore) if err != nil { return err } if d.Type == "StoragePod" { sp := object.StoragePod{ Folder: object.NewFolder(c.Client, d), } var sps types.StoragePlacementSpec if vm.template != "" { sps = buildStoragePlacementSpecClone(c, dcFolders, template, resourcePool, sp) } else { sps = buildStoragePlacementSpecCreate(dcFolders, resourcePool, sp, configSpec) } datastore, err = findDatastore(c, sps) if err != nil { return err } } else { datastore = object.NewDatastore(c.Client, d) } } } log.Printf("[DEBUG] datastore: %#v", datastore) // network networkDevices := []types.BaseVirtualDeviceConfigSpec{} networkConfigs := []types.CustomizationAdapterMapping{} for _, network := range vm.networkInterfaces { // network device var networkDeviceType string if vm.template == "" { networkDeviceType = "e1000" } else { networkDeviceType = "vmxnet3" } nd, err := buildNetworkDevice(finder, network.label, networkDeviceType, network.macAddress) if err != nil { return err } log.Printf("[DEBUG] network device: %+v", nd.Device) networkDevices = append(networkDevices, nd) if vm.template != "" { var ipSetting types.CustomizationIPSettings if network.ipv4Address == "" { ipSetting.Ip = &types.CustomizationDhcpIpGenerator{} } else { if network.ipv4PrefixLength == 0 { return fmt.Errorf("Error: ipv4_prefix_length argument is empty.") } m := net.CIDRMask(network.ipv4PrefixLength, 32) sm := net.IPv4(m[0], m[1], m[2], m[3]) subnetMask := sm.String() log.Printf("[DEBUG] ipv4 gateway: %v\n", network.ipv4Gateway) log.Printf("[DEBUG] ipv4 address: %v\n", network.ipv4Address) log.Printf("[DEBUG] ipv4 prefix length: %v\n", network.ipv4PrefixLength) log.Printf("[DEBUG] ipv4 subnet mask: %v\n", subnetMask) ipSetting.Gateway = []string{ network.ipv4Gateway, } ipSetting.Ip = &types.CustomizationFixedIp{ IpAddress: network.ipv4Address, } ipSetting.SubnetMask = subnetMask } ipv6Spec := &types.CustomizationIPSettingsIpV6AddressSpec{} if network.ipv6Address == "" { ipv6Spec.Ip = []types.BaseCustomizationIpV6Generator{ &types.CustomizationDhcpIpV6Generator{}, } } else { log.Printf("[DEBUG] ipv6 gateway: %v\n", network.ipv6Gateway) log.Printf("[DEBUG] ipv6 address: %v\n", network.ipv6Address) log.Printf("[DEBUG] ipv6 prefix length: %v\n", network.ipv6PrefixLength) ipv6Spec.Ip = []types.BaseCustomizationIpV6Generator{ &types.CustomizationFixedIpV6{ IpAddress: network.ipv6Address, SubnetMask: int32(network.ipv6PrefixLength), }, } ipv6Spec.Gateway = []string{network.ipv6Gateway} } ipSetting.IpV6Spec = ipv6Spec // network config config := types.CustomizationAdapterMapping{ Adapter: ipSetting, } networkConfigs = append(networkConfigs, config) } } log.Printf("[DEBUG] network devices: %#v", networkDevices) log.Printf("[DEBUG] network configs: %#v", networkConfigs) var task *object.Task if vm.template == "" { var mds mo.Datastore if err = datastore.Properties(context.TODO(), datastore.Reference(), []string{"name"}, &mds); err != nil { return err } log.Printf("[DEBUG] datastore: %#v", mds.Name) scsi, err := object.SCSIControllerTypes().CreateSCSIController("scsi") if err != nil { log.Printf("[ERROR] %s", err) } configSpec.DeviceChange = append(configSpec.DeviceChange, &types.VirtualDeviceConfigSpec{ Operation: types.VirtualDeviceConfigSpecOperationAdd, Device: scsi, }) configSpec.Files = &types.VirtualMachineFileInfo{VmPathName: fmt.Sprintf("[%s]", mds.Name)} task, err = folder.CreateVM(context.TODO(), configSpec, resourcePool, nil) if err != nil { log.Printf("[ERROR] %s", err) } err = task.Wait(context.TODO()) if err != nil { log.Printf("[ERROR] %s", err) } } else { relocateSpec, err := buildVMRelocateSpec(resourcePool, datastore, template, vm.linkedClone, vm.hardDisks[0].initType) if err != nil { return err } log.Printf("[DEBUG] relocate spec: %v", relocateSpec) // make vm clone spec cloneSpec := types.VirtualMachineCloneSpec{ Location: relocateSpec, Template: false, Config: &configSpec, PowerOn: false, } if vm.linkedClone { if template_mo.Snapshot == nil { return fmt.Errorf("`linkedClone=true`, but image VM has no snapshots") } cloneSpec.Snapshot = template_mo.Snapshot.CurrentSnapshot } log.Printf("[DEBUG] clone spec: %v", cloneSpec) task, err = template.Clone(context.TODO(), folder, vm.name, cloneSpec) if err != nil { return err } } err = task.Wait(context.TODO()) if err != nil { log.Printf("[ERROR] %s", err) } newVM, err := finder.VirtualMachine(context.TODO(), vm.Path()) if err != nil { return err } log.Printf("[DEBUG] new vm: %v", newVM) devices, err := newVM.Device(context.TODO()) if err != nil { log.Printf("[DEBUG] Template devices can't be found") return err } for _, dvc := range devices { // Issue 3559/3560: Delete all ethernet devices to add the correct ones later if devices.Type(dvc) == "ethernet" { err := newVM.RemoveDevice(context.TODO(), false, dvc) if err != nil { return err } } } // Add Network devices for _, dvc := range networkDevices { err := newVM.AddDevice( context.TODO(), dvc.GetVirtualDeviceConfigSpec().Device) if err != nil { return err } } // Create the cdroms if needed. if err := createCdroms(c, newVM, dc, vm.cdroms); err != nil { return err } newVM.Properties(context.TODO(), newVM.Reference(), []string{"summary", "config"}, &vm_mo) firstDisk := 0 if vm.template != "" { firstDisk++ } for i := firstDisk; i < len(vm.hardDisks); i++ { log.Printf("[DEBUG] disk index: %v", i) var diskPath string switch { case vm.hardDisks[i].vmdkPath != "": diskPath = vm.hardDisks[i].vmdkPath case vm.hardDisks[i].name != "": snapshotFullDir := vm_mo.Config.Files.SnapshotDirectory split := strings.Split(snapshotFullDir, " ") if len(split) != 2 { return fmt.Errorf("[ERROR] setupVirtualMachine - failed to split snapshot directory: %v", snapshotFullDir) } vmWorkingPath := split[1] diskPath = vmWorkingPath + vm.hardDisks[i].name default: return fmt.Errorf("[ERROR] setupVirtualMachine - Neither vmdk path nor vmdk name was given: %#v", vm.hardDisks[i]) } err = addHardDisk(newVM, vm.hardDisks[i].size, vm.hardDisks[i].iops, vm.hardDisks[i].initType, datastore, diskPath, vm.hardDisks[i].controller) if err != nil { err2 := addHardDisk(newVM, vm.hardDisks[i].size, vm.hardDisks[i].iops, vm.hardDisks[i].initType, datastore, diskPath, vm.hardDisks[i].controller) if err2 != nil { return err2 } return err } } if vm.skipCustomization || vm.template == "" { log.Printf("[DEBUG] VM customization skipped") } else { var identity_options types.BaseCustomizationIdentitySettings if strings.HasPrefix(template_mo.Config.GuestId, "win") { var timeZone int if vm.timeZone == "Etc/UTC" { vm.timeZone = "085" } timeZone, err := strconv.Atoi(vm.timeZone) if err != nil { return fmt.Errorf("Error converting TimeZone: %s", err) } guiUnattended := types.CustomizationGuiUnattended{ AutoLogon: false, AutoLogonCount: 1, TimeZone: int32(timeZone), } customIdentification := types.CustomizationIdentification{} userData := types.CustomizationUserData{ ComputerName: &types.CustomizationFixedName{ Name: strings.Split(vm.name, ".")[0], }, ProductId: vm.windowsOptionalConfig.productKey, FullName: "terraform", OrgName: "terraform", } if vm.windowsOptionalConfig.domainUserPassword != "" && vm.windowsOptionalConfig.domainUser != "" && vm.windowsOptionalConfig.domain != "" { customIdentification.DomainAdminPassword = &types.CustomizationPassword{ PlainText: true, Value: vm.windowsOptionalConfig.domainUserPassword, } customIdentification.DomainAdmin = vm.windowsOptionalConfig.domainUser customIdentification.JoinDomain = vm.windowsOptionalConfig.domain } if vm.windowsOptionalConfig.adminPassword != "" { guiUnattended.Password = &types.CustomizationPassword{ PlainText: true, Value: vm.windowsOptionalConfig.adminPassword, } } identity_options = &types.CustomizationSysprep{ GuiUnattended: guiUnattended, Identification: customIdentification, UserData: userData, } } else { identity_options = &types.CustomizationLinuxPrep{ HostName: &types.CustomizationFixedName{ Name: strings.Split(vm.name, ".")[0], }, Domain: vm.domain, TimeZone: vm.timeZone, HwClockUTC: types.NewBool(true), } } // create CustomizationSpec customSpec := types.CustomizationSpec{ Identity: identity_options, GlobalIPSettings: types.CustomizationGlobalIPSettings{ DnsSuffixList: vm.dnsSuffixes, DnsServerList: vm.dnsServers, }, NicSettingMap: networkConfigs, } log.Printf("[DEBUG] custom spec: %v", customSpec) log.Printf("[DEBUG] VM customization starting") taskb, err := newVM.Customize(context.TODO(), customSpec) if err != nil { return err } _, err = taskb.WaitForResult(context.TODO(), nil) if err != nil { return err } log.Printf("[DEBUG] VM customization finished") } if vm.hasBootableVmdk || vm.template != "" { t, err := newVM.PowerOn(context.TODO()) if err != nil { return err } _, err = t.WaitForResult(context.TODO(), nil) if err != nil { return err } err = newVM.WaitForPowerState(context.TODO(), types.VirtualMachinePowerStatePoweredOn) if err != nil { return err } } return nil }