package aws import ( "log" "encoding/json" "fmt" "io/ioutil" "net/http" "strings" "time" "github.com/aws/aws-sdk-go/aws" "github.com/aws/aws-sdk-go/aws/awserr" "github.com/aws/aws-sdk-go/service/emr" "github.com/hashicorp/terraform/helper/resource" "github.com/hashicorp/terraform/helper/schema" ) func resourceAwsEMRCluster() *schema.Resource { return &schema.Resource{ Create: resourceAwsEMRClusterCreate, Read: resourceAwsEMRClusterRead, Update: resourceAwsEMRClusterUpdate, Delete: resourceAwsEMRClusterDelete, Schema: map[string]*schema.Schema{ "name": { Type: schema.TypeString, ForceNew: true, Required: true, }, "release_label": { Type: schema.TypeString, ForceNew: true, Required: true, }, "master_instance_type": { Type: schema.TypeString, Required: true, ForceNew: true, }, "core_instance_type": { Type: schema.TypeString, Optional: true, ForceNew: true, Computed: true, }, "core_instance_count": { Type: schema.TypeInt, Optional: true, Default: 1, }, "cluster_state": { Type: schema.TypeString, Computed: true, }, "log_uri": { Type: schema.TypeString, ForceNew: true, Optional: true, }, "master_public_dns": { Type: schema.TypeString, Computed: true, }, "applications": { Type: schema.TypeSet, Optional: true, ForceNew: true, Elem: &schema.Schema{Type: schema.TypeString}, Set: schema.HashString, }, "termination_protection": { Type: schema.TypeBool, Optional: true, Computed: true, }, "keep_job_flow_alive_when_no_steps": { Type: schema.TypeBool, ForceNew: true, Optional: true, Computed: true, }, "ec2_attributes": { Type: schema.TypeList, MaxItems: 1, Optional: true, ForceNew: true, Elem: &schema.Resource{ Schema: map[string]*schema.Schema{ "key_name": { Type: schema.TypeString, Optional: true, }, "subnet_id": { Type: schema.TypeString, Optional: true, }, "additional_master_security_groups": { Type: schema.TypeString, Optional: true, }, "additional_slave_security_groups": { Type: schema.TypeString, Optional: true, }, "emr_managed_master_security_group": { Type: schema.TypeString, Optional: true, }, "emr_managed_slave_security_group": { Type: schema.TypeString, Optional: true, }, "instance_profile": { Type: schema.TypeString, Required: true, }, "service_access_security_group": { Type: schema.TypeString, Optional: true, }, }, }, }, "bootstrap_action": { Type: schema.TypeSet, Optional: true, ForceNew: true, Elem: &schema.Resource{ Schema: map[string]*schema.Schema{ "name": { Type: schema.TypeString, Required: true, }, "path": { Type: schema.TypeString, Required: true, }, "args": { Type: schema.TypeSet, Optional: true, Elem: &schema.Schema{Type: schema.TypeString}, Set: schema.HashString, }, }, }, }, "tags": tagsSchema(), "configurations": { Type: schema.TypeString, ForceNew: true, Optional: true, }, "service_role": { Type: schema.TypeString, ForceNew: true, Required: true, }, "autoscaling_role": &schema.Schema{ Type: schema.TypeString, ForceNew: true, Optional: true, }, "visible_to_all_users": { Type: schema.TypeBool, Optional: true, Default: true, }, }, } } func resourceAwsEMRClusterCreate(d *schema.ResourceData, meta interface{}) error { conn := meta.(*AWSClient).emrconn log.Printf("[DEBUG] Creating EMR cluster") masterInstanceType := d.Get("master_instance_type").(string) coreInstanceType := masterInstanceType if v, ok := d.GetOk("core_instance_type"); ok { coreInstanceType = v.(string) } coreInstanceCount := d.Get("core_instance_count").(int) applications := d.Get("applications").(*schema.Set).List() keepJobFlowAliveWhenNoSteps := true if v, ok := d.GetOk("keep_job_flow_alive_when_no_steps"); ok { keepJobFlowAliveWhenNoSteps = v.(bool) } terminationProtection := false if v, ok := d.GetOk("termination_protection"); ok { terminationProtection = v.(bool) } instanceConfig := &emr.JobFlowInstancesConfig{ MasterInstanceType: aws.String(masterInstanceType), SlaveInstanceType: aws.String(coreInstanceType), InstanceCount: aws.Int64(int64(coreInstanceCount)), KeepJobFlowAliveWhenNoSteps: aws.Bool(keepJobFlowAliveWhenNoSteps), TerminationProtected: aws.Bool(terminationProtection), } var instanceProfile string if a, ok := d.GetOk("ec2_attributes"); ok { ec2Attributes := a.([]interface{}) attributes := ec2Attributes[0].(map[string]interface{}) if v, ok := attributes["key_name"]; ok { instanceConfig.Ec2KeyName = aws.String(v.(string)) } if v, ok := attributes["subnet_id"]; ok { instanceConfig.Ec2SubnetId = aws.String(v.(string)) } if v, ok := attributes["subnet_id"]; ok { instanceConfig.Ec2SubnetId = aws.String(v.(string)) } if v, ok := attributes["additional_master_security_groups"]; ok { strSlice := strings.Split(v.(string), ",") for i, s := range strSlice { strSlice[i] = strings.TrimSpace(s) } instanceConfig.AdditionalMasterSecurityGroups = aws.StringSlice(strSlice) } if v, ok := attributes["additional_slave_security_groups"]; ok { strSlice := strings.Split(v.(string), ",") for i, s := range strSlice { strSlice[i] = strings.TrimSpace(s) } instanceConfig.AdditionalSlaveSecurityGroups = aws.StringSlice(strSlice) } if v, ok := attributes["emr_managed_master_security_group"]; ok { instanceConfig.EmrManagedMasterSecurityGroup = aws.String(v.(string)) } if v, ok := attributes["emr_managed_slave_security_group"]; ok { instanceConfig.EmrManagedSlaveSecurityGroup = aws.String(v.(string)) } if len(strings.TrimSpace(attributes["instance_profile"].(string))) != 0 { instanceProfile = strings.TrimSpace(attributes["instance_profile"].(string)) } if v, ok := attributes["service_access_security_group"]; ok { instanceConfig.ServiceAccessSecurityGroup = aws.String(v.(string)) } } emrApps := expandApplications(applications) params := &emr.RunJobFlowInput{ Instances: instanceConfig, Name: aws.String(d.Get("name").(string)), Applications: emrApps, ReleaseLabel: aws.String(d.Get("release_label").(string)), ServiceRole: aws.String(d.Get("service_role").(string)), VisibleToAllUsers: aws.Bool(d.Get("visible_to_all_users").(bool)), } if v, ok := d.GetOk("log_uri"); ok { params.LogUri = aws.String(v.(string)) } if v, ok := d.GetOk("autoscaling_role"); ok { params.AutoScalingRole = aws.String(v.(string)) } if instanceProfile != "" { params.JobFlowRole = aws.String(instanceProfile) } if v, ok := d.GetOk("bootstrap_action"); ok { bootstrapActions := v.(*schema.Set).List() params.BootstrapActions = expandBootstrapActions(bootstrapActions) } if v, ok := d.GetOk("tags"); ok { tagsIn := v.(map[string]interface{}) params.Tags = expandTags(tagsIn) } if v, ok := d.GetOk("configurations"); ok { confUrl := v.(string) params.Configurations = expandConfigures(confUrl) } log.Printf("[DEBUG] EMR Cluster create options: %s", params) resp, err := conn.RunJobFlow(params) if err != nil { log.Printf("[ERROR] %s", err) return err } d.SetId(*resp.JobFlowId) log.Println( "[INFO] Waiting for EMR Cluster to be available") stateConf := &resource.StateChangeConf{ Pending: []string{"STARTING", "BOOTSTRAPPING"}, Target: []string{"WAITING", "RUNNING"}, Refresh: resourceAwsEMRClusterStateRefreshFunc(d, meta), Timeout: 75 * time.Minute, MinTimeout: 10 * time.Second, Delay: 30 * time.Second, // Wait 30 secs before starting } _, err = stateConf.WaitForState() if err != nil { return fmt.Errorf("[WARN] Error waiting for EMR Cluster state to be \"WAITING\" or \"RUNNING\": %s", err) } return resourceAwsEMRClusterRead(d, meta) } func resourceAwsEMRClusterRead(d *schema.ResourceData, meta interface{}) error { emrconn := meta.(*AWSClient).emrconn req := &emr.DescribeClusterInput{ ClusterId: aws.String(d.Id()), } resp, err := emrconn.DescribeCluster(req) if err != nil { return fmt.Errorf("Error reading EMR cluster: %s", err) } if resp.Cluster == nil { log.Printf("[DEBUG] EMR Cluster (%s) not found", d.Id()) d.SetId("") return nil } cluster := resp.Cluster if cluster.Status != nil { if *cluster.Status.State == "TERMINATED" { log.Printf("[DEBUG] EMR Cluster (%s) was TERMINATED already", d.Id()) d.SetId("") return nil } if *cluster.Status.State == "TERMINATED_WITH_ERRORS" { log.Printf("[DEBUG] EMR Cluster (%s) was TERMINATED_WITH_ERRORS already", d.Id()) d.SetId("") return nil } d.Set("cluster_state", cluster.Status.State) } instanceGroups, err := fetchAllEMRInstanceGroups(meta, d.Id()) if err == nil { coreGroup := findGroup(instanceGroups, "CORE") if coreGroup != nil { d.Set("core_instance_type", coreGroup.InstanceType) } } d.Set("name", cluster.Name) d.Set("service_role", cluster.ServiceRole) d.Set("autoscaling_role", cluster.AutoScalingRole) d.Set("release_label", cluster.ReleaseLabel) d.Set("log_uri", cluster.LogUri) d.Set("master_public_dns", cluster.MasterPublicDnsName) d.Set("visible_to_all_users", cluster.VisibleToAllUsers) d.Set("tags", tagsToMapEMR(cluster.Tags)) if err := d.Set("applications", flattenApplications(cluster.Applications)); err != nil { log.Printf("[ERR] Error setting EMR Applications for cluster (%s): %s", d.Id(), err) } // Configurations is a JSON document. It's built with an expand method but a // simple string should be returned as JSON if err := d.Set("configurations", cluster.Configurations); err != nil { log.Printf("[ERR] Error setting EMR configurations for cluster (%s): %s", d.Id(), err) } if err := d.Set("ec2_attributes", flattenEc2Attributes(cluster.Ec2InstanceAttributes)); err != nil { log.Printf("[ERR] Error setting EMR Ec2 Attributes: %s", err) } return nil } func resourceAwsEMRClusterUpdate(d *schema.ResourceData, meta interface{}) error { conn := meta.(*AWSClient).emrconn d.Partial(true) if d.HasChange("core_instance_count") { d.SetPartial("core_instance_count") log.Printf("[DEBUG] Modify EMR cluster") groups, err := fetchAllEMRInstanceGroups(meta, d.Id()) if err != nil { log.Printf("[DEBUG] Error finding all instance groups: %s", err) return err } coreInstanceCount := d.Get("core_instance_count").(int) coreGroup := findGroup(groups, "CORE") if coreGroup == nil { return fmt.Errorf("[ERR] Error finding core group") } params := &emr.ModifyInstanceGroupsInput{ InstanceGroups: []*emr.InstanceGroupModifyConfig{ { InstanceGroupId: coreGroup.Id, InstanceCount: aws.Int64(int64(coreInstanceCount) - 1), }, }, } _, errModify := conn.ModifyInstanceGroups(params) if errModify != nil { log.Printf("[ERROR] %s", errModify) return errModify } log.Printf("[DEBUG] Modify EMR Cluster done...") log.Println("[INFO] Waiting for EMR Cluster to be available") stateConf := &resource.StateChangeConf{ Pending: []string{"STARTING", "BOOTSTRAPPING"}, Target: []string{"WAITING", "RUNNING"}, Refresh: resourceAwsEMRClusterStateRefreshFunc(d, meta), Timeout: 40 * time.Minute, MinTimeout: 10 * time.Second, Delay: 5 * time.Second, } _, err = stateConf.WaitForState() if err != nil { return fmt.Errorf("[WARN] Error waiting for EMR Cluster state to be \"WAITING\" or \"RUNNING\" after modification: %s", err) } } if d.HasChange("visible_to_all_users") { d.SetPartial("visible_to_all_users") _, errModify := conn.SetVisibleToAllUsers(&emr.SetVisibleToAllUsersInput{ JobFlowIds: []*string{aws.String(d.Id())}, VisibleToAllUsers: aws.Bool(d.Get("visible_to_all_users").(bool)), }) if errModify != nil { log.Printf("[ERROR] %s", errModify) return errModify } } if d.HasChange("termination_protection") { d.SetPartial("termination_protection") _, errModify := conn.SetTerminationProtection(&emr.SetTerminationProtectionInput{ JobFlowIds: []*string{aws.String(d.Id())}, TerminationProtected: aws.Bool(d.Get("termination_protection").(bool)), }) if errModify != nil { log.Printf("[ERROR] %s", errModify) return errModify } } if err := setTagsEMR(conn, d); err != nil { return err } else { d.SetPartial("tags") } d.Partial(false) return resourceAwsEMRClusterRead(d, meta) } func resourceAwsEMRClusterDelete(d *schema.ResourceData, meta interface{}) error { conn := meta.(*AWSClient).emrconn req := &emr.TerminateJobFlowsInput{ JobFlowIds: []*string{ aws.String(d.Id()), }, } _, err := conn.TerminateJobFlows(req) if err != nil { log.Printf("[ERROR], %s", err) return err } err = resource.Retry(10*time.Minute, func() *resource.RetryError { resp, err := conn.ListInstances(&emr.ListInstancesInput{ ClusterId: aws.String(d.Id()), }) if err != nil { return resource.NonRetryableError(err) } instanceCount := len(resp.Instances) if resp == nil || instanceCount == 0 { log.Printf("[DEBUG] No instances found for EMR Cluster (%s)", d.Id()) return nil } // Collect instance status states, wait for all instances to be terminated // before moving on var terminated []string for j, i := range resp.Instances { if i.Status != nil { if *i.Status.State == "TERMINATED" { terminated = append(terminated, *i.Ec2InstanceId) } } else { log.Printf("[DEBUG] Cluster instance (%d : %s) has no status", j, *i.Ec2InstanceId) } } if len(terminated) == instanceCount { log.Printf("[DEBUG] All (%d) EMR Cluster (%s) Instances terminated", instanceCount, d.Id()) return nil } return resource.RetryableError(fmt.Errorf("[DEBUG] EMR Cluster (%s) has (%d) Instances remaining, retrying", d.Id(), len(resp.Instances))) }) if err != nil { log.Printf("[ERR] Error waiting for EMR Cluster (%s) Instances to drain", d.Id()) } d.SetId("") return nil } func expandApplications(apps []interface{}) []*emr.Application { appOut := make([]*emr.Application, 0, len(apps)) for _, appName := range expandStringList(apps) { app := &emr.Application{ Name: appName, } appOut = append(appOut, app) } return appOut } func flattenApplications(apps []*emr.Application) []interface{} { appOut := make([]interface{}, 0, len(apps)) for _, app := range apps { appOut = append(appOut, *app.Name) } return appOut } func flattenEc2Attributes(ia *emr.Ec2InstanceAttributes) []map[string]interface{} { attrs := map[string]interface{}{} result := make([]map[string]interface{}, 0) if ia.Ec2KeyName != nil { attrs["key_name"] = *ia.Ec2KeyName } if ia.Ec2SubnetId != nil { attrs["subnet_id"] = *ia.Ec2SubnetId } if ia.IamInstanceProfile != nil { attrs["instance_profile"] = *ia.IamInstanceProfile } if ia.EmrManagedMasterSecurityGroup != nil { attrs["emr_managed_master_security_group"] = *ia.EmrManagedMasterSecurityGroup } if ia.EmrManagedSlaveSecurityGroup != nil { attrs["emr_managed_slave_security_group"] = *ia.EmrManagedSlaveSecurityGroup } if len(ia.AdditionalMasterSecurityGroups) > 0 { strs := aws.StringValueSlice(ia.AdditionalMasterSecurityGroups) attrs["additional_master_security_groups"] = strings.Join(strs, ",") } if len(ia.AdditionalSlaveSecurityGroups) > 0 { strs := aws.StringValueSlice(ia.AdditionalSlaveSecurityGroups) attrs["additional_slave_security_groups"] = strings.Join(strs, ",") } if ia.ServiceAccessSecurityGroup != nil { attrs["service_access_security_group"] = *ia.ServiceAccessSecurityGroup } result = append(result, attrs) return result } func loadGroups(d *schema.ResourceData, meta interface{}) ([]*emr.InstanceGroup, error) { emrconn := meta.(*AWSClient).emrconn reqGrps := &emr.ListInstanceGroupsInput{ ClusterId: aws.String(d.Id()), } respGrps, errGrps := emrconn.ListInstanceGroups(reqGrps) if errGrps != nil { return nil, fmt.Errorf("Error reading EMR cluster: %s", errGrps) } return respGrps.InstanceGroups, nil } func findGroup(grps []*emr.InstanceGroup, typ string) *emr.InstanceGroup { for _, grp := range grps { if grp.InstanceGroupType != nil { if *grp.InstanceGroupType == typ { return grp } } } return nil } func expandTags(m map[string]interface{}) []*emr.Tag { var result []*emr.Tag for k, v := range m { result = append(result, &emr.Tag{ Key: aws.String(k), Value: aws.String(v.(string)), }) } return result } func tagsToMapEMR(ts []*emr.Tag) map[string]string { result := make(map[string]string) for _, t := range ts { result[*t.Key] = *t.Value } return result } func diffTagsEMR(oldTags, newTags []*emr.Tag) ([]*emr.Tag, []*emr.Tag) { // First, we're creating everything we have create := make(map[string]interface{}) for _, t := range newTags { create[*t.Key] = *t.Value } // Build the list of what to remove var remove []*emr.Tag for _, t := range oldTags { old, ok := create[*t.Key] if !ok || old != *t.Value { // Delete it! remove = append(remove, t) } } return expandTags(create), remove } func setTagsEMR(conn *emr.EMR, d *schema.ResourceData) error { if d.HasChange("tags") { oraw, nraw := d.GetChange("tags") o := oraw.(map[string]interface{}) n := nraw.(map[string]interface{}) create, remove := diffTagsEMR(expandTags(o), expandTags(n)) // Set tags if len(remove) > 0 { log.Printf("[DEBUG] Removing tags: %s", remove) k := make([]*string, len(remove), len(remove)) for i, t := range remove { k[i] = t.Key } _, err := conn.RemoveTags(&emr.RemoveTagsInput{ ResourceId: aws.String(d.Id()), TagKeys: k, }) if err != nil { return err } } if len(create) > 0 { log.Printf("[DEBUG] Creating tags: %s", create) _, err := conn.AddTags(&emr.AddTagsInput{ ResourceId: aws.String(d.Id()), Tags: create, }) if err != nil { return err } } } return nil } func expandBootstrapActions(bootstrapActions []interface{}) []*emr.BootstrapActionConfig { actionsOut := []*emr.BootstrapActionConfig{} for _, raw := range bootstrapActions { actionAttributes := raw.(map[string]interface{}) actionName := actionAttributes["name"].(string) actionPath := actionAttributes["path"].(string) actionArgs := actionAttributes["args"].(*schema.Set).List() action := &emr.BootstrapActionConfig{ Name: aws.String(actionName), ScriptBootstrapAction: &emr.ScriptBootstrapActionConfig{ Path: aws.String(actionPath), Args: expandStringList(actionArgs), }, } actionsOut = append(actionsOut, action) } return actionsOut } func expandConfigures(input string) []*emr.Configuration { configsOut := []*emr.Configuration{} if strings.HasPrefix(input, "http") { if err := readHttpJson(input, &configsOut); err != nil { log.Printf("[ERR] Error reading HTTP JSON: %s", err) } } else if strings.HasSuffix(input, ".json") { if err := readLocalJson(input, &configsOut); err != nil { log.Printf("[ERR] Error reading local JSON: %s", err) } } else { if err := readBodyJson(input, &configsOut); err != nil { log.Printf("[ERR] Error reading body JSON: %s", err) } } log.Printf("[DEBUG] Expanded EMR Configurations %s", configsOut) return configsOut } func readHttpJson(url string, target interface{}) error { r, err := http.Get(url) if err != nil { return err } defer r.Body.Close() return json.NewDecoder(r.Body).Decode(target) } func readLocalJson(localFile string, target interface{}) error { file, e := ioutil.ReadFile(localFile) if e != nil { log.Printf("[ERROR] %s", e) return e } return json.Unmarshal(file, target) } func readBodyJson(body string, target interface{}) error { log.Printf("[DEBUG] Raw Body %s\n", body) err := json.Unmarshal([]byte(body), target) if err != nil { log.Printf("[ERROR] parsing JSON %s", err) return err } return nil } func resourceAwsEMRClusterStateRefreshFunc(d *schema.ResourceData, meta interface{}) resource.StateRefreshFunc { return func() (interface{}, string, error) { conn := meta.(*AWSClient).emrconn log.Printf("[INFO] Reading EMR Cluster Information: %s", d.Id()) params := &emr.DescribeClusterInput{ ClusterId: aws.String(d.Id()), } resp, err := conn.DescribeCluster(params) if err != nil { if awsErr, ok := err.(awserr.Error); ok { if "ClusterNotFound" == awsErr.Code() { return 42, "destroyed", nil } } log.Printf("[WARN] Error on retrieving EMR Cluster (%s) when waiting: %s", d.Id(), err) return nil, "", err } emrc := resp.Cluster if emrc == nil { return 42, "destroyed", nil } if resp.Cluster.Status != nil { log.Printf("[DEBUG] EMR Cluster status (%s): %s", d.Id(), *resp.Cluster.Status) } return emrc, *emrc.Status.State, nil } }