diff --git a/examples/data-solutions/gcs-to-bq-with-least-privileges/README.md b/examples/data-solutions/gcs-to-bq-with-least-privileges/README.md index 79e09c31..1b51e218 100644 --- a/examples/data-solutions/gcs-to-bq-with-least-privileges/README.md +++ b/examples/data-solutions/gcs-to-bq-with-least-privileges/README.md @@ -47,6 +47,10 @@ In this example you can also configure users or group of user to assign them vie ## Deploy your enviroment +We assume the identiy running the following steps has the following role: + - `resourcemanager.projectCreator` in case a new project will be created. + - `owner` on the project in case you use an existing project. + Run Terraform init: ``` @@ -56,10 +60,9 @@ $ terraform init Configure the Terraform variable in your `terraform.tfvars` file. You need to spefify at least the following variables: ``` -billing_account = "0011322-334455-667788" -root_node = "folders/123456789012" -project_name = "test-demo-tf-001" -data_eng_users = ["your_email@domani.example"] +data_eng_principals = ["user:data-eng@domain.com"] +project_id = "datalake-001" +prefix = "prefix" ``` You can run now: @@ -105,15 +108,15 @@ gcloud --impersonate-service-account=orch-test@PROJECT.iam.gserviceaccount.com d --region REGION \ --disable-public-ips \ --subnetwork https://www.googleapis.com/compute/v1/projects/PROJECT/regions/REGION/subnetworks/subnet \ - --staging-location gs://PROJECT-eu-df-tmplocation \ - --service-account-email df-test@PROJECT.iam.gserviceaccount.com \ + --staging-location gs://PREFIX-df-tmp \ + --service-account-email df-loading@PROJECT.iam.gserviceaccount.com \ --parameters \ javascriptTextTransformFunctionName=transform,\ -JSONPath=gs://PROJECT-eu-data/person_schema.json,\ -javascriptTextTransformGcsPath=gs://PROJECT-eu-data/person_udf.js,\ -inputFilePattern=gs://PROJECT-eu-data/person.csv,\ -outputTable=PROJECT:bq_dataset.person,\ -bigQueryLoadingTemporaryDirectory=gs://PROJECT-eu-df-tmplocation +JSONPath=gs://PREFIX-data/person_schema.json,\ +javascriptTextTransformGcsPath=gs://PREFIX-data/person_udf.js,\ +inputFilePattern=gs://PREFIX-data/person.csv,\ +outputTable=PROJECT:datalake.person,\ +bigQueryLoadingTemporaryDirectory=gs://PREFIX-df-tmp ``` You can check data imported into Google BigQuery using the command returned in the terraform output as `command-03-bq`. Below an example: diff --git a/examples/third-party-solutions/openshift/README.md b/examples/third-party-solutions/openshift/README.md index 37e58e88..b0be97ba 100644 --- a/examples/third-party-solutions/openshift/README.md +++ b/examples/third-party-solutions/openshift/README.md @@ -172,7 +172,7 @@ terraform init terraform apply ``` -If you want to preserve state (which is always a good idea), configure a [GCS backend](https://www.terraform.io/docs/language/settings/backends/gcs.html) as you would do for any other Terraform GCP setup. +If you want to preserve state (which is always a good idea), configure a [GCS backend](https://www.terraform.io/language/settings/backends/gcs) as you would do for any other Terraform GCP setup. ### Waiting for bootstrap to complete