ETL Processing on Google Cloud Using Dataflow and BigQuery (Python) Reviews

ETL Processing on Google Cloud Using Dataflow and BigQuery (Python) Reviews

68565 reviews

hari hara K. · Reviewed about 2 hours ago

Mauricio D. · Reviewed about 7 hours ago

Tshepiso M. · Reviewed about 9 hours ago

Edson H. · Reviewed about 10 hours ago

Barry S. · Reviewed about 15 hours ago

Grover Javier R. · Reviewed about 16 hours ago

JOIN section Workflow keeps failing. Must be a problem with command: python dataflow_python_examples/data_lake_to_mart.py \ --worker_disk_type="compute.googleapis.com/projects//zones//diskTypes/pd-ssd" \ --project=$PROJECT \ --runner=DataflowRunner \ --machine_type=e2-standard-2 \ --staging_location=gs://$PROJECT/test \ --temp_location gs://$PROJECT/test \ --save_main_session \ --region=us-east1

Barry S. · Reviewed about 17 hours ago

Aleksandra O. · Reviewed about 17 hours ago

Tulasi V. · Reviewed about 18 hours ago

Tom J. · Reviewed about 19 hours ago

Kshitija B. · Reviewed about 19 hours ago

tallapally g. · Reviewed about 20 hours ago

one of the commands is missing a default setting and fails

Chris D. · Reviewed about 22 hours ago

good lab

Adhi K. · Reviewed about 23 hours ago

尚太 神. · Reviewed 1 day ago

Benjamin S. · Reviewed 1 day ago

Ezequiel R. · Reviewed 1 day ago

Ceргeй В. · Reviewed 1 day ago

Shalin M. · Reviewed 1 day ago

christian m. · Reviewed 1 day ago

Christian F. · Reviewed 1 day ago

I have problem with the Console, when I was on the docker container with python, on the second step to run the data_trasformation.py script, I kept having this message: root@48034960a23f:/dataflow# python dataflow_python_examples/data_transformation.py --project=$PROJECT --region=us-east4 --runner=DataflowRunner --machine_type=e2-standard-2 --staging_location=gs://$PROJECT/test --temp_location gs: //$PROJECT/test --input gs://$PROJECT/data_files/head_usa_names.csv --save_main_session Traceback (most recent call last): File "dataflow_python_examples/data_transformation.py", line 166, in <module> run() File "dataflow_python_examples/data_transformation.py", line 132, in run p = beam.Pipeline(options=PipelineOptions(pipeline_args)) File "/usr/local/lib/python3.8/site-packages/apache_beam/pipeline.py", line 212, in __init__ raise ValueError( ValueError: Pipeline has validations errors: Invalid GCS path (gs:///test), given for the option: temp_location. Invalid GCS path (gs:///test), given for the option: staging_location. Invalid Project ID (). Please make sure you specified the Project ID, not project number. So i Had to start from the beggining again, and this made me unable to finish the lab on time!! What do I have to do now with my credits to redo this lab??

Almudena O. · Reviewed 1 day ago

Yves E. · Reviewed 1 day ago

Béla K. · Reviewed 1 day ago

Dheeraj K. · Reviewed 1 day ago

We do not ensure the published reviews originate from consumers who have purchased or used the products. Reviews are not verified by Google.