Automating your BigQuery Data Pipeline with Cloud Dataprep avaliações

Automating your BigQuery Data Pipeline with Cloud Dataprep avaliações

2059 avaliações

Subhan M. · Revisado há over 3 years

Subhan M. · Revisado há over 3 years

Lubna S. · Revisado há over 3 years

Subhan M. · Revisado há over 3 years

Ashok K. · Revisado há over 3 years

dil b. · Revisado há over 3 years

jonu c. · Revisado há over 3 years

This is the worst ever written lab in qwiklabs. A guy has to discover what is meant to be done at several steps and in 1 silly hour and 30 silly minutes. Finally done at my 4th attempt! The time required is like 2H 30M. And there is no problem even if it requires 3H, but make it clear. Several issues: 1. Once again, Add Datasets to the flow. Choose Import Datasets, then BigQuery again. --> Use the right terms. Add Dataset and Add Datasets are different functions in this context because both exist. 2. By default, Dataprep has inferred this column to be an array/list. --> Not always. In fact, i found dataprep to not be stable in terms of suggestions and columns formats as this may vary usage by usage. On my first time it inferred it was a string and that changes everything because "Flatten array values into new rows" doesn't appear as suggestion. Instructions shall be provided for the case this happens because one first need to change for Array type so the right Flatten suggestion can appear. 3. (...) If you mouse over the histogram of the column, you can see that the bars represent individual keys inside of the object. --> Nope... this behaviour didn't happen to me. 6. In the first row, highlight the middle portion of the citation_publication_number. You will see suggestions to extract a digit pattern. --> Nope. Didn't happend to me. There are several option and you shall make clear that one shall not chose the one that includes the specific number. 2. Modify the join keys to use citation_publication_id == patent_id as the join keys. --> What about "Click Next."? Still in point 2 and after "Keep the following columns and save the step:" + variables: CLICK Review. CLICK Add to Recipe. 7. One by one, add the following columns to the ‘Group by' field and observe how the preview column updates: --> In fact is ‘Group rows by' ... otherwise, why write everything at all? 9. Now to do the partial deduplication, you will simply need to get rid of any rows where partial_dupe is greater than 1. Click on the histogram in the partial_dupe column and add the suggestion to Keep rows where (1 <= partial_dupe) && (partial_dupe < 2). --> Not allways. If the suggestion doesn't appear, what one can do? --> Maybe you want to write: If this is not working, keep reading to get the idea and resume the process on Step 12. 18. Finally, inner join the inventor dataset, using inventor_id == id as the join keys. Keep all columns from both datasets and Add the step. --> Oh boy. What if you would stepwise this line as provided in proper technological education? And that's all! Goes with a 1 star because the time to execute this lab is horrible and the instructional wording leads to frustration. Thanks.

Américo A. · Revisado há over 3 years

This is the worst ever written lab in qwiklabs. A guy has to discover what is meant to be done at several steps and in 1 silly hour and 30 silly minutes.

Américo A. · Revisado há over 3 years

arshad a. · Revisado há over 3 years

Jimmy H. · Revisado há over 3 years

ashraf n. · Revisado há over 3 years

Hal B. · Revisado há over 3 years

Cemile A. · Revisado há over 3 years

dfdf d. · Revisado há over 3 years

Corey W. · Revisado há over 3 years

Hal B. · Revisado há over 3 years

dfdf d. · Revisado há over 3 years

dfdf d. · Revisado há over 3 years

dfdf d. · Revisado há over 3 years

dil b. · Revisado há over 3 years

Good

King P. · Revisado há over 3 years

Good

King P. · Revisado há over 3 years

Good

King P. · Revisado há over 3 years

Não garantimos que as avaliações publicadas sejam de consumidores que compraram ou usaram os produtos. As avaliações não são verificadas pelo Google.