All Coursera Quiz Answers

Microsoft Azure Databricks for Data Engineering Week 3 | Test prep Quiz Answers

In this article i am gone to share Coursera Course: Microsoft Azure Databricks for Data Engineering Week 3 | Test prep Quiz Answers with you..

Enrol Link: Microsoft Azure Databricks for Data Engineering

Microsoft Azure Databricks for Data Engineering Week 3 | Test prep Quiz Answers


 

Test prep Quiz Answers

Question 1)
When creating a new cluster in Azure Databricks there are three Cluster Modes that can be set. Which of the following are valid Cluster Modes? Select three valid options.

  • Low Concurrency
  • Single Node
  • Multi Node
  • Standard
  • High Concurrency

Question 2)
Which DataFrame method do you use to create a temporary view?

  • createTempView()
  • createTempViewDF()
  • createOrReplaceTempView()

Question 3)
How do you define a DataFrame object?

  • Use the DF.create() syntax
  • Use the createDataFrame() function
  • Introduce a variable name and equate it to something like myDataFrameDF =

Question 4)
How do you cache data into the memory of the local executor for instant access?

  • .save().inMemory()
  • .cache()
  • .inMemory().save()

Question 5)
What is the Python syntax for defining a DataFrame in Spark from an existing Parquet file in DBFS?

  • IPGeocodeDF = spark.read.parquet(“dbfs:/mnt/training/ip-geocode.parquet”)
  • IPGeocodeDF = spark.parquet.read(“dbfs:/mnt/training/ip-geocode.parquet”)
  • IPGeocodeDF = parquet.read(“dbfs:/mnt/training/ip-geocode.parquet”)

Question 6)
Among the most powerful components of Spark are Spark SQL. At its core lies the Catalyst optimizer. When you execute code, Spark SQL uses Catalyst’s general tree transformation framework in four phases. In which order are these phases carried out?

1: logical plan optimization
2: analyzing a logical plan to resolve references
3: code generation to compile parts of the query to Java bytecode
4: physical planning

  • 2, 3, 1, 4
  • 1, 2, 3, 4
  • 2, 1, 4, 3
  • 3, 2, 1, 4

Question 7)
Which of the following statements describes a wide transformation?

  • A wide transformation applies data transformation over a large number of columns
  • A wide transformation can be applied per partition/worker with no need to share or shuffle data to other workers
  • A wide transformation requires sharing data across workers. It does so by shuffling data.

Question 8)
Which of the following statements describes a narrow transformation?

  • A narrow transformation can be applied per partition/worker with no need to share or shuffle data to other workers
  • A narrow transformation applies data transformation over a large number of columns
  • A narrow transformation requires sharing data across workers. It does so by shuffling data.

Question 9)
Which feature of Spark determines how your code is executed?

  • Tungsten Record Format
  • Java Garbage Collection
  • Catalyst Optimizer

Question 10)
Which feature of Spark of optimization is used in shuffling operations during wide transformations?

  • Catalyst Optimizer
  • Lazy Execution
  • Tungsten Record Format