Associate-Developer-Apache-Spark-3.5 無料問題集「Databricks Certified Associate Developer for Apache Spark 3.5 - Python」

Given:
python
CopyEdit
spark.sparkContext.setLogLevel("<LOG_LEVEL>")
Which set contains the suitable configuration settings for Spark driver LOG_LEVELs?

解説: (JPNTest メンバーにのみ表示されます)
What is the difference betweendf.cache()anddf.persist()in Spark DataFrame?

解説: (JPNTest メンバーにのみ表示されます)
Which UDF implementation calculates the length of strings in a Spark DataFrame?

解説: (JPNTest メンバーにのみ表示されます)
What is the relationship between jobs, stages, and tasks during execution in Apache Spark?
Options:

解説: (JPNTest メンバーにのみ表示されます)
A Spark developer is building an app to monitor task performance. They need to track the maximum task processing time per worker node and consolidate it on the driver for analysis.
Which technique should be used?

解説: (JPNTest メンバーにのみ表示されます)
A data engineer wants to create a Streaming DataFrame that reads from a Kafka topic called feed.

Which code fragment should be inserted in line 5 to meet the requirement?
Code context:
spark \
.readStream \
.format("kafka") \
.option("kafka.bootstrap.servers","host1:port1,host2:port2") \
.[LINE5] \
.load()
Options:

解説: (JPNTest メンバーにのみ表示されます)
A developer is running Spark SQL queries and notices underutilization of resources. Executors are idle, and the number of tasks per stage is low.
What should the developer do to improve cluster utilization?

解説: (JPNTest メンバーにのみ表示されます)
Given a DataFramedfthat has 10 partitions, after running the code:
result = df.coalesce(20)
How many partitions will the result DataFrame have?

解説: (JPNTest メンバーにのみ表示されます)

弊社を連絡する

我々は12時間以内ですべてのお問い合わせを答えます。

オンラインサポート時間:( UTC+9 ) 9:00-24:00
月曜日から土曜日まで

サポート:現在連絡