Free Databricks Databricks-Certified-Professional-Data-Scientist Exam Actual Questions

The questions for Databricks-Certified-Professional-Data-Scientist were last updated On Apr 24, 2024

Question No. 1

Feature Hashing approach is "SGD-based classifiers avoid the need to predetermine vector size by simply picking a reasonable size and shoehorning the training data into vectors of that size" now with large vectors or with multiple locations per feature in Feature hashing?

Show Answer Hide Answer
Correct Answer: B

Question No. 2

What are the advantages of the Hashing Features?

Show Answer Hide Answer
Correct Answer: A, B

Question No. 3

Question-3: In machine learning, feature hashing, also known as the hashing trick (by analogy to the kernel trick), is a fast and space-efficient way of vectorizing features (such as the words in a language), i.e., turning arbitrary features into indices in a vector or matrix. It works by applying a hash function to the features and using their hash values modulo the number of features as indices directly, rather than looking the indices up in an associative array. So what is the primary reason of the hashing trick for building classifiers?

Show Answer Hide Answer
Correct Answer: B

Question No. 4

Suppose A, B , and C are events. The probability of A given B , relative to P(|C), is the same as the probability of A given B and C (relative to P ). That is,

Show Answer Hide Answer
Correct Answer: A

From the definition, P(A,B|C) P(B|C) =P(A,B.C)/P(C) P(B.C)/P(C) =P(A,B.C)

P(B,C) =P(A|BC)

This follows from the definition of conditional probability, applied twice: P(A,B)=(PA|B)P(B)


Question No. 5

What is the considerable difference between L1 and L2 regularization?

Show Answer Hide Answer
Correct Answer: B