Azure Data Lake Storage Interview Question-Answer

Q.1 ______ is build on top of Azure Blob Storage.

       A. Azure data lake store gen 1

       B. Azure data lake store gen 2

       C. None of the options mentioned

Ans : Azure data lake store gen 2


Q.2 How may copies of the data does ADLS maintain for high availability of data provisioning.

       A. 1

       B. 2

       C. 3

       D. 4

Ans : 3


Q.3 While creating ADLS Gen 1 using CLI which parameter ensures that the folder is to be created with the given name.

       A. –folder

       B. –file

       C. –directory

       D. None of the options

Ans : –folder


Q.4 ADLS manages what form of data.

       A. Semi-structured

       B. Unstructured

       C. Structured

       D. All of the options

Ans : All of the options


Q.5 ______ function writes the custom queries to bring the desired data-sets for the queries.

       A. Table value function

       B. Procedures

       C. Database

       D. Views

Ans : Table value function


Q.6 ______ language is used to create a script that runs the job which query the data to generate output for analysis.

       A. P-SQL

       B. SQL

       C. C#

       D. U-SQL

Ans : U-SQL


Q.7 ______ is not a build in Extractor in U-SQL

       A. Extractors.Tsv

       B. Extractors.Csv

       C. Extractors.Text

       D. Extractors.Xlsv

Ans : Extractors.Xlsv


Q.8 In ADLS Schema of the data is required to be defined before data is loaded

       A. True

       B. False

Ans : False


Q.9 ADLS Gen1 provides a command line tool, AdlCopy, to copy data from Azure Blob to ADLS and vice-versa.

       A. True

       B. False

Ans : False


Q.10 Account management / Authorization in microsoft Azure is implemented by RBAC which is also a built in feature.

       A. True

       B. False

Ans : True


Q.11 You can enable access to only trusted clients by specifying an IP address or defining a range of IP addresses by setting up the firewall rules to cut down access to your data lake analytics at network level.

       A. True

       B. False

Ans : True


Q.12 What is the purpose of data lake store?

       A. Optimized and dedicated storage for bigdata analytics workload

       B. General purpose object store for variety of storage scenarios

Ans : Optimized and dedicated storage for bigdata analytics workload


Q.13 _______is an analytics job service that writes queries and extracts valuable insights from any scale of data.

       A. Azure data lake store

       B. Azure SQL database

       C. Azure data lake analytics

       D. None of the options

Ans : Azure data lake analytics


Q.14 ADLS Gen1 provides which command line tool to copy data ?

       A. AdlCopy

       B. Azure Data Factory

       C. Azure copy

       D. All of the options

Ans : AdlCopy


Q.15 Which extractor parameter is used to specify column separator whose default value is comma?

       A. rowDelimeter

       B. skipFirstNRows

       C. silent

       D. delimeter

Ans : delimeter


Q.16 Which of the following is not an operation performed by data lake store gen 1?

       A. ngest

       B. Prepare

       C. Analyze

       D. Scripting queries

       E. Store

Ans : Scripting queries


Q.17 The standard RBAC role is

       A. Contributor

       B. Owner

       C. Reader

       D. All of the options

Ans : All of the options


Q.18 If you has not mentioned the –folder parameter while creating the storage account using cli, it creates _____ named folder at the root location?

       A. myfolder

       B. newfolder

       C. mynewfolder

       D. emptyfolder

Ans : mynewfolder


Q.19 Key Capabilities of data lake store gen 2 include which of the following?

       A. Cost effective

       B. Hadoop Compatiblity access

       C. A superset of POSIX permissions

       D. All of the options

Ans : All of the options


Q.20 From which menu of the azure data lake analytics you can explore the added data sources?

       A. Data Extractor

       B. Data explorer

       C. New Job

       D. None of the options

Ans : Data explorer


Q.21 What is the cost per GB that azure incurs in order to store first 100 TB of data in your data lake storage gen 1 account?

       A. 3.00 INR

       B. 2.45 INR

       C. 2.52 INR

       D. 2.58 INR

Ans : 2.58 INR


Q.22 What are the permissions required to append a file that is located in a folder.

       A. Not required any permission

       B. Write and Execute

       C. Read and Write

       D. Read and Execute

Ans : Write and Execute


Q.23 Azure data lake store gen 2 uses which component to organize data files into a hierarchy of directories for efficient data access?

       A. Diversified storage

       B. Hadoop compatible access

       C. Hierarcial namespace

       D. None of the options

Ans : Hierarcial namespace


Q.24 _______ object is used to encapsulate the code that performs certain tasks regularly?

       A. View

       B. Table value function

       C. Procedures

       D. Table

Ans : Procedures


Q.25 Data Operations – Authentication in ADLS is based on

       A. Azure active directory

       B. Account access keys

       C. Shared access key

       D. All of the options

Ans : Azure active directory


Q.26 _________ is used for managing clusters after ingesting large volume of data clusters by extending various open sources such as hadoop, spark, pig, hive etc…

       A. Analytics service

       B. HDInsight

       C. Diversified storage

       D. All of the options

Ans : HDInsight


Q.27 What is the account type that needs to be selected while creating an azure data lake storage gen 2 instance?

       A. Storage gen 2

       B. General Account

       C. StorageV1

       D. StorageV2

Ans : StorageV2


Q.28 _________ helps in creating service alerts and control the cost of Azure data lake implementation.

       A. Security mechanism

       B. Log analytics

       C. Azure data lake analytics

       D. None of the options

Ans : Log analytics


Q.29 Key Capabilities of data lake store gen 1 include which of the following?

       A. Built for Hadoop

       B. Enterprise ready and highly available

       C. Unlimited storage

       D. All of the options

Ans : All of the options


Q.30 What is the file size limit imposed to be stored in ADLS ?

       A. No limit

       B. 100TB

       C. 500TB

       D. 50GB

Ans : No limit


Q.31 Data Storage and analytics jobs combined performed in which layer of the streamed data management?

       A. Data generation

       B. Storage

       C. Data processing

       D. None of the options

Ans : Storage


Q.32 Organizing data into larger files than many small files decreases the performance of the data lake store.

       A. True

       B. False

Ans : False


Leave a Comment

Your email address will not be published. Required fields are marked *